repo
stringclasses
856 values
pull_number
int64
3
127k
instance_id
stringlengths
12
58
issue_numbers
sequencelengths
1
5
base_commit
stringlengths
40
40
patch
stringlengths
67
1.54M
test_patch
stringlengths
0
107M
problem_statement
stringlengths
3
307k
hints_text
stringlengths
0
908k
created_at
timestamp[s]
ray-project/ray
7,443
ray-project__ray-7443
[ "7438" ]
4198db5038dd465a5d5db91b0ca84d98391aec2d
diff --git a/rllib/agents/trainer_template.py b/rllib/agents/trainer_template.py --- a/rllib/agents/trainer_template.py +++ b/rllib/agents/trainer_template.py @@ -38,40 +38,43 @@ def build_trainer(name, Arguments: name (str): name of the trainer (e.g., "PPO") default_policy (cls): the default Policy class to use - default_config (dict): The default config dict of the algorithm, - otherwise uses the Trainer default config. - validate_config (func): optional callback that checks a given config - for correctness. It may mutate the config as needed. - get_initial_state (func): optional function that returns the initial - state dict given the trainer instance as an argument. The state - dict must be serializable so that it can be checkpointed, and will - be available as the `trainer.state` variable. - get_policy_class (func): optional callback that takes a config and - returns the policy class to override the default with - before_init (func): optional function to run at the start of trainer - init that takes the trainer instance as argument - make_workers (func): override the method that creates rollout workers. - This takes in (trainer, env_creator, policy, config) as args. - make_policy_optimizer (func): optional function that returns a - PolicyOptimizer instance given (WorkerSet, config) - after_init (func): optional function to run at the end of trainer init - that takes the trainer instance as argument - before_train_step (func): optional callback to run before each train() - call. It takes the trainer instance as an argument. - after_optimizer_step (func): optional callback to run after each - step() call to the policy optimizer. It takes the trainer instance - and the policy gradient fetches as arguments. - after_train_result (func): optional callback to run at the end of each - train() call. It takes the trainer instance and result dict as - arguments, and may mutate the result dict as needed. - collect_metrics_fn (func): override the method used to collect metrics. - It takes the trainer instance as argumnt. - before_evaluate_fn (func): callback to run before evaluation. This - takes the trainer instance as argument. - mixins (list): list of any class mixins for the returned trainer class. - These mixins will be applied in order and will have higher - precedence than the Trainer class - training_pipeline (func): Experimental support for custom + default_config (Optional[dict]): The default config dict of the + algorithm. If None, uses the Trainer default config. + validate_config (Optional[callable]): Optional callback that checks a + given config for correctness. It may mutate the config as needed. + get_initial_state (Optional[callable]): Optional callable that returns + the initial state dict given the trainer instance as an argument. + The state dict must be serializable so that it can be checkpointed, + and will be available as the `trainer.state` variable. + get_policy_class (Optional[callable]): Optional callable that takes a + Trainer config and returns the policy class to override the default + with. + before_init (Optional[callable]): Optional callable to run at the start + of trainer init that takes the trainer instance as argument. + make_workers (Optional[callable]): Override the default method that + creates rollout workers. This takes in (trainer, env_creator, + policy, config) as args. + make_policy_optimizer (Optional[callable]): Optional callable that + returns a PolicyOptimizer instance given (WorkerSet, config). + after_init (Optional[callable]): Optional callable to run at the end of + trainer init that takes the trainer instance as argument. + before_train_step (Optional[callable]): Optional callable to run before + each train() call. It takes the trainer instance as an argument. + after_optimizer_step (Optional[callable]): Optional callable to run + after each step() call to the policy optimizer. It takes the + trainer instance and the policy gradient fetches as arguments. + after_train_result (Optional[callable]): Optional callable to run at + the end of each train() call. It takes the trainer instance and + result dict as arguments, and may mutate the result dict as needed. + collect_metrics_fn (Optional[callable]): Optional callable to override + the default method used to collect metrics. Takes the trainer + instance as argumnt. + before_evaluate_fn (Optional[callable]): Optional callable to run + before evaluation. Takes the trainer instance as argument. + mixins (Optional[List[class]]): Optional list of mixin class(es) for + the returned trainer class. These mixins will be applied in order + and will have higher precedence than the Trainer class. + training_pipeline (Optional[callable]): Experimental support for custom training pipelines. This overrides `make_policy_optimizer`. Returns: @@ -92,20 +95,26 @@ def __init__(self, config=None, env=None, logger_creator=None): def _init(self, config, env_creator): if validate_config: validate_config(config) + if get_initial_state: self.state = get_initial_state(self) else: self.state = {} - if get_policy_class is None: - policy = default_policy - else: - policy = get_policy_class(config) + + # Override default policy if `get_policy_class` is provided. + if get_policy_class is not None: + self._policy = get_policy_class(config) + if before_init: before_init(self) + + # Creating all workers (excluding evaluation workers). if make_workers: - self.workers = make_workers(self, env_creator, policy, config) + self.workers = make_workers(self, env_creator, self._policy, + config) else: - self.workers = self._make_workers(env_creator, policy, config, + self.workers = self._make_workers(env_creator, self._policy, + config, self.config["num_workers"]) self.train_pipeline = None self.optimizer = None
[rllib] Evaluation doesn’t work properly for PyTorch. This is not a contribution. Ray version: 0.8.1 Python version: 3.6.8 Pytorch version: 1.4 OS: Ubuntu 18.04 Docker Training with Pytorch with evaluation enabled doesn’t work properly as it ends up calling tensorflow code. Here's a script that reproduces the problem: ```python import gym from gym.spaces import Box from ray import tune class DummyEnv(gym.Env): def __init__(self, config): self.action_space = Box(0.0, 1.0, shape=(1,)) self.observation_space = Box(0.0, 1.0, shape=(1, )) def reset(self): self.num_steps = 0 return [0.0] def step(self, action): self.num_steps += 1 return [0.0], 1.0, True if self.num_steps == 30 else False, {} tune.run( "PPO", config={ "env": DummyEnv, "use_pytorch": True, "num_workers": 1, "evaluation_interval": 1, "evaluation_config": {"seed": 1}, } ) ``` This will result in the following exception: ``` ray.exceptions.RayTaskError(AssertionError): ray::PPO.train() (pid=33990) File "python/ray/_raylet.pyx", line 452, in ray._raylet.execute_task File "python/ray/_raylet.pyx", line 430, in ray._raylet.execute_task.function_executor File "/usr/local/lib/python3.6/dist-packages/ray/rllib/agents/trainer.py", line 513, in train evaluation_metrics = self._evaluate() File "/usr/local/lib/python3.6/dist-packages/ray/rllib/agents/trainer.py", line 685, in _evaluate lambda w: w.restore(ray.get(weights))) File "/usr/local/lib/python3.6/dist-packages/ray/rllib/evaluation/worker_set.py", line 110, in foreach_worker local_result = [func(self.local_worker())] File "/usr/local/lib/python3.6/dist-packages/ray/rllib/agents/trainer.py", line 685, in <lambda> lambda w: w.restore(ray.get(weights))) File "/usr/local/lib/python3.6/dist-packages/ray/rllib/evaluation/rollout_worker.py", line 764, in restore self.policy_map[pid].set_state(state) File "/usr/local/lib/python3.6/dist-packages/ray/rllib/policy/policy.py", line 320, in set_state self.set_weights(state) File "/usr/local/lib/python3.6/dist-packages/ray/rllib/policy/tf_policy.py", line 290, in set_weights return self._variables.set_weights(weights) File "/usr/local/lib/python3.6/dist-packages/ray/experimental/tf_utils.py", line 185, in set_weights assert assign_list, ("No variables in the input matched those in the " AssertionError: No variables in the input matched those in the network. Possible cause: Two networks were defined in the same TensorFlow graph. To fix this, place each network definition in its own tf.Graph. ```
This is a bug (seems to have been there for a while). Fixing this now ...
2020-03-04T09:46:15
ray-project/ray
7,444
ray-project__ray-7444
[ "7136" ]
476b5c6196fa734794e395a53d2506e7c8485d12
diff --git a/rllib/agents/ars/ars.py b/rllib/agents/ars/ars.py --- a/rllib/agents/ars/ars.py +++ b/rllib/agents/ars/ars.py @@ -165,8 +165,7 @@ def _init(self, config, env_creator): # PyTorch check. if config["use_pytorch"]: raise ValueError( - "ARS does not support PyTorch yet! Use tf instead." - ) + "ARS does not support PyTorch yet! Use tf instead.") env = env_creator(config["env_config"]) from ray.rllib import models @@ -301,7 +300,7 @@ def _stop(self): w.__ray_terminate__.remote() @override(Trainer) - def compute_action(self, observation): + def compute_action(self, observation, *args, **kwargs): return self.policy.compute(observation, update=True)[0] def _collect_results(self, theta_id, min_episodes): diff --git a/rllib/agents/es/es.py b/rllib/agents/es/es.py --- a/rllib/agents/es/es.py +++ b/rllib/agents/es/es.py @@ -171,8 +171,7 @@ def _init(self, config, env_creator): # PyTorch check. if config["use_pytorch"]: raise ValueError( - "ES does not support PyTorch yet! Use tf instead." - ) + "ES does not support PyTorch yet! Use tf instead.") policy_params = {"action_noise_std": 0.01} @@ -292,7 +291,7 @@ def _train(self): return result @override(Trainer) - def compute_action(self, observation): + def compute_action(self, observation, *args, **kwargs): return self.policy.compute(observation, update=False)[0] @override(Trainer) diff --git a/rllib/rollout.py b/rllib/rollout.py --- a/rllib/rollout.py +++ b/rllib/rollout.py @@ -15,6 +15,7 @@ from ray.rllib.env import MultiAgentEnv from ray.rllib.env.base_env import _DUMMY_AGENT_ID from ray.rllib.evaluation.episode import _flatten_action +from ray.rllib.evaluation.worker_set import WorkerSet from ray.rllib.policy.sample_batch import DEFAULT_POLICY_ID from ray.rllib.utils.deprecation import deprecation_warning from ray.tune.utils import merge_dicts @@ -339,7 +340,7 @@ def rollout(agent, if saver is None: saver = RolloutSaver() - if hasattr(agent, "workers"): + if hasattr(agent, "workers") and isinstance(agent.workers, WorkerSet): env = agent.workers.local_worker().env multiagent = isinstance(env, MultiAgentEnv) if agent.workers.local_worker().multiagent: @@ -349,15 +350,22 @@ def rollout(agent, policy_map = agent.workers.local_worker().policy_map state_init = {p: m.get_initial_state() for p, m in policy_map.items()} use_lstm = {p: len(s) > 0 for p, s in state_init.items()} - action_init = { - p: _flatten_action(m.action_space.sample()) - for p, m in policy_map.items() - } else: env = gym.make(env_name) multiagent = False + try: + policy_map = {DEFAULT_POLICY_ID: agent.policy} + except AttributeError: + raise AttributeError( + "Agent ({}) does not have a `policy` property! This is needed " + "for performing (trained) agent rollouts.".format(agent)) use_lstm = {DEFAULT_POLICY_ID: False} + action_init = { + p: _flatten_action(m.action_space.sample()) + for p, m in policy_map.items() + } + # If monitoring has been requested, manually wrap our environment with a # gym monitor, which is set to record every episode. if video_dir:
[RLlib] Cannot rollout ES after training. Using ray 0.8.1. Having trained an agent with ES: ``` rllib train --run=ES --env=LunarLander-v2 --checkpoint-freq=10 ``` I can't play it. (It works with PPO). Executing this: ``` rllib rollout checkpoint_50/checkpoint-50 --run ES --env LunarLander-v2 --steps 600 ``` results in: ``` 2020-02-12 05:41:30,922 INFO trainable.py:423 -- Current state after restoring: {'_iteration': 50, '_timesteps_total': 16979666, '_time_total': 2578.4048051834106, '_episodes_total': None} Traceback (most recent call last): File "/home/andriy/miniconda3/envs/myproj/bin/rllib", line 8, in <module> sys.exit(cli()) File "/home/andriy/miniconda3/envs/myproj/lib/python3.7/site-packages/ray/rllib/scripts.py", line 36, in cli rollout.run(options, rollout_parser) File "/home/andriy/miniconda3/envs/myproj/lib/python3.7/site-packages/ray/rllib/rollout.py", line 265, in run args.no_render, args.monitor) File "/home/andriy/miniconda3/envs/myproj/lib/python3.7/site-packages/ray/rllib/rollout.py", line 364, in rollout prev_action=prev_actions[agent_id], File "/home/andriy/miniconda3/envs/myproj/lib/python3.7/site-packages/ray/rllib/rollout.py", line 272, in __missing__ self[key] = value = self.default_factory(key) File "/home/andriy/miniconda3/envs/myproj/lib/python3.7/site-packages/ray/rllib/rollout.py", line 340, in <lambda> lambda agent_id: action_init[mapping_cache[agent_id]]) NameError: free variable 'action_init' referenced before assignment in enclosing scope ```
I'll take a look.
2020-03-04T10:20:33
ray-project/ray
7,501
ray-project__ray-7501
[ "7412" ]
a644060daa4c22aabedcfd9b891d43f9afa88f97
diff --git a/python/ray/tune/analysis/experiment_analysis.py b/python/ray/tune/analysis/experiment_analysis.py --- a/python/ray/tune/analysis/experiment_analysis.py +++ b/python/ray/tune/analysis/experiment_analysis.py @@ -7,7 +7,6 @@ except ImportError: pd = None -from ray.tune.checkpoint_manager import Checkpoint from ray.tune.error import TuneError from ray.tune.result import EXPR_PROGRESS_FILE, EXPR_PARAM_FILE,\ CONFIG_PREFIX, TRAINING_ITERATION @@ -43,7 +42,6 @@ def dataframe(self, metric=None, mode=None): metric (str): Key for trial info to order on. If None, uses last result. mode (str): One of [min, max]. - """ rows = self._retrieve_rows(metric=metric, mode=mode) all_configs = self.get_all_configs(prefix=True) @@ -59,7 +57,6 @@ def get_best_config(self, metric, mode="max"): Args: metric (str): Key for trial info to order on. mode (str): One of [min, max]. - """ rows = self._retrieve_rows(metric=metric, mode=mode) all_configs = self.get_all_configs() @@ -73,7 +70,6 @@ def get_best_logdir(self, metric, mode="max"): Args: metric (str): Key for trial info to order on. mode (str): One of [min, max]. - """ df = self.dataframe(metric=metric, mode=mode) if mode == "max": @@ -98,7 +94,7 @@ def fetch_trial_dataframes(self): def get_all_configs(self, prefix=False): """Returns a list of all configurations. - Parameters: + Args: prefix (bool): If True, flattens the config dict and prepends `config/`. """ @@ -120,32 +116,30 @@ def get_all_configs(self, prefix=False): return self._configs def get_trial_checkpoints_paths(self, trial, metric=TRAINING_ITERATION): - """Returns a list of [path, metric] lists for all disk checkpoints of - a trial. + """Gets paths and metrics of all persistent checkpoints of a trial. - Arguments: - trial(Trial): The log directory of a trial, or a trial instance. + Args: + trial (Trial): The log directory of a trial, or a trial instance. metric (str): key for trial info to return, e.g. "mean_accuracy". "training_iteration" is used by default. - """ + Returns: + A list of [path, metric] lists for all persistent checkpoints of + the trial. + """ if isinstance(trial, str): trial_dir = os.path.expanduser(trial) - - # get checkpoints from logdir + # Get checkpoints from logdir. chkpt_df = TrainableUtil.get_checkpoints_paths(trial_dir) - # join with trial dataframe to get metrics + # Join with trial dataframe to get metrics. trial_df = self.trial_dataframes[trial_dir] path_metric_df = chkpt_df.merge( trial_df, on="training_iteration", how="inner") return path_metric_df[["chkpt_path", metric]].values.tolist() elif isinstance(trial, Trial): checkpoints = trial.checkpoint_manager.best_checkpoints() - # TODO(ujvl): Remove condition once the checkpoint manager is - # modified to only track PERSISTENT checkpoints. - return [[c.value, c.result[metric]] for c in checkpoints - if c.storage == Checkpoint.PERSISTENT] + return [[c.value, c.result[metric]] for c in checkpoints] else: raise ValueError("trial should be a string or a Trial instance.") @@ -198,7 +192,8 @@ def __init__(self, experiment_checkpoint_path, trials=None): """Initializer. Args: - experiment_path (str): Path to where experiment is located. + experiment_checkpoint_path (str): Path to where experiment is + located. trials (list|None): List of trials that can be accessed via `analysis.trials`. """ diff --git a/python/ray/tune/trial.py b/python/ray/tune/trial.py --- a/python/ray/tune/trial.py +++ b/python/ray/tune/trial.py @@ -205,10 +205,9 @@ def __init__(self, self.checkpoint_manager = CheckpointManager( keep_checkpoints_num, checkpoint_score_attr, checkpoint_deleter(self._trainable_name(), self.runner)) - checkpoint = Checkpoint(Checkpoint.PERSISTENT, restore_path) - self.checkpoint_manager.newest_persistent_checkpoint = checkpoint # Restoration fields + self.restore_path = restore_path self.restoring_from = None self.num_failures = 0 @@ -243,8 +242,10 @@ def checkpoint(self): if self.status == Trial.PAUSED: assert self.checkpoint_manager.newest_memory_checkpoint.value return self.checkpoint_manager.newest_memory_checkpoint - else: - return self.checkpoint_manager.newest_persistent_checkpoint + checkpoint = self.checkpoint_manager.newest_persistent_checkpoint + if checkpoint.value is None: + checkpoint = Checkpoint(Checkpoint.PERSISTENT, self.restore_path) + return checkpoint @classmethod def generate_id(cls):
diff --git a/python/ray/tune/tests/test_trial_runner_2.py b/python/ray/tune/tests/test_trial_runner_2.py --- a/python/ray/tune/tests/test_trial_runner_2.py +++ b/python/ray/tune/tests/test_trial_runner_2.py @@ -239,7 +239,6 @@ def testRestoreMetricsAfterCheckpointing(self): runner.step() # Start trial self.assertEqual(trials[0].status, Trial.RUNNING) self.assertEqual(ray.get(trials[0].runner.set_info.remote(1)), 1) - # checkpoint = runner.trial_executor.save(trials[0]) runner.step() # Process result, dispatch save runner.step() # Process save runner.trial_executor.stop_trial(trials[0]) diff --git a/python/ray/tune/tests/test_tune_restore.py b/python/ray/tune/tests/test_tune_restore.py --- a/python/ray/tune/tests/test_tune_restore.py +++ b/python/ray/tune/tests/test_tune_restore.py @@ -58,6 +58,22 @@ def testTuneRestore(self): }, ) + def testPostRestoreCheckpointExistence(self): + """Tests that checkpoint restored from is not deleted post-restore.""" + self.assertTrue(os.path.isfile(self.checkpoint_path)) + tune.run( + "PG", + name="TuneRestoreTest", + stop={"training_iteration": 2}, + checkpoint_freq=1, + keep_checkpoints_num=1, + restore=self.checkpoint_path, + config={ + "env": "CartPole-v0", + }, + ) + self.assertTrue(os.path.isfile(self.checkpoint_path)) + class TuneExampleTest(unittest.TestCase): def setUp(self):
"restore" deletes the restored from checkpoint If I specify a checkpoint in the `tune.run(DQNTrainer, restore=X)` - the checkpoint `X` gets deleted! So if there is a bug or something during the initialization, I loose my checkpoint and can no longer restart from it!
When I restore from a checkpoint in the following way, the checkpoint is not deleted for me: `tune.run("DDPG", config=config, restore=checkpoint_path)` I'm on ray v0.8.2 and tensorflow 2.0.0 **EDIT** nvm also have this problem, didn't check/test properly Hm... are you setting check-pointing to be done during this trial? E.g.: ``` tune.run(DQNTrainer, restore=checkpoint, checkpoint_freq=5, checkpoint_at_end=True, checkpoint_score_attr='episode_reward_mean', keep_checkpoints_num=10) ``` Because when I restore it starts a new trial and starts saving checkpoints in a different trial directory (under `~/ray_results`), but for some reason it deletes the last checkpoint from the old directory. cc @ujvl @drozzy thanks for flagging this, will push a fix this weekend.
2020-03-07T23:34:33
ray-project/ray
7,567
ray-project__ray-7567
[ "7310" ]
b38ed4be71d4be096f324c8c666e01e8af5f6c8b
diff --git a/python/ray/state.py b/python/ray/state.py --- a/python/ray/state.py +++ b/python/ray/state.py @@ -318,9 +318,9 @@ def _actor_table(self, actor_id): return {} gcs_entries = gcs_utils.GcsEntry.FromString(message) - assert len(gcs_entries.entries) == 1 + assert len(gcs_entries.entries) > 0 actor_table_data = gcs_utils.ActorTableData.FromString( - gcs_entries.entries[0]) + gcs_entries.entries[-1]) actor_info = { "ActorID": binary_to_hex(actor_table_data.actor_id),
diff --git a/python/ray/tests/test_global_state.py b/python/ray/tests/test_global_state.py --- a/python/ray/tests/test_global_state.py +++ b/python/ray/tests/test_global_state.py @@ -76,6 +76,36 @@ def test_add_remove_cluster_resources(ray_start_cluster_head): assert ray.cluster_resources()["CPU"] == 6 +def test_global_state_actor_table(ray_start_regular): + @ray.remote + class Actor: + def ready(self): + pass + + # actor table should be empty at first + assert len(ray.actors()) == 0 + + # actor table should contain only one entry + a = Actor.remote() + ray.get(a.ready.remote()) + assert len(ray.actors()) == 1 + + # actor table should contain only this entry + # even when the actor goes out of scope + del a + + def get_state(): + return list(ray.actors().values())[0]["State"] + + dead_state = ray.gcs_utils.ActorTableData.DEAD + for _ in range(10): + if get_state() == dead_state: + break + else: + time.sleep(0.5) + assert get_state() == dead_state + + if __name__ == "__main__": import pytest import sys
[dashboard] dashboard.py crashes in example where actors are quickly created and go out of scope. ### What is the problem? I'm using - `ray==0.8.2` - Python 3.7.2 - MacOS ### Reproduction (REQUIRED) #### Bug 1 This is likely somewhat timing dependent, but when I run the following script ```python import ray ray.init() @ray.remote class GrandChild: def __init__(self): pass @ray.remote class Child: def __init__(self): grand_child_handle = GrandChild.remote() @ray.remote class Parent: def __init__(self): children_handles = [Child.remote() for _ in range(2)] parent_handle = Parent.remote() ``` the `dashboard.py` process crashes with the following error. ``` Traceback (most recent call last): File "/Users/rkn/opt/anaconda3/lib/python3.7/threading.py", line 926, in _bootstrap_inner self.run() File "/Users/rkn/opt/anaconda3/lib/python3.7/site-packages/ray/dashboard/dashboard.py", line 546, in run current_actor_table = ray.actors() File "/Users/rkn/opt/anaconda3/lib/python3.7/site-packages/ray/state.py", line 1133, in actors return state.actor_table(actor_id=actor_id) File "/Users/rkn/opt/anaconda3/lib/python3.7/site-packages/ray/state.py", line 369, in actor_table ray.ActorID(actor_id_binary)) File "/Users/rkn/opt/anaconda3/lib/python3.7/site-packages/ray/state.py", line 321, in _actor_table assert len(gcs_entries.entries) == 1 AssertionError ``` #### Bug 2 As a consequence, the dashboard doesn't display anything. There is a second bug here, which is that the error message wasn't pushed to the driver or displayed in the UI in any way. We have a test testing that tests that an error message is pushed to the driver when the monitor process dies. https://github.com/ray-project/ray/blob/007333b96006e090b14a415a8c5993d8b907f02a/python/ray/tests/test_failure.py#L544 I thought we had a similar test for the dashboard, but I don't see it anywhere. Note that we appear to have the plumbing in place to push error messages to the driver when the dashboard process dies, but I didn't see the relevant error message. See https://github.com/ray-project/ray/blob/f2faf8d26e489e0d879fd691abddeb3140c86bee/python/ray/dashboard/dashboard.py#L928-L929 cc @pcmoritz @simon-mo @mitchellstern
2020-03-11T21:36:29
ray-project/ray
7,662
ray-project__ray-7662
[ "7645" ]
90b553ed058a546e036374cd0919e00604892514
diff --git a/rllib/policy/eager_tf_policy.py b/rllib/policy/eager_tf_policy.py --- a/rllib/policy/eager_tf_policy.py +++ b/rllib/policy/eager_tf_policy.py @@ -5,7 +5,6 @@ import logging import functools import numpy as np -import tree from ray.util.debug import log_once from ray.rllib.evaluation.episode import _flatten_action @@ -24,12 +23,12 @@ def _convert_to_tf(x): if isinstance(x, SampleBatch): x = {k: v for k, v in x.items() if k != SampleBatch.INFOS} - return tree.map_structure(_convert_to_tf, x) + return tf.nest.map_structure(_convert_to_tf, x) if isinstance(x, Policy): return x if x is not None: - x = tree.map_structure( + x = tf.nest.map_structure( lambda f: tf.convert_to_tensor(f) if f is not None else None, x) return x @@ -38,7 +37,7 @@ def _convert_to_numpy(x): if x is None: return None try: - return tree.map_structure(lambda component: component.numpy(), x) + return tf.nest.map_structure(lambda component: component.numpy(), x) except AttributeError: raise TypeError( ("Object of type {} has no method to convert to numpy.").format( @@ -66,7 +65,7 @@ def convert_eager_outputs(func): def _func(*args, **kwargs): out = func(*args, **kwargs) if tf.executing_eagerly(): - out = tree.map_structure(_convert_to_numpy, out) + out = tf.nest.map_structure(_convert_to_numpy, out) return out return _func @@ -551,7 +550,7 @@ def _initialize_loss_with_dummy_batch(self): SampleBatch.NEXT_OBS: np.array( [self.observation_space.sample()]), SampleBatch.DONES: np.array([False], dtype=np.bool), - SampleBatch.ACTIONS: tree.map_structure( + SampleBatch.ACTIONS: tf.nest.map_structure( lambda c: np.array([c]), self.action_space.sample()), SampleBatch.REWARDS: np.array([0], dtype=np.float32), } @@ -568,7 +567,8 @@ def _initialize_loss_with_dummy_batch(self): dummy_batch["seq_lens"] = np.array([1], dtype=np.int32) # Convert everything to tensors. - dummy_batch = tree.map_structure(tf.convert_to_tensor, dummy_batch) + dummy_batch = tf.nest.map_structure(tf.convert_to_tensor, + dummy_batch) # for IMPALA which expects a certain sample batch size. def tile_to(tensor, n): @@ -576,7 +576,7 @@ def tile_to(tensor, n): [n] + [1 for _ in tensor.shape.as_list()[1:]]) if get_batch_divisibility_req: - dummy_batch = tree.map_structure( + dummy_batch = tf.nest.map_structure( lambda c: tile_to(c, get_batch_divisibility_req(self)), dummy_batch) @@ -595,7 +595,7 @@ def tile_to(tensor, n): # overwrite any tensor state from that call) self.model.from_batch(dummy_batch) - postprocessed_batch = tree.map_structure( + postprocessed_batch = tf.nest.map_structure( lambda c: tf.convert_to_tensor(c), postprocessed_batch.data) loss_fn(self, self.model, self.dist_class, postprocessed_batch) diff --git a/rllib/policy/tf_policy.py b/rllib/policy/tf_policy.py --- a/rllib/policy/tf_policy.py +++ b/rllib/policy/tf_policy.py @@ -1,7 +1,6 @@ import errno import logging import os -import tree import numpy as np import ray @@ -212,10 +211,8 @@ def _initialize_loss(self, loss, loss_inputs): self._loss = loss self._optimizer = self.optimizer() - self._grads_and_vars = [ - (g, v) for (g, v) in self.gradients(self._optimizer, self._loss) - if g is not None - ] + self._grads_and_vars = [(g, v) for (g, v) in self.gradients( + self._optimizer, self._loss) if g is not None] self._grads = [g for (g, v) in self._grads_and_vars] # TODO(sven/ekl): Deprecate support for v1 models. @@ -493,7 +490,7 @@ def _build_signature_def(self): # build output signatures output_signature = self._extra_output_signature_def() - for i, a in enumerate(tree.flatten(self._sampled_action)): + for i, a in enumerate(tf.nest.flatten(self._sampled_action)): output_signature["actions_{}".format(i)] = \ tf.saved_model.utils.build_tensor_info(a) diff --git a/rllib/utils/torch_ops.py b/rllib/utils/torch_ops.py --- a/rllib/utils/torch_ops.py +++ b/rllib/utils/torch_ops.py @@ -1,9 +1,17 @@ -import tree +import logging from ray.rllib.utils.framework import try_import_torch torch, _ = try_import_torch() +logger = logging.getLogger(__name__) + +try: + import tree +except (ImportError, ModuleNotFoundError) as e: + logger.warning("`dm-tree` is not installed! Run `pip install dm-tree`.") + raise e + def sequence_mask(lengths, maxlen, dtype=None): """ @@ -34,6 +42,7 @@ def convert_to_non_torch_type(stats): dict: A new dict with the same structure as stats_dict, but with all values converted to non-torch Tensor types. """ + # The mapping function used to numpyize torch Tensors. def mapping(item): if isinstance(item, torch.Tensor):
diff --git a/rllib/policy/tests/test_compute_log_likelihoods.py b/rllib/policy/tests/test_compute_log_likelihoods.py --- a/rllib/policy/tests/test_compute_log_likelihoods.py +++ b/rllib/policy/tests/test_compute_log_likelihoods.py @@ -1,5 +1,6 @@ import numpy as np from scipy.stats import norm +from tensorflow.python.eager.context import eager_mode import unittest import ray.rllib.agents.dqn as dqn @@ -43,6 +44,14 @@ def do_test_log_likelihood(run, config["eager"] = fw == "eager" config["use_pytorch"] = fw == "torch" + eager_ctx = None + if fw == "eager": + eager_ctx = eager_mode() + eager_ctx.__enter__() + assert tf.executing_eagerly() + elif fw == "tf": + assert not tf.executing_eagerly() + trainer = run(config=config, env=env) policy = trainer.get_policy() vars = policy.get_weights() @@ -104,6 +113,9 @@ def do_test_log_likelihood(run, prev_reward_batch=np.array([prev_r])) check(np.exp(logp), expected_prob, atol=0.2) + if eager_ctx: + eager_ctx.__exit__(None, None, None) + class TestComputeLogLikelihood(unittest.TestCase): def test_dqn(self):
[rllib] ModuleNotFoundError: No module named 'tree' *Ray version and other system information (Python version, TensorFlow version, OS):* Python3.6 pip install -U ray-0.9.0.dev0-cp36-cp36m-manylinux1_x86_64.whl On 2020-03-18 ### Reproduction (REQUIRED) $ rllib train -f atari-ddppo.yaml Traceback (most recent call last): File "/home/simon/anaconda3/bin/rllib", line 5, in <module> from ray.rllib.scripts import cli File "/home/simon/anaconda3/lib/python3.6/site-packages/ray/rllib/__init__.py", line 9, in <module> from ray.rllib.evaluation.policy_graph import PolicyGraph File "/home/simon/anaconda3/lib/python3.6/site-packages/ray/rllib/evaluation/__init__.py", line 2, in <module> from ray.rllib.evaluation.rollout_worker import RolloutWorker File "/home/simon/anaconda3/lib/python3.6/site-packages/ray/rllib/evaluation/rollout_worker.py", line 19, in <module> from ray.rllib.evaluation.sampler import AsyncSampler, SyncSampler File "/home/simon/anaconda3/lib/python3.6/site-packages/ray/rllib/evaluation/sampler.py", line 11, in <module> from ray.rllib.evaluation.sample_batch_builder import \ File "/home/simon/anaconda3/lib/python3.6/site-packages/ray/rllib/evaluation/sample_batch_builder.py", line 6, in <module> from ray.rllib.policy.sample_batch import SampleBatch, MultiAgentBatch File "/home/simon/anaconda3/lib/python3.6/site-packages/ray/rllib/policy/__init__.py", line 2, in <module> from ray.rllib.policy.torch_policy import TorchPolicy File "/home/simon/anaconda3/lib/python3.6/site-packages/ray/rllib/policy/torch_policy.py", line 10, in <module> from ray.rllib.utils.torch_ops import convert_to_non_torch_type File "/home/simon/anaconda3/lib/python3.6/site-packages/ray/rllib/utils/torch_ops.py", line 1, in <module> import tree ModuleNotFoundError: No module named 'tree'
2020-03-19T19:21:25
ray-project/ray
7,665
ray-project__ray-7665
[ "7663" ]
90b553ed058a546e036374cd0919e00604892514
diff --git a/python/setup.py b/python/setup.py --- a/python/setup.py +++ b/python/setup.py @@ -180,7 +180,6 @@ def find_version(*filepath): "packaging", "pytest", "pyyaml", - "jsonschema", "redis>=3.3.2", # NOTE: Don't upgrade the version of six! Doing so causes installation # problems. See https://github.com/ray-project/ray/issues/4169.
[Python] jsonschema included twice in setup.py requires list. <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? `jsonschema` is included twice in the Python package [setup.py `requires` list](https://github.com/ray-project/ray/blob/master/python/setup.py#L176-L183). This is causing the usage of the Ray Python library within Bazel to fail during the analysis phase due to label duplication in the generated `py_library` target's `'deps'`: ``` ERROR: .../external/requirements_py3_pypi__ray_0_9_0_dev0/BUILD:6:1: Label '@requirements_py3_pypi__jsonschema_3_2_0//:pkg' is duplicated in the 'deps' attribute of rule 'pkg' ``` This bug was introduced in the [cluster json schema validator PR](https://github.com/ray-project/ray/pull/7261/files#diff-8cf6167d58ce775a08acafcfe6f40966). *Ray version and other system information (Python version, TensorFlow version, OS):* Ray master commit 90b553ed058a546e036374cd0919e00604892514 (most recent commit as of this issue filing) ### Reproduction (REQUIRED) - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://ray.readthedocs.io/en/latest/installation.html).
2020-03-19T19:43:14
ray-project/ray
7,705
ray-project__ray-7705
[ "7695" ]
8b4f5a9431fed9454ff116d7370c72bcbb46fbd9
diff --git a/python/ray/tune/logger.py b/python/ray/tune/logger.py --- a/python/ray/tune/logger.py +++ b/python/ray/tune/logger.py @@ -239,9 +239,10 @@ def flush(self): def close(self): if self._file_writer is not None: if self.trial and self.trial.evaluated_params and self.last_result: + flat_result = flatten_dict(self.last_result, delimiter="/") scrubbed_result = { k: value - for k, value in self.last_result.items() + for k, value in flat_result.items() if type(value) in VALID_SUMMARY_TYPES } self._try_log_hparams(scrubbed_result) @@ -249,9 +250,10 @@ def close(self): def _try_log_hparams(self, result): # TBX currently errors if the hparams value is None. + flat_params = flatten_dict(self.trial.evaluated_params) scrubbed_params = { k: v - for k, v in self.trial.evaluated_params.items() if v is not None + for k, v in flat_params.items() if v is not None } from tensorboardX.summary import hparams experiment_tag, session_start_tag, session_end_tag = hparams( diff --git a/python/ray/tune/trainable.py b/python/ray/tune/trainable.py --- a/python/ray/tune/trainable.py +++ b/python/ray/tune/trainable.py @@ -178,7 +178,7 @@ def __init__(self, config=None, logger_creator=None): "slow to initialize, consider setting " "reuse_actors=True to reduce actor creation " "overheads.".format(setup_time)) - self._local_ip = ray.services.get_node_ip_address() + self._local_ip = self.get_current_ip() log_sys_usage = self.config.get("log_sys_usage", False) self._monitor = UtilMonitor(start=log_sys_usage) @@ -213,7 +213,7 @@ def resource_help(cls, config): """ return "" - def current_ip(self): + def get_current_ip(self): logger.info("Getting current IP.") self._local_ip = ray.services.get_node_ip_address() return self._local_ip @@ -419,8 +419,8 @@ def restore(self, checkpoint_path): self._timesteps_since_restore = 0 self._iterations_since_restore = 0 self._restored = True - logger.info("Restored on %s from checkpoint: %s", self.current_ip(), - checkpoint_path) + logger.info("Restored on %s from checkpoint: %s", + self.get_current_ip(), checkpoint_path) state = { "_iteration": self._iteration, "_timesteps_total": self._timesteps_total,
diff --git a/python/ray/tune/tests/test_logger.py b/python/ray/tune/tests/test_logger.py --- a/python/ray/tune/tests/test_logger.py +++ b/python/ray/tune/tests/test_logger.py @@ -28,30 +28,30 @@ def tearDown(self): shutil.rmtree(self.test_dir, ignore_errors=True) def testCSV(self): - config = {"a": 2, "b": 5} + config = {"a": 2, "b": 5, "c": {"c": {"D": 123}, "e": None}} t = Trial(evaluated_params=config, trial_id="csv") logger = CSVLogger(config=config, logdir=self.test_dir, trial=t) logger.on_result(result(2, 4)) logger.on_result(result(2, 4)) - logger.on_result(result(2, 4, score=[1, 2, 3])) + logger.on_result(result(2, 4, score=[1, 2, 3], hello={"world": 1})) logger.close() def testJSON(self): - config = {"a": 2, "b": 5} + config = {"a": 2, "b": 5, "c": {"c": {"D": 123}, "e": None}} t = Trial(evaluated_params=config, trial_id="json") logger = JsonLogger(config=config, logdir=self.test_dir, trial=t) logger.on_result(result(0, 4)) logger.on_result(result(1, 4)) - logger.on_result(result(2, 4, score=[1, 2, 3])) + logger.on_result(result(2, 4, score=[1, 2, 3], hello={"world": 1})) logger.close() def testTBX(self): - config = {"a": 2, "b": 5} + config = {"a": 2, "b": 5, "c": {"c": {"D": 123}, "e": None}} t = Trial(evaluated_params=config, trial_id="tbx") logger = TBXLogger(config=config, logdir=self.test_dir, trial=t) logger.on_result(result(0, 4)) logger.on_result(result(1, 4)) - logger.on_result(result(2, 4, score=[1, 2, 3])) + logger.on_result(result(2, 4, score=[1, 2, 3], hello={"world": 1})) logger.close()
[tune] Reporter fails with nested config keys and {'done': True} ### What is the problem? When a `Trainable` reports `{'done': True}` with some other nested values, the tensorboard logger fails with `NotImplementedError: Got <class 'dict'>, but expected numpy array or torch tensor.`. The same report works for all the preceding logs that don't include `{'done': True}` though. *Ray version and other system information (Python version, TensorFlow version, OS):* `0.8.2` ### Reproduction (REQUIRED) ```python from ray import tune class MyTrainableClass(tune.Trainable): def _setup(self, config): self.timestep = 0 def _train(self): self.timestep += 1 result = {"episode_reward_mean": self.timestep} result['what'] = {'1': '2', '3': 4, '5': {'6': 4}} if self.timestep == 3: result['done'] = True return result tune.run( MyTrainableClass, name="done-test", num_samples=1, config={ 'b': {'c': {'d': '4'}}, 'a': { 'b': tune.sample_from(lambda spec: spec['config']['b']['c']), }, }) ``` Fails with: ``` $ python ./tests/tune_done_test.py 2020-03-22 12:17:21,455 INFO resource_spec.py:212 -- Starting Ray with 10.3 GiB memory available for workers and up to 5.17 GiB for objects. You can adjust these settings with ray.init(memory=<bytes>, object_store_memory=<bytes>). 2020-03-22 12:17:21,746 INFO services.py:1078 -- View the Ray dashboard at localhost:8265 == Status == Memory usage on this node: 17.8/32.0 GiB Using FIFO scheduling algorithm. Resources requested: 1/16 CPUs, 0/0 GPUs, 0.0/10.3 GiB heap, 0.0/3.56 GiB objects Result logdir: /Users/hartikainen/ray_results/done-test Number of trials: 1 (1 RUNNING) +---------------------------+----------+-------+-------+ | Trial name | status | loc | a/b | |---------------------------+----------+-------+-------| | MyTrainableClass_14fda478 | RUNNING | | | +---------------------------+----------+-------+-------+ Result for MyTrainableClass_14fda478: date: 2020-03-22_12-17-23 done: false episode_reward_mean: 1 experiment_id: 3549c7334c884d3abc309232da4f6679 experiment_tag: '0_b={''d'': ''4''}' hostname: catz-48fe.stcatz.ox.ac.uk iterations_since_restore: 1 node_ip: 129.67.48.254 pid: 73820 time_since_restore: 3.0994415283203125e-06 time_this_iter_s: 3.0994415283203125e-06 time_total_s: 3.0994415283203125e-06 timestamp: 1584879443 timesteps_since_restore: 0 training_iteration: 1 trial_id: 14fda478 what: '1': '2' '3': 4 '5': '6': 4 Result for MyTrainableClass_14fda478: date: 2020-03-22_12-17-23 done: true episode_reward_mean: 3 experiment_id: 3549c7334c884d3abc309232da4f6679 experiment_tag: '0_b={''d'': ''4''}' hostname: catz-48fe.stcatz.ox.ac.uk iterations_since_restore: 3 node_ip: 129.67.48.254 pid: 73820 time_since_restore: 9.775161743164062e-06 time_this_iter_s: 2.86102294921875e-06 time_total_s: 9.775161743164062e-06 timestamp: 1584879443 timesteps_since_restore: 0 training_iteration: 3 trial_id: 14fda478 what: '1': '2' '3': 4 '5': '6': 4 2020-03-22 12:17:23,204 ERROR trial_runner.py:513 -- Trial MyTrainableClass_14fda478: Error processing event. Traceback (most recent call last): File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/trial_runner.py", line 511, in _process_trial self._execute_action(trial, decision) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/trial_runner.py", line 595, in _execute_action self.trial_executor.stop_trial(trial) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/ray_trial_executor.py", line 263, in stop_trial trial, error=error, error_msg=error_msg, stop_logger=stop_logger) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/ray_trial_executor.py", line 204, in _stop_trial trial.close_logger() File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/trial.py", line 315, in close_logger self.result_logger.close() File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/logger.py", line 305, in close _logger.close() File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/logger.py", line 233, in close self._try_log_hparams(self.last_result) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/logger.py", line 244, in _try_log_hparams hparam_dict=scrubbed_params, metric_dict=result) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/tensorboardX/summary.py", line 102, in hparams v = make_np(v)[0] File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/tensorboardX/x2num.py", line 34, in make_np 'Got {}, but expected numpy array or torch tensor.'.format(type(x))) NotImplementedError: Got <class 'dict'>, but expected numpy array or torch tensor. Traceback (most recent call last): File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/trial_runner.py", line 511, in _process_trial self._execute_action(trial, decision) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/trial_runner.py", line 595, in _execute_action self.trial_executor.stop_trial(trial) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/ray_trial_executor.py", line 263, in stop_trial trial, error=error, error_msg=error_msg, stop_logger=stop_logger) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/ray_trial_executor.py", line 204, in _stop_trial trial.close_logger() File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/trial.py", line 315, in close_logger self.result_logger.close() File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/logger.py", line 305, in close _logger.close() File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/logger.py", line 233, in close self._try_log_hparams(self.last_result) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/logger.py", line 244, in _try_log_hparams hparam_dict=scrubbed_params, metric_dict=result) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/tensorboardX/summary.py", line 102, in hparams v = make_np(v)[0] File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/tensorboardX/x2num.py", line 34, in make_np 'Got {}, but expected numpy array or torch tensor.'.format(type(x))) NotImplementedError: Got <class 'dict'>, but expected numpy array or torch tensor. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "./tests/tune_done_test.py", line 24, in <module> 'b': tune.sample_from(lambda spec: spec['config']['b']['c']), File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/tune.py", line 324, in run runner.step() File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/trial_runner.py", line 335, in step self._process_events() # blocking File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/trial_runner.py", line 444, in _process_events self._process_trial(trial) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/trial_runner.py", line 514, in _process_trial self._process_trial_failure(trial, traceback.format_exc()) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/trial_runner.py", line 580, in _process_trial_failure trial, error=True, error_msg=error_msg) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/ray_trial_executor.py", line 263, in stop_trial trial, error=error, error_msg=error_msg, stop_logger=stop_logger) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/ray_trial_executor.py", line 204, in _stop_trial trial.close_logger() File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/trial.py", line 315, in close_logger self.result_logger.close() File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/logger.py", line 305, in close _logger.close() File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/logger.py", line 233, in close self._try_log_hparams(self.last_result) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/ray/tune/logger.py", line 244, in _try_log_hparams hparam_dict=scrubbed_params, metric_dict=result) File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/tensorboardX/summary.py", line 102, in hparams v = make_np(v)[0] File "/Users/hartikainen/conda/envs/softlearning-tf2/lib/python3.7/site-packages/tensorboardX/x2num.py", line 34, in make_np 'Got {}, but expected numpy array or torch tensor.'.format(type(x))) NotImplementedError: Got <class 'dict'>, but expected numpy array or torch tensor. ``` - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://ray.readthedocs.io/en/latest/installation.html).
Actually, this seems to be related to the nested config. If the above script is run with: ``` 'b': tune.sample_from(lambda spec: spec['config']['b']['c']['d']) ``` instead of ``` 'b': tune.sample_from(lambda spec: spec['config']['b']['c']), ``` it finishes as expected.
2020-03-23T02:21:38
ray-project/ray
7,752
ray-project__ray-7752
[ "7751" ]
ca6eabc9cb728a829d5305a4dd19216ed925f2e4
diff --git a/python/ray/worker.py b/python/ray/worker.py --- a/python/ray/worker.py +++ b/python/ray/worker.py @@ -1391,6 +1391,9 @@ def register_custom_serializer(cls, use_dict: Deprecated. class_id (str): Unique ID of the class. Autogenerated if None. """ + worker = global_worker + worker.check_connected() + if use_pickle: raise DeprecationWarning( "`use_pickle` is no longer a valid parameter and will be removed "
AttributeError: 'Worker' object has no attribute 'lock' the following code ```python import ray # ray.init() class Foo(object): def __init__(self, value): self.value = value def custom_serializer(obj): return obj.value def custom_deserializer(value): object = Foo(value) return object ray.register_custom_serializer( Foo, serializer=custom_serializer, deserializer=custom_deserializer) object_id = ray.put(Foo(100)) ``` trigger an exception when calling `ray.register_custom_serializer`, ``` Traceback (most recent call last): File "ray_err.py", line 20, in <module> Foo, serializer=custom_serializer, deserializer=custom_deserializer) File "/home/cloudhan/ray/python/ray/worker.py", line 1405, in register_custom_serializer context = global_worker.get_serialization_context() File "/home/cloudhan/ray/python/ray/worker.py", line 198, in get_serialization_context with self.lock: AttributeError: 'Worker' object has no attribute 'lock' ``` inspecting into the code you will find the lock is missing, https://github.com/ray-project/ray/blob/ca6eabc9cb728a829d5305a4dd19216ed925f2e4/python/ray/worker.py#L182-L202 the reason is simply not initialize ray, but the error is confusing.
2020-03-26T10:10:14
ray-project/ray
7,758
ray-project__ray-7758
[ "7757" ]
e196fcdbaf1a226e8f5ef5f28308f13d9be8261d
diff --git a/rllib/rollout.py b/rllib/rollout.py --- a/rllib/rollout.py +++ b/rllib/rollout.py @@ -11,7 +11,6 @@ import gym import ray -from ray.rllib.agents.registry import get_agent_class from ray.rllib.env import MultiAgentEnv from ray.rllib.env.base_env import _DUMMY_AGENT_ID from ray.rllib.evaluation.episode import _flatten_action @@ -19,6 +18,7 @@ from ray.rllib.policy.sample_batch import DEFAULT_POLICY_ID from ray.rllib.utils.deprecation import deprecation_warning from ray.tune.utils import merge_dicts +from ray.tune.registry import get_trainable_cls EXAMPLE_USAGE = """ Example Usage via RLlib CLI: @@ -274,7 +274,7 @@ def run(args, parser): ray.init() # Create the Trainer from config. - cls = get_agent_class(args.run) + cls = get_trainable_cls(args.run) agent = cls(env=args.env, config=config) # Load state from checkpoint. agent.restore(args.checkpoint)
[rllib] ray/rllib/rollout.py does not find registered trainables ### What is the problem? When using rollout.py with custom trainable (registered via `tune.register_trainable`) the script does not find the trainable. This seems to be caused because rollout.py uses `ray.rllib.agents.registry.get_agent_class` instead of `ray.tune.registry.get_trainable_cls`. *Ray version and other system information (Python version, TensorFlow version, OS):* python=3.7.3, ray[rllib,debug]==ray-0.9.0.dev0, tensorflow==1.15.0, Ubuntu 18.04 LTS ### Reproduction ``` #generate checkpoint rllib train --run DQN --env CartPole-v0 --stop '{"timesteps_total": 5000}' --checkpoint-freq 1 python rollout.py PATH_TO_CHECKPOINT --run OtherDQN --episodes 10 --out rollout.pkl 2020-03-26 16:28:25,858 INFO resource_spec.py:212 -- Starting Ray with 11.62 GiB memory available for workers and up to 5.83 GiB for objects. You can adjust these settings with ray.init(memory=<bytes>, object_store_memory=<bytes>). 2020-03-26 16:28:26,332 INFO services.py:1123 -- View the Ray dashboard at localhost:8265 Traceback (most recent call last): File "rollout.py", line 475, in <module> run(args, parser) File "rollout.py", line 285, in run cls = get_agent_class(args.run) File "/home/carl/miniconda3/envs/rollout_test/lib/python3.7/site-packages/ray/rllib/agents/registry.py", line 130, in get_agent_class return _get_agent_class(alg) File "/home/carl/miniconda3/envs/rollout_test/lib/python3.7/site-packages/ray/rllib/agents/registry.py", line 154, in _get_agent_class raise Exception(("Unknown algorithm {}.").format(alg)) Exception: Unknown algorithm OtherDQN. ``` The provided rollout.py adds the following lines at the start: ``` from ray.rllib.agents.dqn import DQNTrainer OtherDQNTrainer = DQNTrainer.with_updates( name="OtherDQN") ray.tune.register_trainable("OtherDQN", OtherDQNTrainer) ``` If `ray.rllib.agents.registry.get_agent_class` is replaces with `ray.tune.registry.get_trainable_cls` it works (with both "OtherDQN" and "DQN") [rollout.zip](https://github.com/ray-project/ray/files/4388161/rollout.zip) - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://ray.readthedocs.io/en/latest/installation.html).
2020-03-26T15:58:21
ray-project/ray
7,846
ray-project__ray-7846
[ "7843" ]
1b477c6dd04bb6adf1b910212e5578e18f774bf6
diff --git a/rllib/agents/ppo/ppo.py b/rllib/agents/ppo/ppo.py --- a/rllib/agents/ppo/ppo.py +++ b/rllib/agents/ppo/ppo.py @@ -67,6 +67,9 @@ # usually slower, but you might want to try it if you run into issues with # the default optimizer. "simple_optimizer": False, + # Whether to fake GPUs (using CPUs). + # Set this to True for debugging on non-GPU machines (set `num_gpus` > 0). + "_fake_gpus": False, # Use PyTorch as framework? "use_pytorch": False }) @@ -92,7 +95,8 @@ def choose_policy_optimizer(workers, config): num_envs_per_worker=config["num_envs_per_worker"], train_batch_size=config["train_batch_size"], standardize_fields=["advantages"], - shuffle_sequences=config["shuffle_sequences"]) + shuffle_sequences=config["shuffle_sequences"], + _fake_gpus=config["_fake_gpus"]) def update_kl(trainer, fetches): diff --git a/rllib/optimizers/multi_gpu_optimizer.py b/rllib/optimizers/multi_gpu_optimizer.py --- a/rllib/optimizers/multi_gpu_optimizer.py +++ b/rllib/optimizers/multi_gpu_optimizer.py @@ -29,8 +29,9 @@ class LocalMultiGPUOptimizer(PolicyOptimizer): A number of SGD passes are then taken over the in-memory data. For more details, see `multi_gpu_impl.LocalSyncParallelOptimizer`. - This optimizer is Tensorflow-specific and require the underlying - Policy to be a TFPolicy instance that support `.copy()`. + This optimizer is Tensorflow-specific and requires the underlying + Policy to be a TFPolicy instance that implements the `copy()` method + for multi-GPU tower generation. Note that all replicas of the TFPolicy will merge their extra_compute_grad and apply_grad feed_dicts and fetches. This @@ -46,7 +47,8 @@ def __init__(self, train_batch_size=1024, num_gpus=0, standardize_fields=[], - shuffle_sequences=True): + shuffle_sequences=True, + _fake_gpus=False): """Initialize a synchronous multi-gpu optimizer. Arguments: @@ -62,6 +64,9 @@ def __init__(self, to normalize shuffle_sequences (bool): whether to shuffle the train batch prior to SGD to break up correlations + _fake_gpus (bool): Whether to use fake-GPUs (CPUs) instead of + actual GPUs (should only be used for testing on non-GPU + machines). """ PolicyOptimizer.__init__(self, workers) @@ -71,12 +76,16 @@ def __init__(self, self.rollout_fragment_length = rollout_fragment_length self.train_batch_size = train_batch_size self.shuffle_sequences = shuffle_sequences + + # Collect actual devices to use. if not num_gpus: - self.devices = ["/cpu:0"] - else: - self.devices = [ - "/gpu:{}".format(i) for i in range(int(math.ceil(num_gpus))) - ] + _fake_gpus = True + num_gpus = 1 + type_ = "cpu" if _fake_gpus else "gpu" + self.devices = [ + "/{}:{}".format(type_, i) for i in range(int(math.ceil(num_gpus))) + ] + self.batch_size = int(sgd_batch_size / len(self.devices)) * len( self.devices) assert self.batch_size % len(self.devices) == 0 diff --git a/rllib/policy/dynamic_tf_policy.py b/rllib/policy/dynamic_tf_policy.py --- a/rllib/policy/dynamic_tf_policy.py +++ b/rllib/policy/dynamic_tf_policy.py @@ -112,6 +112,8 @@ def __init__(self, prev_actions = existing_inputs[SampleBatch.PREV_ACTIONS] prev_rewards = existing_inputs[SampleBatch.PREV_REWARDS] action_input = existing_inputs[SampleBatch.ACTIONS] + explore = existing_inputs["is_exploring"] + timestep = existing_inputs["timestep"] else: obs = tf.placeholder( tf.float32, @@ -123,8 +125,9 @@ def __init__(self, action_space, "prev_action") prev_rewards = tf.placeholder( tf.float32, [None], name="prev_reward") - - explore = tf.placeholder_with_default(False, (), name="is_exploring") + explore = tf.placeholder_with_default( + True, (), name="is_exploring") + timestep = tf.placeholder(tf.int32, (), name="timestep") self._input_dict = { SampleBatch.CUR_OBS: obs, @@ -175,8 +178,6 @@ def __init__(self, for s in self.model.get_initial_state() ] - timestep = tf.placeholder(tf.int32, (), name="timestep") - # Fully customized action generation (e.g., custom policy). if action_sampler_fn: sampled_action, sampled_action_logp = action_sampler_fn( @@ -281,8 +282,9 @@ def copy(self, existing_inputs): existing_inputs[len(self._loss_inputs) + i])) if rnn_inputs: rnn_inputs.append(("seq_lens", existing_inputs[-1])) - input_dict = OrderedDict([(k, existing_inputs[i]) for i, ( - k, _) in enumerate(self._loss_inputs)] + rnn_inputs) + input_dict = OrderedDict([("is_exploring", self._is_exploring), ( + "timestep", self._timestep)] + [(k, existing_inputs[i]) for i, ( + k, _) in enumerate(self._loss_inputs)] + rnn_inputs) instance = self.__class__( self.observation_space, self.action_space,
diff --git a/rllib/agents/ppo/tests/test_ppo.py b/rllib/agents/ppo/tests/test_ppo.py --- a/rllib/agents/ppo/tests/test_ppo.py +++ b/rllib/agents/ppo/tests/test_ppo.py @@ -46,6 +46,33 @@ def test_ppo_compilation(self): for i in range(num_iterations): trainer.train() + def test_ppo_fake_multi_gpu_learning(self): + """Test whether PPOTrainer can learn CartPole w/ faked multi-GPU.""" + config = ppo.DEFAULT_CONFIG.copy() + # Fake GPU setup. + config["num_gpus"] = 2 + config["_fake_gpus"] = True + # Mimick tuned_example for PPO CartPole. + config["num_workers"] = 1 + config["lr"] = 0.0003 + config["observation_filter"] = "MeanStdFilter" + config["num_sgd_iter"] = 6 + config["vf_share_layers"] = True + config["vf_loss_coeff"] = 0.01 + config["model"]["fcnet_hiddens"] = [32] + config["model"]["fcnet_activation"] = "linear" + + trainer = ppo.PPOTrainer(config=config, env="CartPole-v0") + num_iterations = 200 + learnt = False + for i in range(num_iterations): + results = trainer.train() + if results["episode_reward_mean"] > 150: + learnt = True + break + print(results) + assert learnt, "PPO multi-GPU (with fake-GPUs) did not learn CartPole!" + def test_ppo_exploration_setup(self): """Tests, whether PPO runs with different exploration setups.""" config = ppo.DEFAULT_CONFIG.copy() diff --git a/rllib/tuned_examples/regression_tests/cartpole-ppo-tf-multi-gpu.yaml b/rllib/tuned_examples/regression_tests/cartpole-ppo-tf-multi-gpu.yaml new file mode 100644 --- /dev/null +++ b/rllib/tuned_examples/regression_tests/cartpole-ppo-tf-multi-gpu.yaml @@ -0,0 +1,20 @@ +cartpole-ppo-tf-multi-gpu: + env: CartPole-v0 + run: PPO + stop: + episode_reward_mean: 150 + timesteps_total: 100000 + config: + gamma: 0.99 + lr: 0.0003 + num_workers: 1 + observation_filter: MeanStdFilter + num_sgd_iter: 6 + vf_share_layers: true + vf_loss_coeff: 0.01 + model: + fcnet_hiddens: [32] + fcnet_activation: linear + # Use fake-GPU setup to prove towers are working and learning. + num_gpus: 6 + _fake_gpus: true
[rllib] PPO policy return all zeros for action_logp <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? *Ray version and other system information (Python version, TensorFlow version, OS):* PPO policy return all zeros when computing the log probability of actions. ray=0.8.2 (latest) tensorflow=2.1.0 ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): ```python import ray import numpy as np from ray.rllib.agents.ppo import PPOTrainer ray.init() ppo_agent = PPOTrainer(env="BipedalWalker-v2") ppo_agent.get_policy().get_session().run( ppo_agent.get_policy()._action_logp, feed_dict={ ppo_agent.get_policy()._input_dict["obs"]: np.random.random((500, 24)) } ) ``` Result in ``` array([0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0., 0.], dtype=float32) ``` If we cannot run your script, we cannot fix your issue. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://ray.readthedocs.io/en/latest/installation.html).
The issue still happen in Ray=0.8.3 ```python import ray import numpy as np from ray.rllib.agents.ppo import PPOTrainer print(ray.__version__) ray.init(ignore_reinit_error=True) ppo_agent = PPOTrainer(env="BipedalWalker-v2") ppo_agent.get_policy().get_session().run( ppo_agent.get_policy()._sampled_action_logp, feed_dict={ ppo_agent.get_policy()._input_dict["obs"]: np.random.random((500, 24)) } ) ``` Ray 0.8.1 works well! Something wrong with the updates introduced by 0.8.2 I did a bit of digging and it seems related to the is_exploring flag somehow being False for PPO in multi-GPU mode (works fine with simple_optimizer: True). @sven1977 I think this is since the explore placeholder isn't passed to copies in `dynamic_tf_policy.copy()`. Yeah, that's possible. I'll take a look.
2020-04-01T06:41:05
ray-project/ray
7,851
ray-project__ray-7851
[ "7850" ]
e153e3179f54819e06c07df21bbf49e260dec5f2
diff --git a/rllib/agents/ddpg/ddpg_policy.py b/rllib/agents/ddpg/ddpg_policy.py --- a/rllib/agents/ddpg/ddpg_policy.py +++ b/rllib/agents/ddpg/ddpg_policy.py @@ -54,6 +54,7 @@ def postprocess_trajectory(self, feed_dict={ self.cur_observations: states, self._is_exploring: False, + self._timestep: self.global_timestep, }) distance_in_action_space = np.sqrt( np.mean(np.square(clean_actions - noisy_actions))) @@ -414,16 +415,10 @@ def _build_policy_network(self, obs, obs_space, action_space): activation = getattr(tf.nn, self.config["actor_hidden_activation"]) for hidden in self.config["actor_hiddens"]: + action_out = tf.layers.dense( + action_out, units=hidden, activation=activation) if self.config["parameter_noise"]: - import tensorflow.contrib.layers as layers - action_out = layers.fully_connected( - action_out, - num_outputs=hidden, - activation_fn=activation, - normalizer_fn=layers.layer_norm) - else: - action_out = tf.layers.dense( - action_out, units=hidden, activation=activation) + action_out = tf.keras.layers.LayerNormalization()(action_out) action_out = tf.layers.dense( action_out, units=action_space.shape[0], activation=None) diff --git a/rllib/agents/dqn/distributional_q_model.py b/rllib/agents/dqn/distributional_q_model.py --- a/rllib/agents/dqn/distributional_q_model.py +++ b/rllib/agents/dqn/distributional_q_model.py @@ -125,14 +125,12 @@ def build_state_score(model_out): state_out = self._noisy_layer("dueling_hidden_%d" % i, state_out, q_hiddens[i], sigma0) - elif parameter_noise: - state_out = tf.keras.layers.Dense( - units=q_hiddens[i], - activation_fn=tf.nn.relu, - normalizer_fn=tf.contrib.layers.layer_norm)(state_out) else: state_out = tf.keras.layers.Dense( units=q_hiddens[i], activation=tf.nn.relu)(state_out) + if parameter_noise: + state_out = tf.keras.layers.LayerNormalization()( + state_out) if use_noisy: state_score = self._noisy_layer( "dueling_output",
[rllib] ModuleNotFoundError: No module named 'tensorflow.contrib' <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? ``` File "/home/axion/anaconda3/envs/training/lib/python3.7/site-packages/ray/rllib/agents/trainer_template.py", line 93, in __init__ Trainer.__init__(self, config, env, logger_creator) File "/home/axion/anaconda3/envs/training/lib/python3.7/site-packages/ray/rllib/agents/trainer.py", line 455, in __init__ super().__init__(config, logger_creator) File "/home/axion/anaconda3/envs/training/lib/python3.7/site-packages/ray/tune/trainable.py", line 174, in __init__ self._setup(copy.deepcopy(self.config)) File "/home/axion/anaconda3/envs/training/lib/python3.7/site-packages/ray/rllib/agents/trainer.py", line 596, in _setup self._init(self.config, self.env_creator) File "/home/axion/anaconda3/envs/training/lib/python3.7/site-packages/ray/rllib/agents/trainer_template.py", line 117, in _init config) File "/home/axion/anaconda3/envs/training/lib/python3.7/site-packages/ray/rllib/agents/dqn/apex.py", line 48, in defer_make_workers return trainer._make_workers(env_creator, policy, config, 0) File "/home/axion/anaconda3/envs/training/lib/python3.7/site-packages/ray/rllib/agents/trainer.py", line 666, in _make_workers logdir=self.logdir) File "/home/axion/anaconda3/envs/training/lib/python3.7/site-packages/ray/rllib/evaluation/worker_set.py", line 61, in __init__ RolloutWorker, env_creator, policy, 0, self._local_config) File "/home/axion/anaconda3/envs/training/lib/python3.7/site-packages/ray/rllib/evaluation/worker_set.py", line 271, in _make_worker _fake_sampler=config.get("_fake_sampler", False)) File "/home/axion/anaconda3/envs/training/lib/python3.7/site-packages/ray/rllib/evaluation/rollout_worker.py", line 360, in __init__ self._build_policy_map(policy_dict, policy_config) File "/home/axion/anaconda3/envs/training/lib/python3.7/site-packages/ray/rllib/evaluation/rollout_worker.py", line 836, in _build_policy_map policy_map[name] = cls(obs_space, act_space, merged_conf) File "/home/axion/anaconda3/envs/training/lib/python3.7/site-packages/ray/rllib/agents/ddpg/ddpg_policy.py", line 110, in __init__ self.cur_observations, observation_space, action_space) File "/home/axion/anaconda3/envs/training/lib/python3.7/site-packages/ray/rllib/agents/ddpg/ddpg_policy.py", line 410, in _build_policy_network import tensorflow.contrib.layers as layers ModuleNotFoundError: No module named 'tensorflow.contrib' ``` *Ray version and other system information (Python version, TensorFlow version, OS):* Python 3.7.7 I am running Ray 0.8.3 but this bug is currently still in the master branch tensorflow==2.1.0 ### Reproduction (REQUIRED) Setting any DDPG algorithm to `parameter_noise=True` forces this import to run https://github.com/ray-project/ray/blob/master/rllib/agents/ddpg/ddpg_policy.py#L417 **Tensorflow 2 does not have a `contrib` module** - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://ray.readthedocs.io/en/latest/installation.html).
Yeah, we should switch to keras here. I'm taking a look. For now, could you replace the code (~ line 416) in rllib/agents/ddpg/ddpg_policy.py with the following? That should eliminate the dependency on the deprecated tf.contrib. ``` activation = getattr(tf.nn, self.config["actor_hidden_activation"]) for hidden in self.config["actor_hiddens"]: if self.config["parameter_noise"]: action_out = tf.keras.layers.Dense( units=hidden, activation=activation)(action_out) action_out = tf.keras.layers.LayerNormalization()(action_out) else: action_out = tf.keras.layers.Dense( units=hidden, activation=activation)(action_out) action_out = tf.keras.layers.Dense( units=action_space.shape[0], activation=None)(action_out) ``` Cool. I will give this a try! Preliminary tests look ok. Please let me know, if this doesn't help.
2020-04-01T09:56:22
ray-project/ray
7,863
ray-project__ray-7863
[ "7891" ]
c2cb5c2214f5427415f3fff770840cce6881aa28
diff --git a/python/ray/worker.py b/python/ray/worker.py --- a/python/ray/worker.py +++ b/python/ray/worker.py @@ -1349,8 +1349,10 @@ def disconnect(exiting_interpreter=False): @contextmanager def _changeproctitle(title, next_title): setproctitle.setproctitle(title) - yield - setproctitle.setproctitle(next_title) + try: + yield + finally: + setproctitle.setproctitle(next_title) def register_custom_serializer(cls,
Incorrect Title for Failed Tasks <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? When a task fails, the process title of the failed task remains. This appears in both the dashboard and to the OS (`htop`). *Ray version and other system information (Python version, TensorFlow version, OS):* Python: 3.7 OS: MacOS 10.15.4 Ray: Latest Master ### Reproduction (REQUIRED) ``` import ray ray.init() @ray.remote def fail(): import NonRealModule fail.remote() ``` Then run `htop` and "`__main__.fail()`" will remain even after the function errors and returns. If we cannot run your script, we cannot fix your issue. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://ray.readthedocs.io/en/latest/installation.html).
2020-04-02T00:24:35
ray-project/ray
7,877
ray-project__ray-7877
[ "7878" ]
7f9ddfcfd869b16b49e47f2d48c027d13dd45922
diff --git a/python/setup.py b/python/setup.py --- a/python/setup.py +++ b/python/setup.py @@ -191,7 +191,7 @@ def find_version(*filepath): # The BinaryDistribution argument triggers build_ext. distclass=BinaryDistribution, install_requires=requires, - setup_requires=["cython >= 0.29.14"], + setup_requires=["cython >= 0.29.14", "wheel"], extras_require=extras, entry_points={ "console_scripts": [
[ray] Build failure when installing from source <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? I am trying to install ray from source. However, I receive the following error during the install process: ``` HEAD is now at 8ffe41c apply cpython patch bpo-39492 for the reference count issue + /usr/bin/python setup.py bdist_wheel usage: setup.py [global_opts] cmd1 [cmd1_opts] [cmd2 [cmd2_opts] ...] or: setup.py --help [cmd1 cmd2 ...] or: setup.py --help-commands or: setup.py cmd --help error: invalid command 'bdist_wheel' Traceback (most recent call last): File "setup.py", line 178, in <module> setup( File "/usr/lib/python3.8/site-packages/setuptools/__init__.py", line 144, in setup return distutils.core.setup(**attrs) File "/usr/lib/python3.8/distutils/core.py", line 148, in setup dist.run_commands() File "/usr/lib/python3.8/distutils/dist.py", line 966, in run_commands self.run_command(cmd) File "/usr/lib/python3.8/distutils/dist.py", line 985, in run_command cmd_obj.run() File "/usr/lib/python3.8/distutils/command/build.py", line 135, in run self.run_command(cmd_name) File "/usr/lib/python3.8/distutils/cmd.py", line 313, in run_command self.distribution.run_command(command) File "/usr/lib/python3.8/distutils/dist.py", line 985, in run_command cmd_obj.run() File "setup.py", line 107, in run subprocess.check_call(command) File "/usr/lib/python3.8/subprocess.py", line 364, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['../build.sh', '-p', '/usr/bin/python']' returned non-zero exit status 1. ``` Note: `error invalid command 'bdist_wheel'` *Ray version and other system information (Python version, TensorFlow version, OS):* ray: 0.8.4 python: 3.8.2 os: Arch Linux ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): If we cannot run your script, we cannot fix your issue. - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://ray.readthedocs.io/en/latest/installation.html). The above does not apply since this is a build issue.
2020-04-02T21:56:48
ray-project/ray
7,982
ray-project__ray-7982
[ "7981" ]
dd63178e91dbe77b79be8d38bdf67501ceee1d7c
diff --git a/python/ray/tune/progress_reporter.py b/python/ray/tune/progress_reporter.py --- a/python/ray/tune/progress_reporter.py +++ b/python/ray/tune/progress_reporter.py @@ -7,6 +7,11 @@ TRAINING_ITERATION, TIME_TOTAL_S, TIMESTEPS_TOTAL) from ray.tune.utils import flatten_dict +try: + from collections.abc import Mapping +except ImportError: + from collections import Mapping + try: from tabulate import tabulate except ImportError: @@ -101,7 +106,7 @@ def add_metric_column(self, metric, representation=None): if metric in self._metric_columns: raise ValueError("Column {} already exists.".format(metric)) - if isinstance(self._metric_columns, collections.Mapping): + if isinstance(self._metric_columns, Mapping): representation = representation or metric self._metric_columns[metric] = representation else: @@ -291,7 +296,7 @@ def trial_progress_str(trials, metric_columns, fmt="psql", max_rows=None): num_trials, ", ".join(num_trials_strs))) # Pre-process trials to figure out what columns to show. - if isinstance(metric_columns, collections.Mapping): + if isinstance(metric_columns, Mapping): keys = list(metric_columns.keys()) else: keys = metric_columns @@ -303,7 +308,7 @@ def trial_progress_str(trials, metric_columns, fmt="psql", max_rows=None): params = sorted(set().union(*[t.evaluated_params for t in trials])) trial_table = [_get_trial_info(trial, params, keys) for trial in trials] # Format column headings - if isinstance(metric_columns, collections.Mapping): + if isinstance(metric_columns, Mapping): formatted_columns = [metric_columns[k] for k in keys] else: formatted_columns = keys diff --git a/python/ray/util/sgd/torch/torch_runner.py b/python/ray/util/sgd/torch/torch_runner.py --- a/python/ray/util/sgd/torch/torch_runner.py +++ b/python/ray/util/sgd/torch/torch_runner.py @@ -1,4 +1,3 @@ -import collections from filelock import FileLock import logging import inspect @@ -17,6 +16,11 @@ logger = logging.getLogger(__name__) amp = None +try: + from collections.abc import Iterable +except ImportError: + from collections import Iterable + try: from apex import amp except ImportError: @@ -133,7 +137,7 @@ def _create_schedulers_if_available(self): self.schedulers = self.scheduler_creator(self.given_optimizers, self.config) - if not isinstance(self.schedulers, collections.Iterable): + if not isinstance(self.schedulers, Iterable): self.schedulers = [self.schedulers] def _try_setup_apex(self): @@ -153,7 +157,7 @@ def setup_components(self): self._initialize_dataloaders() logger.debug("Creating model") self.models = self.model_creator(self.config) - if not isinstance(self.models, collections.Iterable): + if not isinstance(self.models, Iterable): self.models = [self.models] assert all(isinstance(model, nn.Module) for model in self.models), ( "All models must be PyTorch models: {}.".format(self.models)) @@ -163,7 +167,7 @@ def setup_components(self): logger.debug("Creating optimizer.") self.optimizers = self.optimizer_creator(self.given_models, self.config) - if not isinstance(self.optimizers, collections.Iterable): + if not isinstance(self.optimizers, Iterable): self.optimizers = [self.optimizers] self._create_schedulers_if_available() diff --git a/python/ray/util/sgd/torch/training_operator.py b/python/ray/util/sgd/torch/training_operator.py --- a/python/ray/util/sgd/torch/training_operator.py +++ b/python/ray/util/sgd/torch/training_operator.py @@ -1,4 +1,3 @@ -import collections import torch from ray.util.sgd.utils import (TimerCollection, AverageMeterCollection, @@ -8,6 +7,11 @@ amp = None +try: + from collections.abc import Iterable +except ImportError: + from collections import Iterable + try: from apex import amp except ImportError: @@ -25,7 +29,7 @@ def _is_multiple(component): """Checks if a component (optimizer, model, etc) is not singular.""" - return isinstance(component, collections.Iterable) and len(component) > 1 + return isinstance(component, Iterable) and len(component) > 1 class TrainingOperator: @@ -66,19 +70,24 @@ def __init__(self, use_tqdm=False): # You are not expected to override this method. self._models = models # List of models - assert isinstance(models, collections.Iterable), ( - "Components need to be iterable. Got: {}".format(type(models))) + assert isinstance( + models, + Iterable), ("Components need to be iterable. Got: {}".format( + type(models))) self._optimizers = optimizers # List of optimizers - assert isinstance(optimizers, collections.Iterable), ( - "Components need to be iterable. Got: {}".format(type(optimizers))) + assert isinstance( + optimizers, + Iterable), ("Components need to be iterable. Got: {}".format( + type(optimizers))) self._train_loader = train_loader self._validation_loader = validation_loader self._world_rank = world_rank self._criterion = criterion self._schedulers = schedulers if schedulers: - assert isinstance(schedulers, collections.Iterable), ( - "Components need to be iterable. Got: {}".format( + assert isinstance( + schedulers, + Iterable), ("Components need to be iterable. Got: {}".format( type(schedulers))) self._config = config self._use_fp16 = use_fp16
[python] Importing ABC directly from collections was deprecated and will be removed in Python 3.10 ### What is the problem? Importing ABC directly from collections was deprecated and will be removed in Python 3.10. Use collections.abc instead. ### Reproduction (REQUIRED) ``` python/ray/tune/progress_reporter.py 104: if isinstance(self._metric_columns, collections.Mapping): 294: if isinstance(metric_columns, collections.Mapping): 306: if isinstance(metric_columns, collections.Mapping): python/ray/util/sgd/torch/training_operator.py 28: return isinstance(component, collections.Iterable) and len(component) > 1 69: assert isinstance(models, collections.Iterable), ( 72: assert isinstance(optimizers, collections.Iterable), ( 80: assert isinstance(schedulers, collections.Iterable), ( python/ray/util/sgd/torch/torch_runner.py 136: if not isinstance(self.schedulers, collections.Iterable): 156: if not isinstance(self.models, collections.Iterable): 166: if not isinstance(self.optimizers, collections.Iterable): ``` - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://ray.readthedocs.io/en/latest/installation.html).
2020-04-11T16:27:32
ray-project/ray
8,177
ray-project__ray-8177
[ "8176" ]
9dc625318f51f7d2ccaeb44ca331d94bc3b2a2f7
diff --git a/python/ray/experimental/async_api.py b/python/ray/experimental/async_api.py --- a/python/ray/experimental/async_api.py +++ b/python/ray/experimental/async_api.py @@ -1,7 +1,4 @@ -# Note: asyncio is only compatible with Python 3 - import asyncio -import threading import ray from ray.experimental.async_plasma import PlasmaEventHandler @@ -10,7 +7,10 @@ handler = None -async def _async_init(): +def init(): + """Initialize plasma event handlers for asyncio support.""" + assert ray.is_initialized(), "Please call ray.init before async_api.init" + global handler if handler is None: worker = ray.worker.global_worker @@ -20,31 +20,6 @@ async def _async_init(): logger.debug("AsyncPlasma Connection Created!") -def init(): - """ - Initialize synchronously. - """ - assert ray.is_initialized(), "Please call ray.init before async_api.init" - - # Noop when handler is set. - if handler is not None: - return - - loop = asyncio.get_event_loop() - if loop.is_running(): - if loop._thread_id != threading.get_ident(): - # If the loop is runing outside current thread, we actually need - # to do this to make sure the context is initialized. - asyncio.run_coroutine_threadsafe(_async_init(), loop=loop) - else: - async_init_done = asyncio.get_event_loop().create_task( - _async_init()) - # Block until the async init finishes. - async_init_done.done() - else: - asyncio.get_event_loop().run_until_complete(_async_init()) - - def as_future(object_id): """Turn an object_id into a Future object.
diff --git a/python/ray/tests/test_asyncio.py b/python/ray/tests/test_asyncio.py --- a/python/ray/tests/test_asyncio.py +++ b/python/ray/tests/test_asyncio.py @@ -113,8 +113,8 @@ async def test_asyncio_get(ray_start_regular_shared, event_loop): loop.set_debug(True) # This is needed for async plasma - from ray.experimental.async_api import _async_init - await _async_init() + from ray.experimental.async_api import init + init() # Test Async Plasma @ray.remote
Ray async api is not working with uvloop. <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? Current Ray async api uses asyncio event loop's internal attribute to identify if the loop is running in the current current thread. ```python3 loop = asyncio.get_event_loop() if loop.is_running(): if loop._thread_id != threading.get_ident(): # If the loop is runing outside current thread, we actually need # to do this to make sure the context is initialized. asyncio.run_coroutine_threadsafe(_async_init(), loop=loop) ``` This causes a problem when we uses Ray APIs inside Fast API because Fast API uses uvloop as its main event loop, and uvloop doesn't have `_thread_id` attribute. @simon-mo Any good idea to fix this? It doesn't seem to be trivial. What about we do async_init() whenever asyncio loop is created in a different thread instead of checking if the event loop's thread id? I assume the only use case where asyncio loop is defined in a different thread is only inside async actor? ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): ```python3 import time import asyncio ​ import ray import psutil from fastapi import FastAPI, APIRouter ​ ​ app = FastAPI( title="API template", description="Template to build upon for API serving and distributed computation", version="0.1.0", openapi_url="/openapi.json", docs_url="/docs", ) ​ @app.on_event("startup") def startup_event(): ray.init(num_cpus=2) ​ ​ ​ @app.on_event("shutdown") def shutdown_event(): ray.shutdown() ​ ​ @app.get('/async') async def non_seq_async_process(): """ async distributed execution """ @ray.remote def slow_function(i): time.sleep(i) return i ​ start_time = time.time() ​ # result_ids = [] # for i in range(10, 60, 10): # result_ids.append(slow_function.remote(i)) # results = ray.get(result_ids) ​ results = await asyncio.wait([slow_function.remote(i) for i in range(10, 60, 10)]) ​ duration = time.time() - start_time out = "Executing the for loop took {:.3f} seconds.\n".format(duration) out += f"The results are: {results}\n" ``` If we cannot run your script, we cannot fix your issue. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-04-25T05:46:33
ray-project/ray
8,211
ray-project__ray-8211
[ "8203" ]
eb790bf3a3241b492171c321fc59234c66fae8d5
diff --git a/python/ray/worker.py b/python/ray/worker.py --- a/python/ray/worker.py +++ b/python/ray/worker.py @@ -435,10 +435,6 @@ def get_gpu_ids(): """ # TODO(ilr) Handle inserting resources in local mode - if _mode() == LOCAL_MODE: - logger.info("ray.get_gpu_ids() currently does not work in LOCAL " - "MODE.") - all_resource_ids = global_worker.core_worker.resource_ids() assigned_ids = [ resource_id for resource_id, _ in all_resource_ids.get("GPU", [])
Lots of "INFO worker.py:439 -- ray.get_gpu_ids() currently does not work in LOCAL MODE." messages printed in local mode <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? When I run in local mode I see lots of the following message being printed many times such that the console isn't really useable: ``` INFO worker.py:439 -- ray.get_gpu_ids() currently does not work in LOCAL MODE. ``` I'm not sure why this message is being printed as I'm not using a GPU (num_gpus=0, although my machine does have a GPU enabled). As I recall some time ago I was running in local mode and found that the GPU was being used when it wasn't supposed to be however. I first see this message printed with code version 9bfc2c4b5462871f0fe669b3aabc5f2772bc3a6f, and so assume it was caused by #7670. I'm not sure why this message is being printed when I shouldn't be using a GPU, and also I think the message should only be printed once and then suppressed in future instances when it would otherwise be printed *Ray version and other system information (Python version, TensorFlow version, OS):* Running on Azure NC6s_v2 machine ray: 0.9.0.dev0 (latest wheels installed on 4/27/20) python: 3.7.6 tensorflow: 2.1 torch: 1.4.0 OS: Ubuntu 16.04.6 LTS ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): I'm just running the following example, but changing `ray.init()` to `ray.init(local_mode=True)` ```bash python examples/custom_env.py ``` If we cannot run your script, we cannot fix your issue. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
cc @ijrsvt - can we mute this warning? Yes we can
2020-04-28T06:28:51
ray-project/ray
8,225
ray-project__ray-8225
[ "854" ]
bf25aee39228d25fe34b782ccfb02ded59f64f50
diff --git a/python/ray/worker.py b/python/ray/worker.py --- a/python/ray/worker.py +++ b/python/ray/worker.py @@ -1683,7 +1683,7 @@ def kill(actor): def cancel(object_id, force=False): - """Cancels a locally-submitted task according to the following conditions. + """Cancels a task according to the following conditions. If the specified task is pending execution, it will not be executed. If the task is currently executing, the behavior depends on the ``force`` @@ -1702,8 +1702,7 @@ def cancel(object_id, force=False): force (boolean): Whether to force-kill a running task by killing the worker that is running the task. Raises: - ValueError: This is also raised for actor tasks, already completed - tasks, and non-locally submitted tasks. + TypeError: This is also raised for actor tasks. """ worker = ray.worker.global_worker worker.check_connected()
diff --git a/python/ray/tests/test_cancel.py b/python/ray/tests/test_cancel.py --- a/python/ray/tests/test_cancel.py +++ b/python/ray/tests/test_cancel.py @@ -228,5 +228,31 @@ def wait_for(y): assert isinstance(e, valid_exceptions(use_force)) [email protected]("use_force", [True, False]) +def test_remote_cancel(ray_start_regular, use_force): + signaler = SignalActor.remote() + + @ray.remote + def wait_for(y): + return ray.get(y[0]) + + @ray.remote + def remote_wait(sg): + return [wait_for.remote([sg[0]])] + + sig = signaler.wait.remote() + + outer = remote_wait.remote([sig]) + inner = ray.get(outer)[0] + + with pytest.raises(RayTimeoutError): + ray.get(inner, 1) + + ray.cancel(inner) + + with pytest.raises(valid_exceptions(use_force)): + ray.get(inner, 10) + + if __name__ == "__main__": sys.exit(pytest.main(["-v", __file__]))
Ray kill remote tasks Is there any python interface for terminating remote tasks? In some situations, one may need to stop some tasks before they finish. For example, we may create several NNs to test a dataset. cnn1 = CNN1.remote() cnn2 = CNN2.remote() lstm1 = LSTM1.remote() lstm2 = LSTM2.remote() ... then we use `ray.wait` to wait for these tasks. and results (given by some tasks finished earlier) show that CNN is much better than LSTM for this dataset. So we would like to kill most LSTM tasks (we know their object_ids) to make more space for CNNs. We wonder how to do that with ray. @mitar
Currently there is no interface for terminating tasks, but I agree this could be really useful. One workaround would be to take your training tasks and divide them into smaller segments (e.g., 10 second chunks). Then whenever one training segment finishes, you can decide to launch a new training segment (picking up where the previous one left off) or to stop training that model. Does that achieves what you're looking for? As for canceling tasks, what sort of API do you think would make sense? E.g., `ray.cancel(object_id)` or something like that? I would suggest provide some mechanism for sending messages and then task can check periodically if there is a message. Because in our case we might want to terminate a task, but also maybe just pause a task. So different messages could be beneficial for this. So API could be: ```python ray.send_message(object_id, {message}) ``` If there is no task running for `object_id`, nothing happens (or maybe a return value could signal that). On receiving end, one could do something like `ray.get_messages()` to return any current pending messages available for the current task. I see. You could certainly try that out (e.g., hijacking the existing Redis server for that purpose). I wouldn't want to encourage anyone to actually do this though, but we can see if that provides the functionality that you're looking for. Something like the following. ```python def broadcast_message(key, message): ray.global_worker.redis_client.set(key, message) def check_message(key): return ray.global_worker.redis_client.get(key) ``` E.g., ```python import ray import time ray.init() def broadcast_message(key, message): ray.worker.global_worker.redis_client.set(key, message) def check_message(key): return ray.worker.global_worker.redis_client.get(key) @ray.remote def f(): while True: if check_message("XXX") == b"stop": break time.sleep(1) object_id = f.remote() broadcast_message("XXX", b"stop") # This should return ray.get(object_id) ``` A good way to send messages to tasks in the current API is to use actors. Each training task is a long running actor which has a train() method that does the training for a few steps (which takes say a few seconds); you can then send messages to the actor by invoking methods on it, for example for killing it or for changing some training settings, etc. These can be invoked in between the training tasks. -- Philipp. Does a task have access to its ObjectId? Maybe with some context manager? @pcmoritz: But wouldn't this introduce some latency? After each iteration, method has to return to the caller, then caller decides if to continue? Wouldn't be easier if method just check for any messages which arrived in meantime and then decide locally? While a task is executing, it does not have access to the object IDs of its return values. What would you use that for? As for latency, presumably you would train for 10 seconds or a minute and then you would have 1 millisecond of latency and then do another 10 seconds of training and so it probably wouldn't be an issue. BTW, if we just leave the actor there (stop calling any function of the actor), will the actor try to release all the resources it used? For example, when we initialize an actor, it creates a tf session and occupies some resources of GPU. Then we call `step()` of the actor or `keep_training_until_receive_msg()`. After breaking the training process, will it still stay in GPU until we manually close the tensorflow session in the actor? Once a process using TensorFlow starts to use space on a GPU, it never releases it until the process is killed (see #616), so the actor will continue to occupy space on the GPU until the actor is killed. The actor will naturally be killed when the actor handle goes out of scope in Python. E.g., if you do ```python actor = CNN1.remote() del actor ``` Then the call to `del actor` will send a task to the actor that will cause the actor to kill itself when the task runs. While we still haven't implemented regular cancelation, now that we have distributed actor handles, it is possible to work around it by using an actor to send messages from the driver to the task. Note that this requires the cooperation of the task itself. For example, ```python import ray import time ray.init() @ray.remote class TaskCanceler(object): def __init__(self): self.task_canceled = False def cancel_task(self): """The driver can call this to cancel the task.""" self.task_canceled = True def is_task_canceled(self): """The task calls this to check if it is canceled.""" return self.task_canceled @ray.remote def f(task_canceler): i = 0 while True: print(i) i += 1 time.sleep(1) # Check if the task should abort. if ray.get(task_canceler.is_task_canceled.remote()): return task_canceler = TaskCanceler.remote() # Create the task. f.remote(task_canceler) # Cancel the task. task_canceler.cancel_task.remote() ``` And of course this can be generalized in a lot of ways (e.g., to use one actor to selectively cancel many different tasks). This has just again turned up as an issue for our project. I realized that in fact besides cooperative termination also forceful termination would be useful. Sometimes there is simply no small iterations to have (like in traditional ML algorithms) to check some status. So I would ask for some way to kill a task forcefully, which would probably just terminate whole worker (I assume each worker is running only one task at a time, no). No matter the state or anything. I also need to kill the worker via API, it would be very handy. I'd start remote workers, then I'd do .wait with time limit and then I would like to kill the workers that didn't finish. I tried implementing cooperative stopping in the workers but that's very hard, because the worker is calling compiled 3rd party simulator. Hi guys, Any update on this? Do we have any APIs to kill the remote workers yet? I've ended up doing this by running my task in an actor on a background thread, a la ```python import ray import threading import time @ray.remote class Task: def __init__(self): self._thread = threading.Thread(target=self._run, daemon=True) self._thread.start() def _run(self): print('Started long-running task') time.sleep(10) print('You shouldn\'t see this because the process has been killed') ray.init() task = Task.remote() task.__ray_terminate__.remote() # kills the process with an `exit(0)` ``` This should work as long as your long-running task doesn't hang on `ray.get`/`ray.wait`/the GIL. `__ray_terminate__` will also be called when the original actor handle goes out of scope, so you don't have to call it explicitly if you don't want to. You can now kill actors while they're still processing tasks using the `actor.__ray_kill__()`. This is similar to the solution @andyljones posted except it will work even if the actor is blocking on `ray.get()`, `ray.wait()`, or the GIL. What about non-actors? So a simple Ray task? (BTW, any pointers how this is implemented for actors?) Not currently supported for tasks. See the original PR for details: https://github.com/ray-project/ray/pull/6523 So then it would be reasonable to turn all of my tasks into actors? That looks like a trivial thing to do. BTW, why is method called `__ray_kill__`? This is generally reserved for Python standard method names for interfaces which have language-based syntax to access. Like if there would be `kill(actor)` function which would then internally call `__ray_kill__` or something. That seems like a reasonable workaround for now. That's a good point about the API - what would you think of `ray.kill_actor(<actor>)` instead? What about just `ray.kill(<actor>)`? In the future it could also support killing tasks. BTW, any reason why it would not be as easy to expose API to kill tasks as well? I am not sure how much overhead is to have an actor instead of a task? It would be as easy to expose the API, but not to actually implement it. This is because anyone with a handle to an actor knows its location (RPC address) and can send a kill message, but to kill a task given only an `ObjectID` you would first have to look up where the task is running, which is possible but somewhat complicated and not currently supported. @mitar @suquark We'd love to setup a quick call to go over your requirements here and better understand your use case. We're sketching a design and it'd be helpful to have your input. Can we setup some time in the next week or so? shoot us an email my first name at anyscale.com
2020-04-29T04:42:20
ray-project/ray
8,231
ray-project__ray-8231
[ "7920" ]
bf25aee39228d25fe34b782ccfb02ded59f64f50
diff --git a/rllib/contrib/bandits/examples/LinTS_train_wheel_env.py b/rllib/contrib/bandits/examples/LinTS_train_wheel_env.py --- a/rllib/contrib/bandits/examples/LinTS_train_wheel_env.py +++ b/rllib/contrib/bandits/examples/LinTS_train_wheel_env.py @@ -10,7 +10,7 @@ def plot_model_weights(means, covs): fmts = ["bo", "ro", "yx", "k+", "gx"] - labels = [f"arm{i}" for i in range(5)] + labels = ["arm{}".format(i) for i in range(5)] fig, ax = plt.subplots(figsize=(6, 4)) diff --git a/rllib/contrib/bandits/examples/tune_LinTS_train_wheel_env.py b/rllib/contrib/bandits/examples/tune_LinTS_train_wheel_env.py --- a/rllib/contrib/bandits/examples/tune_LinTS_train_wheel_env.py +++ b/rllib/contrib/bandits/examples/tune_LinTS_train_wheel_env.py @@ -15,7 +15,7 @@ def plot_model_weights(means, covs, ax): fmts = ["bo", "ro", "yx", "k+", "gx"] - labels = [f"arm{i}" for i in range(5)] + labels = ["arm{}".format(i) for i in range(5)] ax.set_title("Weights distributions of arms") diff --git a/rllib/contrib/bandits/models/linear_regression.py b/rllib/contrib/bandits/models/linear_regression.py --- a/rllib/contrib/bandits/models/linear_regression.py +++ b/rllib/contrib/bandits/models/linear_regression.py @@ -85,9 +85,9 @@ def _check_inputs(self, x, y=None): assert x.ndim in [2, 3], \ "Input context tensor must be 2 or 3 dimensional, where the" \ " first dimension is batch size" - assert x.shape[ - 1] == self.d, f"Feature dimensions of weights ({self.d}) and " \ - f"context ({x.shape[1]}) do not match!" + assert x.shape[1] == self.d, \ + "Feature dimensions of weights ({}) and context ({}) do not " \ + "match!".format(self.d, x.shape[1]) if y: assert torch.is_tensor(y) and y.numel() == 1,\ "Target should be a tensor;" \ @@ -134,9 +134,9 @@ def predict(self, x, sample_theta=False, use_ucb=False): return scores def partial_fit(self, x, y, arm): - assert 0 <= arm.item() < len(self.arms),\ - f"Invalid arm: {arm.item()}." \ - f"It should be 0 <= arm < {len(self.arms)}" + assert 0 <= arm.item() < len(self.arms), \ + "Invalid arm: {}. It should be 0 <= arm < {}".format( + arm.item(), len(self.arms)) self.arms[arm].partial_fit(x, y) @override(ModelV2)
[rllib] Python 3.5 compaltibility <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? In the fresh install of deep learning VM on GCE, rllib could not start training PPO with CartPole-v0. There are some f-strings in the code and such code do not run on python 3.5. Tested environment: + Debian9 on Google Compute Engine with Deep learning VM (Tensorflow 1.15.2) + Python 3.5.3 + tensorflow-gpu 1.15.2 (Installed with Deep learning VM) + some python modules > pip3 install dm-tree tabulate lz4 tensorboardX torch torchvision gym + Ray 0.8.4 > pip3 install ray + and Ray 0.9.0.dev0 (installed from snapshot wheel) > pip3 install -U https://s3-us-west-2.amazonaws.com/ray-wheels/latest/ray-0.9.0.dev0-cp35-cp35m-manylinux1_x86_64.whl ### Reproduction (REQUIRED) From the command line on the document "rllib in 60 seconds" > rllib train --run=PPO --env=CartPole-v0 Results ``` Traceback (most recent call last): File "/home/yos/bin/rllib", line 5, in <module> from ray.rllib.scripts import cli File "/home/yos/.local/lib/python3.5/site-packages/ray/rllib/__init__.py", line 60, in <module> _register_all() File "/home/yos/.local/lib/python3.5/site-packages/ray/rllib/__init__.py", line 37, in _register_all register_trainable(key, get_agent_class(key)) File "/home/yos/.local/lib/python3.5/site-packages/ray/rllib/agents/registry.py", line 130, in get_agent_class return _get_agent_class(alg) File "/home/yos/.local/lib/python3.5/site-packages/ray/rllib/agents/registry.py", line 140, in _get_agent_class return CONTRIBUTED_ALGORITHMS[alg]() File "/home/yos/.local/lib/python3.5/site-packages/ray/rllib/contrib/registry.py", line 26, in _import_bandit_linucb from ray.rllib.contrib.bandits.agents.lin_ucb import LinUCBTrainer File "/home/yos/.local/lib/python3.5/site-packages/ray/rllib/contrib/bandits/agents/__init__.py", line 1, in <module> from ray.rllib.contrib.bandits.agents.lin_ts import LinTSTrainer File "/home/yos/.local/lib/python3.5/site-packages/ray/rllib/contrib/bandits/agents/lin_ts.py", line 5, in <module> from ray.rllib.contrib.bandits.agents.policy import BanditPolicy File "/home/yos/.local/lib/python3.5/site-packages/ray/rllib/contrib/bandits/agents/policy.py", line 6, in <module> from ray.rllib.contrib.bandits.models.linear_regression import \ File "/home/yos/.local/lib/python3.5/site-packages/ray/rllib/contrib/bandits/models/linear_regression.py", line 87 1] == self.d, f"Feature dimensions of weights ({self.d}) and " \ ^ SyntaxError: invalid syntax ``` I've searched f-string in the source tree and patched like this commit(73ecb585b9f43a021e74f802ae83c9ed9889b4bf), and syntax error is gone. However, even with this patch, training of CartPole-v0 could not be started because of another error: ``` 2020-04-07 14:13:30,625 ERROR trial_runner.py:521 -- Trial PPO_CartPole-v0_00000: Error processing event. Traceback (most recent call last): File "/home/yos/.local/lib/python3.5/site-packages/ray/tune/trial_runner.py", line 467, in _process_trial result = self.trial_executor.fetch_result(trial) File "/home/yos/.local/lib/python3.5/site-packages/ray/tune/ray_trial_executor.py", line 381, in fetch_result result = ray.get(trial_future[0], DEFAULT_GET_TIMEOUT) File "/home/yos/.local/lib/python3.5/site-packages/ray/worker.py", line 1486, in get raise value.as_instanceof_cause() ray.exceptions.RayTaskError(NameError): ray::PPO.__init__() (pid=1684, ip=10.140.0.3) File "/home/yos/.local/lib/python3.5/site-packages/ray/rllib/utils/from_config.py", line 157, in from_config constructor = getattr(module, type_) AttributeError: module 'ray.rllib.utils.exploration.exploration' has no attribute 'StochasticSampling' ``` I have no idea with this error. Any help please? - [V ] I have verified my script runs in a clean environment and reproduces the issue. - [V ] I have verified the issue also occurs with the [latest wheels]
2020-04-29T14:58:05
ray-project/ray
8,279
ray-project__ray-8279
[ "6993" ]
2f01776d09835ffe93a8c608391059176c7688c5
diff --git a/python/ray/scripts/scripts.py b/python/ray/scripts/scripts.py --- a/python/ray/scripts/scripts.py +++ b/python/ray/scripts/scripts.py @@ -7,6 +7,8 @@ import subprocess import sys import time +import urllib +import urllib.parse import ray.services as services from ray.autoscaler.commands import ( @@ -566,6 +568,16 @@ def create_or_update(cluster_config_file, min_workers, max_workers, no_restart, if restart_only or no_restart: assert restart_only != no_restart, "Cannot set both 'restart_only' " \ "and 'no_restart' at the same time!" + if urllib.parse.urlparse(cluster_config_file).scheme in ("http", "https"): + try: + response = urllib.request.urlopen(cluster_config_file, timeout=5) + content = response.read() + file_name = cluster_config_file.split("/")[-1] + with open(file_name, "wb") as f: + f.write(content) + cluster_config_file = file_name + except urllib.error.HTTPError as e: + logger.info("Error downloading file: ", e) create_or_update_cluster(cluster_config_file, min_workers, max_workers, no_restart, restart_only, yes, cluster_name)
diff --git a/python/ray/tests/test_autoscaler_yaml.py b/python/ray/tests/test_autoscaler_yaml.py --- a/python/ray/tests/test_autoscaler_yaml.py +++ b/python/ray/tests/test_autoscaler_yaml.py @@ -2,6 +2,8 @@ import os import unittest import yaml +import urllib +import tempfile from ray.autoscaler.autoscaler import fillout_defaults, validate_config from ray.test_utils import recursive_fnmatch @@ -25,6 +27,21 @@ def testValidateDefaultConfig(self): except Exception: self.fail("Config did not pass validation test!") + def testValidateNetworkConfig(self): + web_yaml = "https://raw.githubusercontent.com/ray-project/ray/" \ + "master/python/ray/autoscaler/aws/example-full.yaml" + response = urllib.request.urlopen(web_yaml, timeout=5) + content = response.read() + with tempfile.TemporaryFile() as f: + f.write(content) + f.seek(0) + config = yaml.safe_load(f) + config = fillout_defaults(config) + try: + validate_config(config) + except Exception: + self.fail("Config did not pass validation test!") + def _test_invalid_config(self, config_path): with open(os.path.join(RAY_PATH, config_path)) as f: config = yaml.safe_load(f)
Allow `ray up` to take URL as argument. To get started with the autoscaler, we tell people to run a line like `ray up ray/python/ray/autoscaler/aws/example-full.yaml`. In practice, the steps are 1. Figure out that this means you have to clone the Ray repository. 2. Clone the Ray repository. 3. Figure out the new correct relative file path based on where you cloned it. 4. Run `ray up ...`. Instead, we could let people just copy and paste the following command: ``` ray up https://github.com/ray-project/ray/blob/master/python/ray/autoscaler/aws/example-full.yaml ``` In this blog post https://towardsdatascience.com/10x-faster-parallel-python-without-python-multiprocessing-e5017c93cce1, I posted benchmarks at the end that they could run with the autoscaler, but they still have to `wget` the relevant file. It'd be much easier if they could just use the URL. Could also just write stuff like the following, but one line would be better. ``` wget https://raw.githubusercontent.com/ray-project/ray/master/python/ray/autoscaler/aws/example-full.yaml ray up example-full.yaml ```
Actually, even better, maybe we can have default cluster profiles available. For example, ``` ray up examples.tune ray up examples.sgd ray up examples.small ``` The url thing is a standard practice in kubernetes world. for example https://aws.amazon.com/premiumsupport/knowledge-center/eks-cluster-kubernetes-dashboard/ @richardliaw the thing I like about the URL approach is that it is extensible, so if I write a blog post, I can host my own cluster yaml somewhere and then give people one command to run to reproduce the experiments. This might be within my capability (found via `good first issue` tag). Can someone reference me to the `ray up` section of the codebase? Sure! Here's the code that correspond to `ray up` https://github.com/ray-project/ray/blob/b23b6addfcfd4be7a27e58aefbecb3fd7dd6510c/python/ray/scripts/scripts.py#L523-L566 https://github.com/ray-project/ray/blob/b23b6addfcfd4be7a27e58aefbecb3fd7dd6510c/python/ray/scripts/scripts.py#L999 Thanks! Yeah, I think I can do that. I'll take a stab over the weekend and get you a PR.
2020-05-01T22:12:54
ray-project/ray
8,324
ray-project__ray-8324
[ "8319" ]
5f278c6411eda9d0d5e69e5e14e3496a17daf298
diff --git a/rllib/examples/env/simple_corridor.py b/rllib/examples/env/simple_corridor.py --- a/rllib/examples/env/simple_corridor.py +++ b/rllib/examples/env/simple_corridor.py @@ -12,24 +12,21 @@ def __init__(self, config): self.end_pos = config["corridor_length"] self.cur_pos = 0 self.action_space = Discrete(2) - self.observation_space = Box( - 0.0, self.end_pos, shape=(1, ), dtype=np.float32) + self.observation_space = Box(0.0, 999.0, shape=(1, ), dtype=np.float32) def set_corridor_length(self, length): self.end_pos = length - self.observation_space = Box( - 0.0, self.end_pos, shape=(1, ), dtype=np.float32) print("Updated corridor length to {}".format(length)) def reset(self): - self.cur_pos = 0 + self.cur_pos = 0.0 return [self.cur_pos] def step(self, action): assert action in [0, 1], action if action == 0 and self.cur_pos > 0: - self.cur_pos -= 1 + self.cur_pos -= 1.0 elif action == 1: - self.cur_pos += 1 + self.cur_pos += 1.0 done = self.cur_pos >= self.end_pos return [self.cur_pos], 1 if done else 0, done, {} diff --git a/rllib/models/tf/tf_action_dist.py b/rllib/models/tf/tf_action_dist.py --- a/rllib/models/tf/tf_action_dist.py +++ b/rllib/models/tf/tf_action_dist.py @@ -326,6 +326,11 @@ def _unsquash(self, values): unsquashed = tf.math.atanh(save_normed_values) return unsquashed + @staticmethod + @override(ActionDistribution) + def required_model_output_shape(action_space, model_config): + return np.prod(action_space.shape) * 2 + class Beta(TFActionDistribution): """ @@ -371,6 +376,11 @@ def _squash(self, raw_values): def _unsquash(self, values): return (values - self.low) / (self.high - self.low) + @staticmethod + @override(ActionDistribution) + def required_model_output_shape(action_space, model_config): + return np.prod(action_space.shape) * 2 + class Deterministic(TFActionDistribution): """Action distribution that returns the input values directly. diff --git a/rllib/utils/exploration/random.py b/rllib/utils/exploration/random.py --- a/rllib/utils/exploration/random.py +++ b/rllib/utils/exploration/random.py @@ -1,4 +1,4 @@ -from gym.spaces import Discrete, MultiDiscrete, Tuple +from gym.spaces import Discrete, Box, MultiDiscrete import numpy as np import tree from typing import Union @@ -9,6 +9,7 @@ from ray.rllib.utils import force_tuple from ray.rllib.utils.framework import try_import_tf, try_import_torch, \ TensorType +from ray.rllib.utils.space_utils import get_base_struct_from_space tf = try_import_tf() torch, _ = try_import_torch() @@ -35,13 +36,8 @@ def __init__(self, action_space, *, model, framework, **kwargs): framework=framework, **kwargs) - # Determine py_func types, depending on our action-space. - if isinstance(self.action_space, (Discrete, MultiDiscrete)) or \ - (isinstance(self.action_space, Tuple) and - isinstance(self.action_space[0], (Discrete, MultiDiscrete))): - self.dtype_sample, self.dtype = (tf.int64, tf.int32) - else: - self.dtype_sample, self.dtype = (tf.float64, tf.float32) + self.action_space_struct = get_base_struct_from_space( + self.action_space) @override(Exploration) def get_exploration_action(self, @@ -59,14 +55,46 @@ def get_exploration_action(self, def get_tf_exploration_action_op(self, action_dist, explore): def true_fn(): - action = tf.py_function(self.action_space.sample, [], - self.dtype_sample) - # Will be unnecessary, once we support batch/time-aware Spaces. - return tf.expand_dims(tf.cast(action, dtype=self.dtype), 0) + batch_size = 1 + req = force_tuple( + action_dist.required_model_output_shape( + self.action_space, self.model.model_config)) + # Add a batch dimension? + if len(action_dist.inputs.shape) == len(req) + 1: + batch_size = tf.shape(action_dist.inputs)[0] + + # Function to produce random samples from primitive space + # components: (Multi)Discrete or Box. + def random_component(component): + if isinstance(component, Discrete): + return tf.random.uniform( + shape=(batch_size, ) + component.shape, + maxval=component.n, + dtype=component.dtype) + elif isinstance(component, MultiDiscrete): + return tf.random.uniform( + shape=(batch_size, ) + component.shape, + maxval=component.nvec, + dtype=component.dtype) + elif isinstance(component, Box): + if component.bounded_above.all() and \ + component.bounded_below.all(): + return tf.random.uniform( + shape=(batch_size, ) + component.shape, + minval=component.low, + maxval=component.high, + dtype=component.dtype) + else: + return tf.random.normal( + shape=(batch_size, ) + component.shape, + dtype=component.dtype) + + actions = tree.map_structure(random_component, + self.action_space_struct) + return actions def false_fn(): - return tf.cast( - action_dist.deterministic_sample(), dtype=self.dtype) + return action_dist.deterministic_sample() action = tf.cond( pred=tf.constant(explore, dtype=tf.bool) @@ -81,15 +109,17 @@ def false_fn(): def get_torch_exploration_action(self, action_dist, explore): if explore: - # Unsqueeze will be unnecessary, once we support batch/time-aware - # Spaces. - a = self.action_space.sample() req = force_tuple( action_dist.required_model_output_shape( self.action_space, self.model.model_config)) - # Add a batch dimension. + # Add a batch dimension? if len(action_dist.inputs.shape) == len(req) + 1: - a = np.expand_dims(a, 0) + batch_size = action_dist.inputs.shape[0] + a = np.stack( + [self.action_space.sample() for _ in range(batch_size)]) + else: + a = self.action_space.sample() + # Convert action to torch tensor. action = torch.from_numpy(a).to(self.device) else: action = action_dist.deterministic_sample()
diff --git a/rllib/agents/ddpg/tests/test_ddpg.py b/rllib/agents/ddpg/tests/test_ddpg.py --- a/rllib/agents/ddpg/tests/test_ddpg.py +++ b/rllib/agents/ddpg/tests/test_ddpg.py @@ -22,11 +22,12 @@ def test_ddpg_compilation(self): """Test whether a DDPGTrainer can be built with both frameworks.""" config = ddpg.DEFAULT_CONFIG.copy() config["num_workers"] = 0 # Run locally. + config["num_envs_per_worker"] = 2 # Run locally. num_iterations = 2 # Test against all frameworks. - for _ in framework_iterator(config, ("torch", "tf")): + for _ in framework_iterator(config, ("tf", "torch")): trainer = ddpg.DDPGTrainer(config=config, env="Pendulum-v0") for i in range(num_iterations): results = trainer.train() @@ -366,6 +367,8 @@ def test_ddpg_loss_function(self): else: check(tf_var, torch_var, rtol=0.07) + trainer.stop() + def _get_batch_helper(self, obs_size, actions, batch_size): return { SampleBatch.CUR_OBS: np.random.random(size=obs_size), diff --git a/rllib/tests/test_multi_agent_pendulum.py b/rllib/tests/test_multi_agent_pendulum.py --- a/rllib/tests/test_multi_agent_pendulum.py +++ b/rllib/tests/test_multi_agent_pendulum.py @@ -2,9 +2,10 @@ import unittest import ray -from ray.rllib.examples.env.multi_agent import MultiAgentPendulum from ray.tune import run_experiments from ray.tune.registry import register_env +from ray.rllib.examples.env.multi_agent import MultiAgentPendulum +from ray.rllib.utils.test_utils import framework_iterator class TestMultiAgentPendulum(unittest.TestCase): @@ -17,34 +18,38 @@ def tearDown(self) -> None: def test_multi_agent_pendulum(self): register_env("multi_agent_pendulum", lambda _: MultiAgentPendulum({"num_agents": 1})) - trials = run_experiments({ - "test": { - "run": "PPO", - "env": "multi_agent_pendulum", - "stop": { - "timesteps_total": 500000, - "episode_reward_mean": -200, - }, - "config": { - "train_batch_size": 2048, - "vf_clip_param": 10.0, - "num_workers": 0, - "num_envs_per_worker": 10, - "lambda": 0.1, - "gamma": 0.95, - "lr": 0.0003, - "sgd_minibatch_size": 64, - "num_sgd_iter": 10, - "model": { - "fcnet_hiddens": [128, 128], + + # Test for both torch and tf. + for fw in framework_iterator(frameworks=["torch", "tf"]): + trials = run_experiments({ + "test": { + "run": "PPO", + "env": "multi_agent_pendulum", + "stop": { + "timesteps_total": 500000, + "episode_reward_mean": -300.0, + }, + "config": { + "train_batch_size": 2048, + "vf_clip_param": 10.0, + "num_workers": 0, + "num_envs_per_worker": 10, + "lambda": 0.1, + "gamma": 0.95, + "lr": 0.0003, + "sgd_minibatch_size": 64, + "num_sgd_iter": 10, + "model": { + "fcnet_hiddens": [128, 128], + }, + "batch_mode": "complete_episodes", + "use_pytorch": fw == "torch", }, - "batch_mode": "complete_episodes", - }, - } - }) - if trials[0].last_result["episode_reward_mean"] < -200: - raise ValueError("Did not get to -200 reward", - trials[0].last_result) + } + }) + if trials[0].last_result["episode_reward_mean"] < -300.0: + raise ValueError("Did not get to -200 reward", + trials[0].last_result) if __name__ == "__main__":
[rllib] Vectorization & multi-agent are broken in DDPG (both TF and Torch) <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? `./train.py --run=DDPG --env=MountainCarContinuous-v0 --config='{"num_envs_per_worker": 2}'` and `./train.py --run=DDPG --env=MountainCarContinuous-v0 --config='{"num_envs_per_worker": 2}' --torch` both currently crash. The issue seems to be that while the input to compute_actions() is a batch of N observations, only 1 action is returned as output. I ran into this while debugging a hang in a multi-agent test case (hung due to a bug in the env triggered by vectorization returning 1 action instead of N actions).
2020-05-05T13:09:48
ray-project/ray
8,366
ray-project__ray-8366
[ "8307" ]
2c599dbf05e41e338920ee2fbe692658bcbec4dd
diff --git a/python/ray/tune/ray_trial_executor.py b/python/ray/tune/ray_trial_executor.py --- a/python/ray/tune/ray_trial_executor.py +++ b/python/ray/tune/ray_trial_executor.py @@ -256,9 +256,6 @@ def _stop_trial(self, trial, error=False, error_msg=None, error_msg (str): Optional error message. stop_logger (bool): Whether to shut down the trial logger. """ - if stop_logger: - trial.close_logger() - self.set_status(trial, Trial.ERROR if error else Trial.TERMINATED) trial.set_location(Location()) @@ -278,6 +275,8 @@ def _stop_trial(self, trial, error=False, error_msg=None, self.set_status(trial, Trial.ERROR) finally: trial.set_runner(None) + if stop_logger: + trial.close_logger() def start_trial(self, trial, checkpoint=None, train=True): """Starts the trial.
[tune] close logger after updating Trial status <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### Describe your feature request Several experiment tracking systems, such as [neptune](https://neptune.ai/) or [Weights&Biases](https://www.wandb.com/) offer to categorize experiments as "Succeeded" or "Failed". This status is set when closing the logger. Today, the `RayTrialExecutor._stop_trial` method closes the loggers before updating the status of the trial and updating its `trial.error_msg` attribute. When loggers are closed, it is impossible to know whether the trial is terminated normally or with an error, and therefore to set correct status in the experiment trackers. To me, the solution seems as simple as moving these two lines: https://github.com/ray-project/ray/blob/8625e09067a0f87b9951b2b41a7c7e0fead66f6e/python/ray/tune/ray_trial_executor.py#L259-L260 at the end of the `_stop_trial` method. Is there any reason to close logger at the beginning of `_stop_trial`? Happy to create a PR otherwise! Thanks,
Yeah that sounds good - in the finally block right? A PR would be really appreciated - thanks!
2020-05-08T10:02:35
ray-project/ray
8,445
ray-project__ray-8445
[ "8147" ]
5f4c196feda42d022c12c07e5cd4182fbbf46cba
diff --git a/python/ray/tune/analysis/experiment_analysis.py b/python/ray/tune/analysis/experiment_analysis.py --- a/python/ray/tune/analysis/experiment_analysis.py +++ b/python/ray/tune/analysis/experiment_analysis.py @@ -220,8 +220,10 @@ def get_best_trial(self, metric, mode="max", scope="all"): Args: metric (str): Key for trial info to order on. mode (str): One of [min, max]. - scope (str): One of [all, last]. If `scope=last`, only look at + scope (str): One of [all, last, avg]. If `scope=last`, only look at each trial's final step for `metric`, and compare across + trials based on `mode=[min,max]`. If `scope=avg`, consider the + simple average over all steps for `metric` and compare across trials based on `mode=[min,max]`. If `scope=all`, find each trial's min/max score for `metric` based on `mode`, and compare trials based on `mode=[min,max]`. @@ -231,11 +233,11 @@ def get_best_trial(self, metric, mode="max", scope="all"): "ExperimentAnalysis: attempting to get best trial for " "metric {} for mode {} not in [\"max\", \"min\"]".format( metric, mode)) - if scope not in ["all", "last"]: + if scope not in ["all", "last", "avg"]: raise ValueError( "ExperimentAnalysis: attempting to get best trial for " - "metric {} for scope {} not in [\"all\", \"last\"]".format( - metric, scope)) + "metric {} for scope {} not in [\"all\", \"last\", \"avg\"]". + format(metric, scope)) best_trial = None best_metric_score = None for trial in self.trials: @@ -244,6 +246,8 @@ def get_best_trial(self, metric, mode="max", scope="all"): if scope == "last": metric_score = trial.metric_analysis[metric]["last"] + elif scope == "avg": + metric_score = trial.metric_analysis[metric]["avg"] else: metric_score = trial.metric_analysis[metric][mode] @@ -269,8 +273,10 @@ def get_best_config(self, metric, mode="max", scope="all"): Args: metric (str): Key for trial info to order on. mode (str): One of [min, max]. - scope (str): One of [all, last]. If `scope=last`, only look at + scope (str): One of [all, last, avg]. If `scope=last`, only look at each trial's final step for `metric`, and compare across + trials based on `mode=[min,max]`. If `scope=avg`, consider the + simple average over all steps for `metric` and compare across trials based on `mode=[min,max]`. If `scope=all`, find each trial's min/max score for `metric` based on `mode`, and compare trials based on `mode=[min,max]`. @@ -286,8 +292,10 @@ def get_best_logdir(self, metric, mode="max", scope="all"): Args: metric (str): Key for trial info to order on. mode (str): One of [min, max]. - scope (str): One of [all, last]. If `scope=last`, only look at + scope (str): One of [all, last, avg]. If `scope=last`, only look at each trial's final step for `metric`, and compare across + trials based on `mode=[min,max]`. If `scope=avg`, consider the + simple average over all steps for `metric` and compare across trials based on `mode=[min,max]`. If `scope=all`, find each trial's min/max score for `metric` based on `mode`, and compare trials based on `mode=[min,max]`. diff --git a/python/ray/tune/trial.py b/python/ray/tune/trial.py --- a/python/ray/tune/trial.py +++ b/python/ray/tune/trial.py @@ -214,7 +214,7 @@ def __init__(self, self.last_result = {} self.last_update_time = -float("inf") - # stores in memory max/min/last result for each metric by trial + # stores in memory max/min/avg/last result for each metric by trial self.metric_analysis = {} self.export_formats = export_formats @@ -476,13 +476,18 @@ def update_last_result(self, result, terminate=False): self.metric_analysis[metric] = { "max": value, "min": value, + "avg": value, "last": value } else: + step = result["training_iteration"] or 1 self.metric_analysis[metric]["max"] = max( value, self.metric_analysis[metric]["max"]) self.metric_analysis[metric]["min"] = min( value, self.metric_analysis[metric]["min"]) + self.metric_analysis[metric]["avg"] = 1 / step * ( + value + + (step - 1) * self.metric_analysis[metric]["avg"]) self.metric_analysis[metric]["last"] = value def get_trainable_cls(self):
diff --git a/python/ray/tune/tests/test_experiment_analysis_mem.py b/python/ray/tune/tests/test_experiment_analysis_mem.py --- a/python/ray/tune/tests/test_experiment_analysis_mem.py +++ b/python/ray/tune/tests/test_experiment_analysis_mem.py @@ -3,6 +3,7 @@ import tempfile import random import pandas as pd +import numpy as np import ray from ray.tune import run, Trainable, sample_from, Analysis, grid_search @@ -12,16 +13,17 @@ class ExperimentAnalysisInMemorySuite(unittest.TestCase): def setUp(self): class MockTrainable(Trainable): + scores_dict = { + 0: [5, 4, 0], + 1: [4, 3, 1], + 2: [2, 1, 8], + 3: [9, 7, 6], + 4: [7, 5, 3] + } + def _setup(self, config): self.id = config["id"] self.idx = 0 - self.scores_dict = { - 0: [5, 0], - 1: [4, 1], - 2: [2, 8], - 3: [9, 6], - 4: [7, 3] - } def _train(self): val = self.scores_dict[self.id][self.idx] @@ -43,14 +45,15 @@ def tearDown(self): def testCompareTrials(self): self.test_dir = tempfile.mkdtemp() - scores_all = [5, 4, 2, 9, 7, 0, 1, 8, 6, 3] + scores = np.asarray(list(self.MockTrainable.scores_dict.values())) + scores_all = scores.flatten("F") scores_last = scores_all[5:] ea = run( self.MockTrainable, name="analysis_exp", local_dir=self.test_dir, - stop={"training_iteration": 2}, + stop={"training_iteration": 3}, num_samples=1, config={"id": grid_search(list(range(5)))}) @@ -60,9 +63,15 @@ def testCompareTrials(self): "min").metric_analysis["score"]["min"] max_last = ea.get_best_trial("score", "max", "last").metric_analysis["score"]["last"] + max_avg = ea.get_best_trial("score", "max", + "avg").metric_analysis["score"]["avg"] + min_avg = ea.get_best_trial("score", "min", + "avg").metric_analysis["score"]["avg"] self.assertEqual(max_all, max(scores_all)) self.assertEqual(min_all, min(scores_all)) self.assertEqual(max_last, max(scores_last)) + self.assertAlmostEqual(max_avg, max(np.mean(scores, axis=1))) + self.assertAlmostEqual(min_avg, min(np.mean(scores, axis=1))) self.assertNotEqual(max_last, max(scores_all))
[tune]add 'last-n-avg' option for scope parameter in get_best_* methods <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### Describe your feature request As described in the [Google Groups](https://mail.google.com/mail/u/0/#inbox/FMfcgxwHMZGcqCZltQQjfxMpKMWpXKdK), it makes sense to evaluate model on every batch of time series data (like _hour-01, hour-02, ..., hour-23_) when we are doing some online learning stuff (for example ads\feeds recommending). When dealing with such situation, it's better to take **average of last n or all** steps' metric of one trial into consideration when we are retrieving the best trial. So I sugguest add **'avg'**, **'last-n-avg'** options to the **scope** parameter in following methods - ray.tune.ExperimentAnalysis.get_best_trial - ray.tune.ExperimentAnalysis.get_best_config - ray.tune.ExperimentAnalysis.get_best_logdir
cc @krfricke
2020-05-14T11:53:14
ray-project/ray
8,466
ray-project__ray-8466
[ "8464" ]
9a83908c460a8bdf38b665c290a003e492a87e16
diff --git a/python/setup.py b/python/setup.py --- a/python/setup.py +++ b/python/setup.py @@ -194,7 +194,7 @@ def find_version(*filepath): "google", "grpcio", "jsonschema", - "msgpack >= 0.6.0, < 1.0.0", + "msgpack >= 0.6.0, < 2.0.0", "numpy >= 1.16", "protobuf >= 3.8.0", "py-spy >= 0.2.0", diff --git a/streaming/python/runtime/gateway_client.py b/streaming/python/runtime/gateway_client.py --- a/streaming/python/runtime/gateway_client.py +++ b/streaming/python/runtime/gateway_client.py @@ -69,4 +69,4 @@ def serialize(obj) -> bytes: def deserialize(data: bytes): """Deserialize the binary data serialized by `PythonGateway`""" - return msgpack.unpackb(data, raw=False) + return msgpack.unpackb(data, raw=False, strict_map_key=False) diff --git a/streaming/python/runtime/serialization.py b/streaming/python/runtime/serialization.py --- a/streaming/python/runtime/serialization.py +++ b/streaming/python/runtime/serialization.py @@ -41,7 +41,7 @@ def serialize(self, obj): return msgpack.packb(fields, use_bin_type=True) def deserialize(self, data): - fields = msgpack.unpackb(data, raw=False) + fields = msgpack.unpackb(data, raw=False, strict_map_key=False) if fields[0] == _RECORD_TYPE_ID: stream, value = fields[1:] record = message.Record(value)
Bump msgpack to 1.0.0 for Python We've started bumping msgpack to 1.0.0 in some of our repos at our company. However, some of our repos use Ray, which doesn't support version 1.0.0 for msgpack. We're hoping that Ray could support msgpack 1.0.0 - it doesn't look like there are too many breaking changes, and it looks like both major versions 0 and 1 can be supported. Thanks!
cc @simon-mo. Thanks for reporting this @alecbrick . We will see if we can have time to work on this, but always feel free to contribute!
2020-05-16T08:10:05
ray-project/ray
8,491
ray-project__ray-8491
[ "5982" ]
5b330de1820b6b0115833dff5b890fd7bbf56080
diff --git a/python/ray/experimental/tf_utils.py b/python/ray/experimental/tf_utils.py --- a/python/ray/experimental/tf_utils.py +++ b/python/ray/experimental/tf_utils.py @@ -161,10 +161,7 @@ def get_weights(self): Dictionary mapping variable names to their weights. """ self._check_sess() - return { - k: v.eval(session=self.sess) - for k, v in self.variables.items() - } + return self.sess.run(self.variables) def set_weights(self, new_weights): """Sets the weights to new_weights.
Time to initialize a policy grows linearly with the number of agents <!-- General questions should be asked on the mailing list [email protected]. Questions about how to use Ray should be asked on [StackOverflow](https://stackoverflow.com/questions/tagged/ray). Before submitting an issue, please fill out the following form. --> ### System information - **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**: Linux Ubuntu 18.04 - **Ray installed from (source or binary)**: Binary - **Ray version**: 0.7.4 - **Python version**: 3.7.4 - **Exact command to reproduce**: N/A <!-- You can obtain the Ray version with python -c "import ray; print(ray.__version__)" --> ### Describe the problem <!-- Describe the problem clearly here. --> I noticed that in multi agent settings, the time to initialize a policy per agent increases as more agents are initialized. In the sample output I provided below, you can see that the time to initialize a single DynamicTFPolicy grows from 4.6 seconds to 15.3 seconds from the first agent to the tenth agent created. Line 291 of `rllib/policy/dynamic_tf_policy.py` is ```python self._sess.run(tf.global_variables_initializer()) ``` which I believe will run one time for each agent initialized. If I'm not mistaken, this means that every variable in the computation graph is being initialized each time that we initialize a DynamicTFPolicy. If initializing a DynamicTFPolicy adds new variables to the computation graph (as I believe it does), this would explain why the time to initialize a DynamicTFPolicy grows over time: We are initializing every variable in the computation graph, and the computation graph is growing. My question is, why does line 291 run a global variables initializer? Is there a reason for this that I can't see inside this method? How hard would it be to modify this to only initialize variables in the individual policy that we care to initialize? I'm asking this because as detailed in #5753, I'm trying to modify rllib to allow initialization and removal of policies during training. The overhead incurred by this initialization quickly slows the training script down enough to be useless. Also, if anyone knows what the resource bottleneck is for policy initialization, that would be very helpful to know for when we're picking new hardware. Does it need a ton of cores to run in parallel, or more memory, or a bigger GPU or more GPUs or something? Thanks. ### Source code / logs <!-- Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached. Try to provide a reproducible test case that is the bare minimum necessary to generate the problem. --> ``` (pytorch) root@e3a955e42cae:~/bees/bees# python trainer.py settings/settings.json 2019-10-23 12:38:04,168 WARNING worker.py:1426 -- WARNING: Not updating worker name since `setproctitle` is not installed. Install this with `pip install setproctitle` (or ray[debug]) to enable monitoring of worker processes. 2019-10-23 12:38:04,169 INFO resource_spec.py:205 -- Starting Ray with 3.52 GiB memory available for workers and up to 1.78 GiB for objects. You can adjust these settings with ray.init(memory=<bytes>, object_store_memory=<bytes>). 2019-10-23 12:38:05,354 INFO trainer.py:344 -- Tip: set 'eager': true or the --eager flag to enable TensorFlow eager execution 2019-10-23 12:38:05,754 WARNING ppo.py:149 -- Using the simple minibatch optimizer. This will significantly reduce performance, consider simple_optimizer=False. DTFP: 4.604122s DTFP: 4.856234s DTFP: 5.630484s DTFP: 6.850456s DTFP: 7.856700s DTFP: 9.624164s DTFP: 10.894944s DTFP: 12.129192s DTFP: 14.210247s DTFP: 15.342738s ``` Line 130 in `tf_policy_template.py` (modified to print debug output above) ```python t = time.time() DynamicTFPolicy.__init__( self, obs_space, action_space, config, loss_fn, stats_fn=stats_fn, grad_stats_fn=grad_stats_fn, before_loss_init=before_loss_init_wrapper, make_model=make_model, action_sampler_fn=action_sampler_fn, existing_model=existing_model, existing_inputs=existing_inputs, get_batch_divisibility_req=get_batch_divisibility_req, obs_include_prev_action_reward=obs_include_prev_action_reward) print("DTFP: %fs" % (time.time() - t)) ``` Snippet of trainer script used. ```python # pylint: disable=invalid-name if __name__ == "__main__": ray.init() # Get ``settings`` file for now. settings_file = sys.argv[1] with open(settings_file, "r") as f: settings = json.load(f) env_config = settings["env"] time_steps = env_config["time_steps"] space_env = create_env(settings) env = create_env(settings) # Register environment register_env("world", lambda _: env) # Build environment instance to get ``obs_space``. obs_space = space_env.observation_space act_space = space_env.action_space # You can also have multiple policies per trainer, but here we just # show one each for PPO and DQN. policies: Dict[str, Tuple[Any, gym.Space, gym.Space, Dict[Any, Any]]] = { "0": (PPOTFPolicy, obs_space, act_space, {}), "1": (PPOTFPolicy, obs_space, act_space, {}), "2": (PPOTFPolicy, obs_space, act_space, {}), "3": (PPOTFPolicy, obs_space, act_space, {}), "4": (PPOTFPolicy, obs_space, act_space, {}), "5": (PPOTFPolicy, obs_space, act_space, {}), "6": (PPOTFPolicy, obs_space, act_space, {}), "7": (PPOTFPolicy, obs_space, act_space, {}), "8": (PPOTFPolicy, obs_space, act_space, {}), "9": (PPOTFPolicy, obs_space, act_space, {}), } def policy_mapping_fn(agent_id: int) -> str: """ Returns the given agent's policy identifier. """ return str(agent_id) ppo_trainer = PPOTrainer( env="bee_world", config={ "multiagent": { "policies": policies, "policy_mapping_fn": policy_mapping_fn, "policies_to_train": ["0", "1", "2", "3", "4", "5", "6", "7", "8", "9"], }, "simple_optimizer": True, # Disable filters, otherwise we would need to synchronize those # as well to the DQN agent. "observation_filter": "NoFilter", "num_workers": 2, "num_gpus": 1, "train_batch_size": 2, "sample_batch_size": 1, "sgd_minibatch_size": 2, }, ) ```
I am a collaborator on the project with @brendanxwhitaker, and I wanted to add another peculiar thing about the initialization code in DynamicTFPolicy. As pointed out above, line 291 of `rllib/policy/dynamic_tf_policy.py` is ```python self._sess.run(tf.global_variables_initializer()) ``` However, line 350 of the same initialization method in the same file is the exact same line, so that initialization runs twice. This is briefly touched on in a comment on line 290 ```python # postprocessing might depend on variable init, so run it first here ``` This is obviously a bit redundant but seems necessary unless something else in the method changes. My question about this is, under what conditions is the postprocessing dependent on the variable initialization, and how hard would it be to change this method so that the initialization is not redundant? Thanks for looking into this... agree we should try to fix the redundant initialization. One possible solution may be to initialize uninitialized variables. It looks like TF doesn't have a way to do this out of the box, but there are some workarounds: https://stackoverflow.com/questions/35164529/in-tensorflow-is-there-any-way-to-just-initialize-uninitialised-variables Maybe one of these would work? > This is obviously a bit redundant but seems necessary unless something else in the method changes. My question about this is, under what conditions is the postprocessing dependent on the variable initialization, and how hard would it be to change this method so that the initialization is not redundant? This happens in for example PPO, where the postprocessor is accessing the value function. If you comment it out, then PPO will crash during policy creation trying to read uninitialized weights. Okay, so just to be clear, when you say redundant, are you referring to calling `tf.global_variables_initializer` instead of only initializing the uninitialized variables, or are you referring to there being two of those calls for every call to the init function? >Maybe one of these would work? That seems like it would. Do you think it's possible to just create a variable scope that only contains the new, uninitialized variables and then just initialize all of those? Or maybe just add them to a list as they're created and pass a reference back up to the init function? Also, is it expected behavior for the initialization time to be growing with each new policy added? Is this something unique to PPO, or is it going to be observed no matter what algorithm we use? Are there other algorithms better suited for large numbers of individual policies running on a single machine? The former (which includes the latter, though double init is less of an issue). On Fri, Oct 25, 2019, 5:11 AM Brendan Whitaker <[email protected]> wrote: > Okay, so just to be clear, when you say redundant, are you referring to > calling tf.global_variables_initializer and initializing everything on > each of those calls, or are you referring to there being two of those calls > for every call to the init function? > > — > You are receiving this because you were assigned. > Reply to this email directly, view it on GitHub > <https://github.com/ray-project/ray/issues/5982?email_source=notifications&email_token=AAADUSWGLFRZ2T4UX4BPOODQQLO6JA5CNFSM4JEA3DT2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOECIFA6Y#issuecomment-546328699>, > or unsubscribe > <https://github.com/notifications/unsubscribe-auth/AAADUSS6IPZ3UYTEBZEYLVLQQLO6JANCNFSM4JEA3DTQ> > . > Looking through the stack overflow page, none of those answers are super clean but one of them is not as slow, with only one sess.run to find which variables are uninitialized. We are going to try a workaround and see how the performance is. Also, I am curious about what Brendan asked about regarding the initialization time growing. Do you know if this is something we should expect from initialization, that the initialization time will be larger for the last agents initialized than for the first, even if we implement one of the workarounds above? Thanks. @mtcrawshaw I don't think it should be slowing down over time once the redundant inits are removed, if so that might be a bug in TensorFlow. So we have implemented a couple of the workarounds from the stack overflow page that you mentioned, and we noticed some weird behavior from RLlib that we wanted to ask you about. In policy/dynamic_tf_policy.py, we replaced ```python self._sess.run(tf.global_variables_initializer()) ``` on lines 291 and 350 with ```python initialize_uninitialized(self._sess) ``` where ``initialize_uninitialized`` defined by ```python def initialize_uninitialized(sess): bytestr_var_names = sess.run(tf.report_uninitialized_variables()) global_vars = tf.global_variables() strnames = [name.decode("utf-8") for name in bytestr_var_names] uninit_vars = [ v for v in tf.global_variables() if v.name.split(":")[0] in set(strnames) ] initializer = tf.variables_initializer(uninit_vars) sess.run(initializer) ``` However, we still were getting an increase in the initialization time as the number of initialized agents grew. We threw a couple of print statements in ``initialize_uninitialized`` to get the names and number of uninitialized variables, modified version below. ```python def initialize_uninitialized(sess): bytestr_var_names = sess.run(tf.report_uninitialized_variables()) global_vars = tf.global_variables() strnames = [name.decode("utf-8") for name in bytestr_var_names] uninit_vars = [ v for v in tf.global_variables() if v.name.split(":")[0] in set(strnames) ] print("Names of uninitialized variables: %s" % str([v.name for v in uninit_vars])) print("Number of uninitvars: %d" % len(uninit_vars)) print("Number of names from report: %d" % len(bytestr_var_names)) print("Number of variables: %d" % len(global_vars)) initializer = tf.variables_initializer(uninit_vars) sess.run(initializer) ``` I won't provide the output here because it is quite long, but it shows that the variables which get initialized during the initialization of the first agent show up as uninitialized during the initialization of the second agent! The variables corresponding to the first and second agent are then initialized, but during the initialization of the third agent, they again reappear as uninitialized. My question is, do you have any idea why these variables might be getting "de-initialized" at any point during the initialization process? That's really weird, perhaps try printing out the current TF graph id? Maybe there are new graphs being created each time. I can't think of any code that would be uninitializing variables otherwise. On Tue, Oct 29, 2019, 5:57 AM Michael Crawshaw <[email protected]> wrote: > So we have implemented a couple of the workarounds from the stack overflow > page that you mentioned, and we noticed some weird behavior from RLlib that > we wanted to ask you about. In policy/dynamic_tf_policy.py, we replaced > > self._sess.run(tf.global_variables_initializer()) > > with a call to initialize_uninitialized(self._sess), where > initialize_uninitialized defined by > > def initialize_uninitialized(sess): > bytestr_var_names = sess.run(tf.report_uninitialized_variables()) > global_vars = tf.global_variables() > strnames = [name.decode("utf-8") for name in bytestr_var_names] > uninit_vars = [ > v > for v in tf.global_variables() > if v.name.split(":")[0] > in set(strnames) > ] > initializer = tf.variables_initializer(uninit_vars) > sess.run(initializer) > > However, we still were getting an increase in the initialization time as > the number of initialized agents grew. We threw a couple of print > statements in initialize_uninitialized to get the names and number of > uninitialized variables, modified version below. > > def initialize_uninitialized(sess): > bytestr_var_names = sess.run(tf.report_uninitialized_variables()) > global_vars = tf.global_variables() > strnames = [name.decode("utf-8") for name in bytestr_var_names] > uninit_vars = [ > v > for v in tf.global_variables() > if v.name.split(":")[0] > in set(strnames) > ] > print("Names of uninitialized variables: %s" % str([v.name for v in uninit_vars])) > print("Number of uninitvars: %d" % len(uninit_vars)) > print("Number of names from report: %d" % len(bytestr_var_names)) > print("Number of variables: %d" % len(global_vars)) > initializer = tf.variables_initializer(uninit_vars) > sess.run(initializer) > > I won't provide the output here because it is quite long, but it shows > that the variables which get initialized during the initialization of the > first agent show up as uninitialized during the initialization of the > second agent! The variables corresponding to the first and second agent are > then initialized, but during the initialization of the third agent, they > again reappear as uninitialized. My question is, do you have any idea why > these variables might be getting "de-initialized" at any point during the > initialization process? > > — > You are receiving this because you were assigned. > Reply to this email directly, view it on GitHub > <https://github.com/ray-project/ray/issues/5982?email_source=notifications&email_token=AAADUSUTNO3MSFXURACF44TQRAXM3A5CNFSM4JEA3DT2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOECQMCAA#issuecomment-547406080>, > or unsubscribe > <https://github.com/notifications/unsubscribe-auth/AAADUSQ56CLKJAL5S577XUTQRAXM3ANCNFSM4JEA3DTQ> > . > Alright, so last week we were actually running that in a fork of RLLib. We ran the same tests in a fresh install, and found that the initialization time still grows, but that the number of uninitialized variables is constant (normal). We think this is entirely because of the (3?) calls to the global variable initializer. We'll investigate the graph id thing. I'm running into the same situation. As the agent initialization time grows linearly per-agent, this means that there is superlinear growth in total initialization time as the number of agents increases. My agents all use the same type of custom policy, but each have their own instance of that policy. Although a single part of the graph does grow exponentially with the amount of agents, it is insignificant compared to the rest of of the parameters: a change of <20k parameters versus hundreds of thousands of other fixed parameters. I ran cprofile/kcachegrind with `local_mode=True` to take a peek. - 2 agents: Initialization takes 44 seconds - 4 agents: Initialization takes 172 seconds - 5 agents: Initialization takes 275 seconds Respectively, their call graphs look like this: ![2_agent_profile](https://user-images.githubusercontent.com/5096835/81180011-be693680-8faa-11ea-9c6b-ff19aad3e788.png) ![4_agent_profile](https://user-images.githubusercontent.com/5096835/81180013-bf01cd00-8faa-11ea-9307-ff97c8dd53ef.png) ![5_agent_profile](https://user-images.githubusercontent.com/5096835/81180014-bf9a6380-8faa-11ea-860c-6d8960787fee.png) As the number of agents grows, the 3 functions that dominate are: 1. _pywrap_tensorflow_internal.TF_SessionRun_wrapper 2. _pywrap_tensorflow_internal.ExtendSession 3. _pywrap_tensorflow_internal.TF_SessionMakeCallable If anyone wants to dig into this, check out these [3 profiler files](https://github.com/ray-project/ray/files/4586880/profiler_files.zip). You can start one in KCachegrind with the following command: `pyprof2calltree -i prof-2.out -k` I experimented a bit, and I've found [this line](https://github.com/ray-project/ray/blob/master/rllib/optimizers/multi_gpu_optimizer.py#L139) `self.workers.local_worker().get_weights()` to be the slowest (makes up for 80-90% of initialization time with 5 agents), but only when it is called for the first time. Subsequent calls are far, far faster. Somewhere down the call stack Ray then calls [this function](https://github.com/ray-project/ray/blob/master/python/ray/experimental/tf_utils.py#L157) for each policy, for each element of that policy. The tf graph is the same size for each agent, and grows linearly (in my case, 2 agents/80k elements, 5 agents/200k elements). Thus, you get a linearly increasing amount of eval calls to a linearly increasingly large graph = exponential growth. I'm not sure whether the graph size increase is intentional: each policy is after all evaluated independently, as shown in these two lines in sampler.py {[1](https://github.com/ray-project/ray/blob/releases/0.8.5/rllib/evaluation/sampler.py#L590), [2](https://github.com/ray-project/ray/blob/releases/0.8.5/rllib/evaluation/sampler.py#L613)}. Is this done so that optimization is faster? Here's a graph for memory use while evaluating/training on a small amount of steps (30k-ish) with respectively `[2,3,4,5]` agents. The longest flat lines in each segment are where get_weights() is being called. ![num_agents](https://user-images.githubusercontent.com/5096835/82066727-181ae080-96d0-11ea-8c28-17179c37423c.png)
2020-05-18T21:18:56
ray-project/ray
8,493
ray-project__ray-8493
[ "8482" ]
aa1cbe8abc845e1a6acf2f237b4b00aa7c3f295d
diff --git a/python/ray/__init__.py b/python/ray/__init__.py --- a/python/ray/__init__.py +++ b/python/ray/__init__.py @@ -9,9 +9,13 @@ # raylet modules. if "pickle5" in sys.modules: - raise ImportError("Ray must be imported before pickle5 because Ray " - "requires a specific version of pickle5 (which is " - "packaged along with Ray).") + import pkg_resources + version_info = pkg_resources.require("pickle5") + version = tuple(int(n) for n in version_info[0].version.split(".")) + if version < (0, 0, 10): + raise ImportError("You are using an old version of pickle5 that " + "leaks memory, please run 'pip install pickle5 -U' " + "to upgrade") if "OMP_NUM_THREADS" not in os.environ: logger.debug("[ray] Forcing OMP_NUM_THREADS=1 to avoid performance "
import error ray 0.8.5,when I try to use ray, it occurs Ray must be imported before pickle5 because Ray requires a specific version of pickle5 (which is packaged along with Ray. I want to know it must import pickle5 before import ray, Right?
yes, we are using an internal version of pickle5, because the current pickle5 has memory leak issues. so we have to import ray before importing the pickle5 library. I have created a PR in the upstream (pickle5-backport), it would address the problem once merged.
2020-05-18T21:44:03
ray-project/ray
8,511
ray-project__ray-8511
[ "8492" ]
f8f7efc24f77e4491bc41daa7fb800cd07e524b2
diff --git a/python/ray/worker.py b/python/ray/worker.py --- a/python/ray/worker.py +++ b/python/ray/worker.py @@ -1195,10 +1195,6 @@ def connect(node, worker.lock = threading.RLock() - # Create an object for interfacing with the global state. - ray.state.state._initialize_global_state( - node.redis_address, redis_password=node.redis_password) - driver_name = "" log_stdout_file_name = "" log_stderr_file_name = "" @@ -1269,6 +1265,12 @@ def connect(node, log_stderr_file_name, ) + # Create an object for interfacing with the global state. + # Note, global state should be intialized after `CoreWorker`, because it will + # use glog, which is intialized in `CoreWorker`. + ray.state.state._initialize_global_state( + node.redis_address, redis_password=node.redis_password) + if driver_object_store_memory is not None: worker.core_worker.set_object_store_client_options( "ray_driver_{}".format(os.getpid()), driver_object_store_memory)
Excess INFO messages to console in latest build <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? Master build logs this to console on ray init: ``` >>> (pid=raylet) WARNING: Logging before InitGoogleLogging() is written to STDERR (pid=raylet) I0518 14:41:09.151492 3131 global_state_accessor.cc:25] Redis server address = 192.168.5.121:6379, is test flag = 0 (pid=raylet) I0518 14:41:09.166244 3131 redis_client.cc:141] RedisClient connected. (pid=raylet) WARNING: Logging before InitGoogleLogging() is written to STDERR (pid=raylet) I0518 14:41:09.168131 3135 global_state_accessor.cc:25] Redis server address = 192.168.5.121:6379, is test flag = 0 (pid=raylet) I0518 14:41:09.174638 3131 redis_gcs_client.cc:88] RedisGcsClient Connected. (pid=raylet) I0518 14:41:09.175307 3131 service_based_gcs_client.cc:66] ServiceBasedGcsClient Connected. (pid=raylet) WARNING: Logging before InitGoogleLogging() is written to STDERR (pid=raylet) I0518 14:41:09.177665 3132 global_state_accessor.cc:25] Redis server address = 192.168.5.121:6379, is test flag = 0 (pid=raylet) I0518 14:41:09.178391 3135 redis_client.cc:141] RedisClient connected. (pid=raylet) I0518 14:41:09.181085 3132 redis_client.cc:141] RedisClient connected. (pid=raylet) I0518 14:41:09.187088 3135 redis_gcs_client.cc:88] RedisGcsClient Connected. (pid=raylet) I0518 14:41:09.187737 3135 service_based_gcs_client.cc:66] ServiceBasedGcsClient Connected. (pid=raylet) I0518 14:41:09.194645 3132 redis_gcs_client.cc:88] RedisGcsClient Connected. (pid=raylet) I0518 14:41:09.195237 3132 service_based_gcs_client.cc:66] ServiceBasedGcsClient Connected. (pid=raylet) WARNING: Logging before InitGoogleLogging() is written to STDERR (pid=raylet) I0518 14:41:09.226564 3133 global_state_accessor.cc:25] Redis server address = 192.168.5.121:6379, is test flag = 0 (pid=raylet) I0518 14:41:09.230108 3133 redis_client.cc:141] RedisClient connected. (pid=raylet) I0518 14:41:09.238509 3133 redis_gcs_client.cc:88] RedisGcsClient Connected. (pid=raylet) I0518 14:41:09.239135 3133 service_based_gcs_client.cc:66] ServiceBasedGcsClient Connected. (pid=raylet) WARNING: Logging before InitGoogleLogging() is written to STDERR (pid=raylet) I0518 14:41:09.279531 3136 global_state_accessor.cc:25] Redis server address = 192.168.5.121:6379, is test flag = 0 (pid=raylet) I0518 14:41:09.280068 3136 redis_client.cc:141] RedisClient connected. (pid=raylet) I0518 14:41:09.288298 3136 redis_gcs_client.cc:88] RedisGcsClient Connected. (pid=raylet) I0518 14:41:09.288733 3136 service_based_gcs_client.cc:66] ServiceBasedGcsClient Connected. (pid=raylet) WARNING: Logging before InitGoogleLogging() is written to STDERR (pid=raylet) I0518 14:41:09.300333 3130 global_state_accessor.cc:25] Redis server address = 192.168.5.121:6379, is test flag = 0 (pid=raylet) I0518 14:41:09.301705 3130 redis_client.cc:141] RedisClient connected. (pid=raylet) I0518 14:41:09.309738 3130 redis_gcs_client.cc:88] RedisGcsClient Connected. (pid=raylet) I0518 14:41:09.310228 3130 service_based_gcs_client.cc:66] ServiceBasedGcsClient Connected. (pid=raylet) WARNING: Logging before InitGoogleLogging() is written to STDERR (pid=raylet) I0518 14:41:09.316042 3134 global_state_accessor.cc:25] Redis server address = 192.168.5.121:6379, is test flag = 0 (pid=raylet) I0518 14:41:09.317946 3134 redis_client.cc:141] RedisClient connected. (pid=raylet) I0518 14:41:09.326037 3134 redis_gcs_client.cc:88] RedisGcsClient Connected. (pid=raylet) I0518 14:41:09.326454 3134 service_based_gcs_client.cc:66] ServiceBasedGcsClient Connected. (pid=raylet) WARNING: Logging before InitGoogleLogging() is written to STDERR (pid=raylet) I0518 14:41:09.357240 3137 global_state_accessor.cc:25] Redis server address = 192.168.5.121:6379, is test flag = 0 (pid=raylet) I0518 14:41:09.357900 3137 redis_client.cc:141] RedisClient connected. (pid=raylet) I0518 14:41:09.366128 3137 redis_gcs_client.cc:88] RedisGcsClient Connected. (pid=raylet) I0518 14:41:09.366888 3137 service_based_gcs_client.cc:66] ServiceBasedGcsClient Connected. ```
@ffbin is this related to your recent change? @ericl This log is related to this PR. https://github.com/ray-project/ray/pull/8401 I will take a look, thanks.
2020-05-20T03:31:57
ray-project/ray
8,521
ray-project__ray-8521
[ "5821" ]
6b53df95990bfac235313f24f80176948339b606
diff --git a/python/ray/tune/examples/mlflow_example.py b/python/ray/tune/examples/mlflow_example.py --- a/python/ray/tune/examples/mlflow_example.py +++ b/python/ray/tune/examples/mlflow_example.py @@ -41,7 +41,9 @@ def easy_objective(config): num_samples=5, loggers=DEFAULT_LOGGERS + (MLFLowLogger, ), config={ - "mlflow_experiment_id": experiment_id, + "logger_config": { + "mlflow_experiment_id": experiment_id, + }, "width": tune.sample_from( lambda spec: 10 + int(90 * random.random())), "height": tune.sample_from(lambda spec: int(100 * random.random())) diff --git a/python/ray/tune/logger.py b/python/ray/tune/logger.py --- a/python/ray/tune/logger.py +++ b/python/ray/tune/logger.py @@ -78,9 +78,12 @@ class MLFLowLogger(Logger): """ def _init(self): + logger_config = self.config.get("logger_config", {}) from mlflow.tracking import MlflowClient - client = MlflowClient() - run = client.create_run(self.config.get("mlflow_experiment_id")) + client = MlflowClient( + tracking_uri=logger_config.get("mlflow_tracking_uri"), + registry_uri=logger_config.get("mlflow_registry_uri")) + run = client.create_run(logger_config.get("mlflow_experiment_id")) self._run_id = run.info.run_id for key, value in self.config.items(): client.log_param(self._run_id, key, value) diff --git a/rllib/agents/trainer.py b/rllib/agents/trainer.py --- a/rllib/agents/trainer.py +++ b/rllib/agents/trainer.py @@ -370,6 +370,10 @@ "replay_mode": "independent", }, + # === Logger === + # Define logger-specific configuration to be used inside Logger + "logger_config": {}, + # === Replay Settings === # The number of contiguous environment steps to replay at once. This may # be set to greater than 1 to support recurrent models.
MLFlow parameter error using rllib agents <!-- General questions should be asked on the mailing list [email protected]. Questions about how to use Ray should be asked on [StackOverflow](https://stackoverflow.com/questions/tagged/ray). Before submitting an issue, please fill out the following form. --> ### System information - **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**: MacOS 10.14.6 - **Ray installed from (source or binary)**: source - **Ray version**: 0.7.5 - **Python version**: 3.7.3 - **Exact command to reproduce**: <!-- You can obtain the Ray version with python -c "import ray; print(ray.__version__)" --> ### Describe the problem <!-- Describe the problem clearly here. --> When I tried to integrate MLFlow to my experiment that uses a DQN agent, I get the following error: ``` Trainable.__init__(self, config, logger_creator) File "/python3.7/site-packages/ray/tune/trainable.py", line 96, in __init__ self._setup(copy.deepcopy(self.config)) File "/python3.7/site-packages/ray/rllib/agents/trainer.py", line 478, in _setup self._allow_unknown_subkeys) File "/python3.7/site-packages/ray/tune/util.py", line 160, in deep_update raise Exception("Unknown config parameter `{}` ".format(k)) Exception: Unknown config parameter mlflow_experiment_id ``` It seems like the agent's config is quite frozen, and there doesn't seem to be an explicit way to allow new keys to that config. When I digged a bit in the code, I found, in the `rllib/agents/trainer.py` that there's an argument to the deep_update function that merges configs together called `allow_unknown_configs` and that is by default set to False. When I set it to True, for testing purposes, it worked. So maybe we should add an abstraction layer to that so that it can easily be set from the experiment itself. ### Source code / logs This is part of the source code of when I ran the experiment on a DQN agent. I got the error above ``` tune.run( ... config={ "mlflow_experiment_id": experiment_id, ... ) ```
I have a similar, but not MLflow related, problem. The basic question is the same: How do you pass arguments via the config if the agent does not allow additional config values. Maybe an abstraction layer to allow additional parameters makes sense. Or maybe allowing a field like `"custom_config"` in the config dict should be allowed all the time. I have the same issue, I have to hack the code in trainer.py put `_allow_unknown_configs=True`. Only in this way I am able to use MLflow @fatenghriss
2020-05-20T14:02:19
ray-project/ray
8,533
ray-project__ray-8533
[ "8523", "8523" ]
d27e6da1b291887707533cc10f066cd0e888607b
diff --git a/rllib/agents/qmix/qmix_policy.py b/rllib/agents/qmix/qmix_policy.py --- a/rllib/agents/qmix/qmix_policy.py +++ b/rllib/agents/qmix/qmix_policy.py @@ -13,11 +13,13 @@ from ray.rllib.models.catalog import ModelCatalog from ray.rllib.models.modelv2 import _unpack_obs from ray.rllib.env.constants import GROUP_REWARDS +from ray.rllib.utils import try_import_tree from ray.rllib.utils.framework import try_import_torch from ray.rllib.utils.annotations import override # Torch must be installed. torch, nn = try_import_torch(error=True) +tree = try_import_tree() logger = logging.getLogger(__name__) @@ -463,25 +465,28 @@ def _unpack_observation(self, obs_batch): state (np.ndarray or None): state tensor of shape [B, state_size] or None if it is not in the batch """ + unpacked = _unpack_obs( np.array(obs_batch, dtype=np.float32), self.observation_space.original_space, tensorlib=np) + + if isinstance(unpacked[0], dict): + unpacked_obs = [ + np.concatenate(tree.flatten(u["obs"]), 1) for u in unpacked + ] + else: + unpacked_obs = unpacked + + obs = np.concatenate( + unpacked_obs, + axis=1).reshape([len(obs_batch), self.n_agents, self.obs_size]) + if self.has_action_mask: - obs = np.concatenate( - [o["obs"] for o in unpacked], - axis=1).reshape([len(obs_batch), self.n_agents, self.obs_size]) action_mask = np.concatenate( [o["action_mask"] for o in unpacked], axis=1).reshape( [len(obs_batch), self.n_agents, self.n_actions]) else: - if isinstance(unpacked[0], dict): - unpacked_obs = [u["obs"] for u in unpacked] - else: - unpacked_obs = unpacked - obs = np.concatenate( - unpacked_obs, - axis=1).reshape([len(obs_batch), self.n_agents, self.obs_size]) action_mask = np.ones( [len(obs_batch), self.n_agents, self.n_actions], dtype=np.float32)
diff --git a/rllib/tests/test_avail_actions_qmix.py b/rllib/tests/test_avail_actions_qmix.py --- a/rllib/tests/test_avail_actions_qmix.py +++ b/rllib/tests/test_avail_actions_qmix.py @@ -1,4 +1,4 @@ -from gym.spaces import Tuple, Discrete, Dict, Box +from gym.spaces import Box, Dict, Discrete, MultiDiscrete, Tuple import numpy as np import unittest @@ -9,10 +9,17 @@ class AvailActionsTestEnv(MultiAgentEnv): - action_space = Discrete(10) + num_actions = 10 + action_space = Discrete(num_actions) observation_space = Dict({ - "obs": Discrete(3), - "action_mask": Box(0, 1, (10, )), + "obs": Dict({ + "test": Dict({ + "a": Discrete(2), + "b": MultiDiscrete([2, 3, 4]) + }), + "state": MultiDiscrete([2, 2, 2]) + }), + "action_mask": Box(0, 1, (num_actions, )), }) def __init__(self, env_config): @@ -25,7 +32,7 @@ def reset(self): self.state = 0 return { "agent_1": { - "obs": self.state, + "obs": self.observation_space["obs"].sample(), "action_mask": self.action_mask } } @@ -36,7 +43,12 @@ def step(self, action_dict): "Failed to obey available actions mask!" self.state += 1 rewards = {"agent_1": 1} - obs = {"agent_1": {"obs": 0, "action_mask": self.action_mask}} + obs = { + "agent_1": { + "obs": self.observation_space["obs"].sample(), + "action_mask": self.action_mask + } + } dones = {"__all__": self.state > 20} return obs, rewards, dones, {}
[rllib] Make QMix support complex observation spaces <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### Describe your feature request The QMix example and the default model only uses a simple MultiDiscrete observation space. It would be nice if more abstract observation spaces and custom models would be supported. The following example using a nested dict observation space ([based on the two step game](https://github.com/ray-project/ray/blob/master/rllib/examples/env/two_step_game.py)) fails [here](https://github.com/ray-project/ray/blob/master/rllib/agents/qmix/qmix_policy.py#L482) (I am aware that the model would have to be adapted as well, but that does not seem to be the cause of this error): ``` import argparse from gym.spaces import Tuple, MultiDiscrete, Dict, Discrete import ray from ray import tune from ray.tune import register_env, grid_search from ray.rllib.env.multi_agent_env import ENV_STATE from ray.rllib.examples.env.two_step_game import TwoStepGame from ray.rllib.utils.test_utils import check_learning_achieved from gym.spaces import MultiDiscrete, Dict, Discrete import numpy as np from ray.rllib.env.multi_agent_env import MultiAgentEnv, ENV_STATE class TwoStepGame(MultiAgentEnv): action_space = Discrete(2) def __init__(self, env_config): self.state = None self.agent_1 = 0 self.agent_2 = 1 # MADDPG emits action logits instead of actual discrete actions self.actions_are_logits = env_config.get("actions_are_logits", False) self.one_hot_state_encoding = env_config.get("one_hot_state_encoding", False) self.with_state = env_config.get("separate_state_space", False) if not self.one_hot_state_encoding: self.observation_space = Discrete(6) self.with_state = False else: # Each agent gets the full state (one-hot encoding of which of the # three states are active) as input with the receiving agent's # ID (1 or 2) concatenated onto the end. if self.with_state: self.observation_space = Dict({ "obs": MultiDiscrete([2, 2, 2, 3]), ENV_STATE: MultiDiscrete([2, 2, 2]) }) else: self.observation_space = MultiDiscrete([2, 2, 2, 3]) def reset(self): self.state = np.array([1, 0, 0]) return self._obs() def step(self, action_dict): if self.actions_are_logits: action_dict = { k: np.random.choice([0, 1], p=v) for k, v in action_dict.items() } state_index = np.flatnonzero(self.state) if state_index == 0: action = action_dict[self.agent_1] assert action in [0, 1], action if action == 0: self.state = np.array([0, 1, 0]) else: self.state = np.array([0, 0, 1]) global_rew = 0 done = False elif state_index == 1: global_rew = 7 done = True else: if action_dict[self.agent_1] == 0 and action_dict[self. agent_2] == 0: global_rew = 0 elif action_dict[self.agent_1] == 1 and action_dict[self. agent_2] == 1: global_rew = 8 else: global_rew = 1 done = True rewards = { self.agent_1: global_rew / 2.0, self.agent_2: global_rew / 2.0 } obs = self._obs() dones = {"__all__": done} infos = {} return obs, rewards, dones, infos def _obs(self): if self.with_state: return { self.agent_1: { "obs": { "data": self.agent_1_obs(), "test": 1 }, ENV_STATE: self.state }, self.agent_2: { "obs": { "data": self.agent_2_obs(), "test": 2 }, ENV_STATE: self.state } } else: return { self.agent_1: self.agent_1_obs(), self.agent_2: self.agent_2_obs() } def agent_1_obs(self): if self.one_hot_state_encoding: return np.concatenate([self.state, [1]]) else: return np.flatnonzero(self.state)[0] def agent_2_obs(self): if self.one_hot_state_encoding: return np.concatenate([self.state, [2]]) else: return np.flatnonzero(self.state)[0] + 3 if __name__ == "__main__": grouping = { "group_1": [0, 1], } obs_space = Tuple([ Dict({ "obs": Dict({ 'data': MultiDiscrete([2, 2, 2, 3]), 'test': Discrete(5) }), ENV_STATE: MultiDiscrete([2, 2, 2]) }), Dict({ "obs": Dict({ 'data': MultiDiscrete([2, 2, 2, 3]), 'test': Discrete(5) }), ENV_STATE: MultiDiscrete([2, 2, 2]) }), ]) act_space = Tuple([ TwoStepGame.action_space, TwoStepGame.action_space, ]) register_env( "grouped_twostep", lambda config: TwoStepGame(config).with_agent_groups( grouping, obs_space=obs_space, act_space=act_space)) ray.init() results = tune.run("QMIX", config={ "rollout_fragment_length": 4, "train_batch_size": 32, "exploration_fraction": .4, "exploration_final_eps": 0.0, "num_workers": 0, "mixer": "qmix", "env_config": { "separate_state_space": True, "one_hot_state_encoding": True }, "use_pytorch": True, "env": "grouped_twostep" }) ray.shutdown() ``` ``` 2020-05-20 15:40:03,477 ERROR trial_runner.py:519 -- Trial QMIX_grouped_twostep_c960c_00002: Error processing event. Traceback (most recent call last): File "[...]/ray/tune/trial_runner.py", line 467, in _process_trial result = self.trial_executor.fetch_result(trial) File "[...]/ray/tune/ray_trial_executor.py", line 430, in fetch_result result = ray.get(trial_future[0], DEFAULT_GET_TIMEOUT) File "[...]/ray/worker.py", line 1516, in get raise value.as_instanceof_cause() ray.exceptions.RayTaskError(ValueError): ray::QMIX.train() (pid=54834, ip=128.232.69.20) File "python/ray/_raylet.pyx", line 460, in ray._raylet.execute_task File "python/ray/_raylet.pyx", line 414, in ray._raylet.execute_task.function_executor File "[...]/ray/rllib/agents/trainer.py", line 504, in train raise e File "[...]/ray/rllib/agents/trainer.py", line 490, in train result = Trainable.train(self) File "[...]/ray/tune/trainable.py", line 260, in train result = self._train() File "[...]/ray/rllib/agents/trainer_template.py", line 138, in _train return self._train_exec_impl() File "[...]/ray/rllib/agents/trainer_template.py", line 173, in _train_exec_impl res = next(self.train_exec_impl) File "[...]/ray/util/iter.py", line 689, in __next__ return next(self.built_iterator) File "[...]/ray/util/iter.py", line 702, in apply_foreach for item in it: File "[...]/ray/util/iter.py", line 772, in apply_filter for item in it: File "[...]/ray/util/iter.py", line 772, in apply_filter for item in it: File "[...]/ray/util/iter.py", line 702, in apply_foreach for item in it: File "[...]/ray/util/iter.py", line 772, in apply_filter for item in it: File "[...]/ray/util/iter.py", line 977, in build_union item = next(it) File "[...]/ray/util/iter.py", line 689, in __next__ return next(self.built_iterator) File "[...]/ray/util/iter.py", line 702, in apply_foreach for item in it: File "[...]/ray/util/iter.py", line 702, in apply_foreach for item in it: File "[...]/ray/util/iter.py", line 702, in apply_foreach for item in it: File "[...]/ray/rllib/execution/rollout_ops.py", line 70, in sampler yield workers.local_worker().sample() File "[...]/ray/rllib/evaluation/rollout_worker.py", line 515, in sample batches = [self.input_reader.next()] File "[...]/ray/rllib/evaluation/sampler.py", line 56, in next batches = [self.get_data()] File "[...]/ray/rllib/evaluation/sampler.py", line 101, in get_data item = next(self.rollout_provider) File "[...]/ray/rllib/evaluation/sampler.py", line 367, in _env_runner active_episodes) File "[...]/ray/rllib/evaluation/sampler.py", line 637, in _do_policy_eval timestep=policy.global_timestep) File "[...]/ray/rllib/agents/qmix/qmix_policy.py", line 260, in compute_actions obs_batch, action_mask, _ = self._unpack_observation(obs_batch) File "[...]/ray/rllib/agents/qmix/qmix_policy.py", line 486, in _unpack_observation axis=1).reshape([len(obs_batch), self.n_agents, self.obs_size]) File "<__array_function__ internals>", line 6, in concatenate ValueError: zero-dimensional arrays cannot be concatenated ``` [rllib] Make QMix support complex observation spaces <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### Describe your feature request The QMix example and the default model only uses a simple MultiDiscrete observation space. It would be nice if more abstract observation spaces and custom models would be supported. The following example using a nested dict observation space ([based on the two step game](https://github.com/ray-project/ray/blob/master/rllib/examples/env/two_step_game.py)) fails [here](https://github.com/ray-project/ray/blob/master/rllib/agents/qmix/qmix_policy.py#L482) (I am aware that the model would have to be adapted as well, but that does not seem to be the cause of this error): ``` import argparse from gym.spaces import Tuple, MultiDiscrete, Dict, Discrete import ray from ray import tune from ray.tune import register_env, grid_search from ray.rllib.env.multi_agent_env import ENV_STATE from ray.rllib.examples.env.two_step_game import TwoStepGame from ray.rllib.utils.test_utils import check_learning_achieved from gym.spaces import MultiDiscrete, Dict, Discrete import numpy as np from ray.rllib.env.multi_agent_env import MultiAgentEnv, ENV_STATE class TwoStepGame(MultiAgentEnv): action_space = Discrete(2) def __init__(self, env_config): self.state = None self.agent_1 = 0 self.agent_2 = 1 # MADDPG emits action logits instead of actual discrete actions self.actions_are_logits = env_config.get("actions_are_logits", False) self.one_hot_state_encoding = env_config.get("one_hot_state_encoding", False) self.with_state = env_config.get("separate_state_space", False) if not self.one_hot_state_encoding: self.observation_space = Discrete(6) self.with_state = False else: # Each agent gets the full state (one-hot encoding of which of the # three states are active) as input with the receiving agent's # ID (1 or 2) concatenated onto the end. if self.with_state: self.observation_space = Dict({ "obs": MultiDiscrete([2, 2, 2, 3]), ENV_STATE: MultiDiscrete([2, 2, 2]) }) else: self.observation_space = MultiDiscrete([2, 2, 2, 3]) def reset(self): self.state = np.array([1, 0, 0]) return self._obs() def step(self, action_dict): if self.actions_are_logits: action_dict = { k: np.random.choice([0, 1], p=v) for k, v in action_dict.items() } state_index = np.flatnonzero(self.state) if state_index == 0: action = action_dict[self.agent_1] assert action in [0, 1], action if action == 0: self.state = np.array([0, 1, 0]) else: self.state = np.array([0, 0, 1]) global_rew = 0 done = False elif state_index == 1: global_rew = 7 done = True else: if action_dict[self.agent_1] == 0 and action_dict[self. agent_2] == 0: global_rew = 0 elif action_dict[self.agent_1] == 1 and action_dict[self. agent_2] == 1: global_rew = 8 else: global_rew = 1 done = True rewards = { self.agent_1: global_rew / 2.0, self.agent_2: global_rew / 2.0 } obs = self._obs() dones = {"__all__": done} infos = {} return obs, rewards, dones, infos def _obs(self): if self.with_state: return { self.agent_1: { "obs": { "data": self.agent_1_obs(), "test": 1 }, ENV_STATE: self.state }, self.agent_2: { "obs": { "data": self.agent_2_obs(), "test": 2 }, ENV_STATE: self.state } } else: return { self.agent_1: self.agent_1_obs(), self.agent_2: self.agent_2_obs() } def agent_1_obs(self): if self.one_hot_state_encoding: return np.concatenate([self.state, [1]]) else: return np.flatnonzero(self.state)[0] def agent_2_obs(self): if self.one_hot_state_encoding: return np.concatenate([self.state, [2]]) else: return np.flatnonzero(self.state)[0] + 3 if __name__ == "__main__": grouping = { "group_1": [0, 1], } obs_space = Tuple([ Dict({ "obs": Dict({ 'data': MultiDiscrete([2, 2, 2, 3]), 'test': Discrete(5) }), ENV_STATE: MultiDiscrete([2, 2, 2]) }), Dict({ "obs": Dict({ 'data': MultiDiscrete([2, 2, 2, 3]), 'test': Discrete(5) }), ENV_STATE: MultiDiscrete([2, 2, 2]) }), ]) act_space = Tuple([ TwoStepGame.action_space, TwoStepGame.action_space, ]) register_env( "grouped_twostep", lambda config: TwoStepGame(config).with_agent_groups( grouping, obs_space=obs_space, act_space=act_space)) ray.init() results = tune.run("QMIX", config={ "rollout_fragment_length": 4, "train_batch_size": 32, "exploration_fraction": .4, "exploration_final_eps": 0.0, "num_workers": 0, "mixer": "qmix", "env_config": { "separate_state_space": True, "one_hot_state_encoding": True }, "use_pytorch": True, "env": "grouped_twostep" }) ray.shutdown() ``` ``` 2020-05-20 15:40:03,477 ERROR trial_runner.py:519 -- Trial QMIX_grouped_twostep_c960c_00002: Error processing event. Traceback (most recent call last): File "[...]/ray/tune/trial_runner.py", line 467, in _process_trial result = self.trial_executor.fetch_result(trial) File "[...]/ray/tune/ray_trial_executor.py", line 430, in fetch_result result = ray.get(trial_future[0], DEFAULT_GET_TIMEOUT) File "[...]/ray/worker.py", line 1516, in get raise value.as_instanceof_cause() ray.exceptions.RayTaskError(ValueError): ray::QMIX.train() (pid=54834, ip=128.232.69.20) File "python/ray/_raylet.pyx", line 460, in ray._raylet.execute_task File "python/ray/_raylet.pyx", line 414, in ray._raylet.execute_task.function_executor File "[...]/ray/rllib/agents/trainer.py", line 504, in train raise e File "[...]/ray/rllib/agents/trainer.py", line 490, in train result = Trainable.train(self) File "[...]/ray/tune/trainable.py", line 260, in train result = self._train() File "[...]/ray/rllib/agents/trainer_template.py", line 138, in _train return self._train_exec_impl() File "[...]/ray/rllib/agents/trainer_template.py", line 173, in _train_exec_impl res = next(self.train_exec_impl) File "[...]/ray/util/iter.py", line 689, in __next__ return next(self.built_iterator) File "[...]/ray/util/iter.py", line 702, in apply_foreach for item in it: File "[...]/ray/util/iter.py", line 772, in apply_filter for item in it: File "[...]/ray/util/iter.py", line 772, in apply_filter for item in it: File "[...]/ray/util/iter.py", line 702, in apply_foreach for item in it: File "[...]/ray/util/iter.py", line 772, in apply_filter for item in it: File "[...]/ray/util/iter.py", line 977, in build_union item = next(it) File "[...]/ray/util/iter.py", line 689, in __next__ return next(self.built_iterator) File "[...]/ray/util/iter.py", line 702, in apply_foreach for item in it: File "[...]/ray/util/iter.py", line 702, in apply_foreach for item in it: File "[...]/ray/util/iter.py", line 702, in apply_foreach for item in it: File "[...]/ray/rllib/execution/rollout_ops.py", line 70, in sampler yield workers.local_worker().sample() File "[...]/ray/rllib/evaluation/rollout_worker.py", line 515, in sample batches = [self.input_reader.next()] File "[...]/ray/rllib/evaluation/sampler.py", line 56, in next batches = [self.get_data()] File "[...]/ray/rllib/evaluation/sampler.py", line 101, in get_data item = next(self.rollout_provider) File "[...]/ray/rllib/evaluation/sampler.py", line 367, in _env_runner active_episodes) File "[...]/ray/rllib/evaluation/sampler.py", line 637, in _do_policy_eval timestep=policy.global_timestep) File "[...]/ray/rllib/agents/qmix/qmix_policy.py", line 260, in compute_actions obs_batch, action_mask, _ = self._unpack_observation(obs_batch) File "[...]/ray/rllib/agents/qmix/qmix_policy.py", line 486, in _unpack_observation axis=1).reshape([len(obs_batch), self.n_agents, self.obs_size]) File "<__array_function__ internals>", line 6, in concatenate ValueError: zero-dimensional arrays cannot be concatenated ```
2020-05-21T10:26:15
ray-project/ray
8,572
ray-project__ray-8572
[ "8368" ]
351839bf698dec8dd8592ef70cd063ceab778d0e
diff --git a/rllib/agents/ppo/ppo_torch_policy.py b/rllib/agents/ppo/ppo_torch_policy.py --- a/rllib/agents/ppo/ppo_torch_policy.py +++ b/rllib/agents/ppo/ppo_torch_policy.py @@ -223,4 +223,7 @@ def setup_mixins(policy, obs_space, action_space, config): extra_grad_process_fn=apply_grad_clipping, before_init=setup_config, after_init=setup_mixins, - mixins=[KLCoeffMixin, ValueNetworkMixin]) + mixins=[ + LearningRateSchedule, EntropyCoeffSchedule, KLCoeffMixin, + ValueNetworkMixin + ])
"lr_schedule" option ignored using torch framework and PPO algorithm <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> *Ray version and other system information (Python version, TensorFlow version, OS):* - Ray: [0.9.0.dev0 (2c599dbf05e41e338920ee2fbe692658bcbec4dd)](https://s3-us-west-2.amazonaws.com/ray-wheels/releases/0.8.5/02c1ab0ec6d615ad54ebf33bd93c51c04000534e/ray-0.8.5-cp36-cp36m-manylinux1_x86_64.whl) - CUDA: 10.1 - Pytorch: 1.4.0 with GPU support - Ubuntu 18.04 - Python 3.6 ### What is the problem? Setting the hyperparameter "lr_schedule" as no effect when using PyTorch as backend framework and PPO learning algorithm. ### Reproduction (REQUIRED) ``` import ray from ray.rllib.agents.ppo import PPOTrainer, DEFAULT_CONFIG config = DEFAULT_CONFIG.copy() for key, val in { "env": "CartPole-v0", "num_workers": 0, "use_pytorch": False, "lr": 1.0e-5, "lr_schedule": [ [0, 1.0e-6], [1, 1.0e-7], ] }.items(): config[key] = val ray.init() for use_pytorch in [False, True]: config["use_pytorch"] = use_pytorch agent = PPOTrainer(config, "CartPole-v0") for _ in range(2): result = agent.train() print(f"use_pytorch: {use_pytorch} - Current learning rate: "\ f"{result['info']['learner']['default_policy']['cur_lr']}") ``` - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
Thanks for filing this issue @duburcqa . Will take a look ... cc @krfricke I can confirm this - would appreciate an update on it
2020-05-23T10:07:20
ray-project/ray
8,617
ray-project__ray-8617
[ "8614", "8614" ]
b0bb0584fb6e2a293eea7620f1a83d0350bdce9f
diff --git a/rllib/agents/a3c/a3c.py b/rllib/agents/a3c/a3c.py --- a/rllib/agents/a3c/a3c.py +++ b/rllib/agents/a3c/a3c.py @@ -54,11 +54,6 @@ def get_policy_class(config): def validate_config(config): if config["entropy_coeff"] < 0: raise DeprecationWarning("entropy_coeff must be >= 0") - if config["sample_async"] and config["use_pytorch"]: - config["sample_async"] = False - logger.warning( - "The sample_async option is not supported with use_pytorch: " - "Multithreading can be lead to crashes if used with pytorch.") def execution_plan(workers, config):
[rllib] PyTorch and SampleAsync validation <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? PyTorch is supposed to be thread-safe, as long as you don't write a tensor using multiple threads. Please see https://discuss.pytorch.org/t/is-pytorch-supposed-to-be-thread-safe/36540/2 It might be worth removing the validation of sample_async and use_pytorch for A3C (and maybe others?). Ray Version 0.9.0dev (but this applies to any ray version actually) ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): If we cannot run your script, we cannot fix your issue. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html). [rllib] PyTorch and SampleAsync validation <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? PyTorch is supposed to be thread-safe, as long as you don't write a tensor using multiple threads. Please see https://discuss.pytorch.org/t/is-pytorch-supposed-to-be-thread-safe/36540/2 It might be worth removing the validation of sample_async and use_pytorch for A3C (and maybe others?). Ray Version 0.9.0dev (but this applies to any ray version actually) ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): If we cannot run your script, we cannot fix your issue. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-05-26T15:16:42
ray-project/ray
8,628
ray-project__ray-8628
[ "8254" ]
cd5a207d69cdaf05b47d956c18e89d928585eec7
diff --git a/python/ray/node.py b/python/ray/node.py --- a/python/ray/node.py +++ b/python/ray/node.py @@ -4,6 +4,7 @@ import errno import os import logging +import random import signal import socket import subprocess @@ -24,6 +25,7 @@ logger = logging.getLogger(__name__) SESSION_LATEST = "session_latest" +NUMBER_OF_PORT_RETRIES = 40 class Node: @@ -167,7 +169,8 @@ def __init__(self, # NOTE: There is a possible but unlikely race condition where # the port is bound by another process between now and when the # raylet starts. - self._ray_params.node_manager_port = self._get_unused_port() + self._ray_params.node_manager_port, self._socket = \ + self._get_unused_port(close_on_exit=False) if not connect_only and spawn_reaper and not self.kernel_fate_share: self.start_reaper_process() @@ -300,6 +303,14 @@ def node_manager_port(self): """Get the node manager's port.""" return self._ray_params.node_manager_port + @property + def socket(self): + """Get the socket reserving the node manager's port""" + try: + return self._socket + except AttributeError: + return None + @property def address_info(self): """Get a dictionary of addresses.""" @@ -395,12 +406,30 @@ def new_log_files(self, name): log_stderr_file = open(log_stderr, "a", buffering=1) return log_stdout_file, log_stderr_file - def _get_unused_port(self): + def _get_unused_port(self, close_on_exit=True): s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) s.bind(("", 0)) port = s.getsockname()[1] - s.close() - return port + + # Try to generate a port that is far above the 'next available' one. + # This solves issue #8254 where GRPC fails because the port assigned + # from this method has been used by a different process. + for _ in range(NUMBER_OF_PORT_RETRIES): + new_port = random.randint(port, 65535) + new_s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) + try: + new_s.bind(("", new_port)) + except OSError: + new_s.close() + continue + s.close() + if close_on_exit: + new_s.close() + return new_port, new_s + logger.error("Unable to succeed in selecting a random port.") + if close_on_exit: + s.close() + return port, s def _prepare_socket_file(self, socket_path, default_prefix): """Prepare the socket file for raylet and plasma. @@ -417,7 +446,7 @@ def _prepare_socket_file(self, socket_path, default_prefix): if sys.platform == "win32": if socket_path is None: result = "tcp://{}:{}".format(self._localhost, - self._get_unused_port()) + self._get_unused_port()[0]) else: if socket_path is None: result = self._make_inc_temp( @@ -598,7 +627,8 @@ def start_raylet(self, use_valgrind=False, use_profiler=False): include_java=self._ray_params.include_java, java_worker_options=self._ray_params.java_worker_options, load_code_from_local=self._ray_params.load_code_from_local, - fate_share=self.kernel_fate_share) + fate_share=self.kernel_fate_share, + socket_to_use=self.socket) assert ray_constants.PROCESS_TYPE_RAYLET not in self.all_processes self.all_processes[ray_constants.PROCESS_TYPE_RAYLET] = [process_info] diff --git a/python/ray/services.py b/python/ray/services.py --- a/python/ray/services.py +++ b/python/ray/services.py @@ -1230,7 +1230,8 @@ def start_raylet(redis_address, include_java=False, java_worker_options=None, load_code_from_local=False, - fate_share=None): + fate_share=None, + socket_to_use=None): """Start a raylet, which is a combined local scheduler and object manager. Args: @@ -1361,6 +1362,8 @@ def start_raylet(redis_address, "--temp_dir={}".format(temp_dir), "--session_dir={}".format(session_dir), ] + if socket_to_use: + socket_to_use.close() process_info = start_ray_process( command, ray_constants.PROCESS_TYPE_RAYLET,
[Core] Node Failure Long Running Test Failed at `grpc::ServerInterface::RegisteredAsyncRequest::IssueRequest()` <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? Node Failure Long Running test fails with the error log below. ``` (pid=raylet) E0429 02:32:06.263886 22036 process.cc:274] Failed to wait for process 22047 with error system:10: No child processes E0429 02:32:12.346844 23272 task_manager.cc:288] 3 retries left for task b48f33dc1265b526ffffffff0100, attempting to resubmit. E0429 02:32:12.346899 23272 core_worker.cc:373] Will resubmit task after a 5000ms delay: Type=NORMAL_TASK, Language=PYTHON, function_descriptor={type=PythonFunctionDescriptor, module_name=__main__, class_name=, function_name=f, function_hash=7d2c6c88e5e801d48a350076f2117e717fe12224}, task_id=b48f33dc1265b526ffffffff0100, job_id=0100, num_args=2, num_returns=1 (pid=raylet) E0429 02:32:12.347446 22089 process.cc:274] Failed to wait for process 22100 with error system:10: No child processes 2020-04-29 02:32:12,653 INFO resource_spec.py:212 -- Starting Ray with 27.88 GiB memory available for workers and up to 0.15 GiB for objects. You can adjust these settings with ray.init(memory=<bytes>, object_store_memory=<bytes>). (pid=raylet) E0429 02:32:12.732946757 22142 server_chttp2.cc:40] {"created":"@1588127532.732848116","description":"No address added out of total 1 resolved","file":"external/com_github_grpc_grpc/src/core/ext/transport/chttp2/server/chttp2_server.cc","file_line":394,"referenced_errors":[{"created":"@1588127532.732846227","description":"Failed to add any wildcard listeners","file":"external/com_github_grpc_grpc/src/core/lib/iomgr/tcp_server_posix.cc","file_line":341,"referenced_errors":[{"created":"@1588127532.732832876","description":"Unable to configure socket","fd":44,"file":"external/com_github_grpc_grpc/src/core/lib/iomgr/tcp_server_utils_posix_common.cc","file_line":208,"referenced_errors":[{"created":"@1588127532.732823689","description":"Address already in use","errno":98,"file":"external/com_github_grpc_grpc/src/core/lib/iomgr/tcp_server_utils_posix_common.cc","file_line":181,"os_error":"Address already in use","syscall":"bind"}]},{"created":"@1588127532.732845812","description":"Unable to configure socket","fd":44,"file":"external/com_github_grpc_grpc/src/core/lib/iomgr/tcp_server_utils_posix_common.cc","file_line":208,"referenced_errors":[{"created":"@1588127532.732843382","description":"Address already in use","errno":98,"file":"external/com_github_grpc_grpc/src/core/lib/iomgr/tcp_server_utils_posix_common.cc","file_line":181,"os_error":"Address already in use","syscall":"bind"}]}]}]} (pid=raylet) *** Aborted at 1588127532 (unix time) try "date -d @1588127532" if you are using GNU date *** (pid=raylet) PC: @ 0x0 (unknown) (pid=raylet) *** SIGSEGV (@0x58) received by PID 22142 (TID 0x7fc3a66d37c0) from PID 88; stack trace: *** (pid=raylet) @ 0x7fc3a5c32390 (unknown) (pid=raylet) @ 0x5596e3957692 grpc::ServerInterface::RegisteredAsyncRequest::IssueRequest() (pid=raylet) @ 0x5596e35b2149 ray::rpc::NodeManagerService::WithAsyncMethod_RequestWorkerLease<>::RequestRequestWorkerLease() (pid=raylet) @ 0x5596e35c7b1b ray::rpc::ServerCallFactoryImpl<>::CreateCall() (pid=raylet) @ 0x5596e380bfe1 ray::rpc::GrpcServer::Run() (pid=raylet) @ 0x5596e3629acc ray::raylet::NodeManager::NodeManager() (pid=raylet) @ 0x5596e35cbc07 ray::raylet::Raylet::Raylet() (pid=raylet) @ 0x5596e359848d main (pid=raylet) @ 0x7fc3a5459830 __libc_start_main (pid=raylet) @ 0x5596e35a9391 (unknown) (pid=22153) E0429 02:32:13.864451 22153 raylet_client.cc:69] Retrying to connect to socket for pathname /tmp/ray/session_2020-04-28_20-19-44_770473_22870/sockets/raylet.8 (num_attempts = 1, num_retries = 5) (pid=22153) E0429 02:32:14.364712 22153 raylet_client.cc:69] Retrying to connect to socket for pathname /tmp/ray/session_2020-04-28_20-19-44_770473_22870/sockets/raylet.8 (num_attempts = 2, num_retries = 5) (pid=22153) E0429 02:32:14.864863 22153 raylet_client.cc:69] Retrying to connect to socket for pathname /tmp/ray/session_2020-04-28_20-19-44_770473_22870/sockets/raylet.8 (num_attempts = 3, num_retries = 5) (pid=22153) E0429 02:32:15.365000 22153 raylet_client.cc:69] Retrying to connect to socket for pathname /tmp/ray/session_2020-04-28_20-19-44_770473_22870/sockets/raylet.8 (num_attempts = 4, num_retries = 5) (pid=22153) F0429 02:32:15.865115 22153 raylet_client.cc:78] Could not connect to socket /tmp/ray/session_2020-04-28_20-19-44_770473_22870/sockets/raylet.8 (pid=22153) *** Check failure stack trace: *** (pid=22153) @ 0x7f5d3b2b40ed google::LogMessage::Fail() (pid=22153) @ 0x7f5d3b2b555c google::LogMessage::SendToLog() (pid=22153) @ 0x7f5d3b2b3dc9 google::LogMessage::Flush() (pid=22153) @ 0x7f5d3b2b3fe1 google::LogMessage::~LogMessage() (pid=22153) @ 0x7f5d3b03bb39 ray::RayLog::~RayLog() (pid=22153) @ 0x7f5d3ae55133 ray::raylet::RayletConnection::RayletConnection() (pid=22153) @ 0x7f5d3ae55abf ray::raylet::RayletClient::RayletClient() (pid=22153) @ 0x7f5d3adf513b ray::CoreWorker::CoreWorker() (pid=22153) @ 0x7f5d3adf8984 ray::CoreWorkerProcess::CreateWorker() (pid=22153) @ 0x7f5d3adf8efb ray::CoreWorkerProcess::CoreWorkerProcess() (pid=22153) @ 0x7f5d3adf93fb ray::CoreWorkerProcess::Initialize() (pid=22153) @ 0x7f5d3ad6c06c __pyx_pw_3ray_7_raylet_10CoreWorker_1__cinit__() (pid=22153) @ 0x7f5d3ad6d155 __pyx_tp_new_3ray_7_raylet_CoreWorker() (pid=22153) @ 0x55db24e47965 type_call (pid=22153) @ 0x55db24db7d7b _PyObject_FastCallDict (pid=22153) @ 0x55db24e477ce call_function (pid=22153) @ 0x55db24e69cba _PyEval_EvalFrameDefault (pid=22153) @ 0x55db24e40dae _PyEval_EvalCodeWithName (pid=22153) @ 0x55db24e41941 fast_function (pid=22153) @ 0x55db24e47755 call_function (pid=22153) @ 0x55db24e6aa7a _PyEval_EvalFrameDefault (pid=22153) @ 0x55db24e42459 PyEval_EvalCodeEx (pid=22153) @ 0x55db24e431ec PyEval_EvalCode (pid=22153) @ 0x55db24ebd9a4 run_mod (pid=22153) @ 0x55db24ebdda1 PyRun_FileExFlags (pid=22153) @ 0x55db24ebdfa4 PyRun_SimpleFileExFlags (pid=22153) @ 0x55db24ec1a9e Py_Main (pid=22153) @ 0x55db24d894be main (pid=22153) @ 0x7f5d3cd85830 __libc_start_main (pid=22153) @ 0x55db24e70773 (unknown) Traceback (most recent call last): File "workloads/node_failures.py", line 57, in <module> cluster.add_node() File "/home/ubuntu/anaconda3/envs/tensorflow_p36/lib/python3.6/site-packages/ray/cluster_utils.py", line 115, in add_node self._wait_for_node(node) File "/home/ubuntu/anaconda3/envs/tensorflow_p36/lib/python3.6/site-packages/ray/cluster_utils.py", line 165, in _wait_for_node raise TimeoutError("Timed out while waiting for nodes to join.") TimeoutError: Timed out while waiting for nodes to join. (pid=raylet) E0429 02:32:42.965368 13125 process.cc:274] Failed to wait for process 13136 with error system:10: No child processes (pid=raylet) E0429 02:32:43.045863 1167 process.cc:274] Failed to wait for process 1178 with error system:10: No child processes 2020-04-29 02:32:43,942 ERROR import_thread.py:93 -- ImportThread: Connection closed by server. 2020-04-29 02:32:43,942 ERROR worker.py:996 -- print_logs: Connection closed by server. 2020-04-29 02:32:43,942 ERROR worker.py:1096 -- listen_error_messages_raylet: Connection closed by server. E0429 02:32:45.999132 22870 raylet_client.cc:90] IOError: [RayletClient] Connection closed unexpectedly. [RayletClient] Failed to disconnect from raylet. ``` *Ray version and other system information (Python version, TensorFlow version, OS):* ### Reproduction (REQUIRED) This is hard to reproduce, but it happens consistently in multiple long running tests. For the logs above, it happens after 2909 iterations. - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
@stephanie-wang Are you working on this issue? I don't think this is a P0. It's just a side effect of rapidly killing and adding ray nodes on the same machine. Raylet gRPC server cannot bind to a port that's just released by killing ray node. This is uncommon in production scenarios. Moving it to P1. However, we do need to fix the long running test by adding some sleep buffer between killing and adding a new node, maybe inside test_reconstruction as well. It looks like the ultimate source of the problem is that the port for GPRC is selected before GRPC actually uses that port: https://github.com/ray-project/ray/blob/137519e19d77debb5427a24279320f21e805867f/python/ray/node.py#L167 @simon-mo What are your thoughts on this? One possible fix is to randomize the port instead of using kernel to get unused port. This can avoid possible race condition. The reason is that current `_get_unused_port` procedure returns unused port in sequence and is likely to duplicate port numbers, since we close the socket immediately and release the ownership of that port. ```python def _get_unused_port(self): s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) s.bind(("", 0)) port = s.getsockname()[1] # fix: add some random integer to the port number, returns if available, else retry s.close() return port ```
2020-05-27T01:18:16
ray-project/ray
8,731
ray-project__ray-8731
[ "8470" ]
64a98e44475b879630dda789b9c9fcd5a1f6931e
diff --git a/python/ray/tune/analysis/experiment_analysis.py b/python/ray/tune/analysis/experiment_analysis.py --- a/python/ray/tune/analysis/experiment_analysis.py +++ b/python/ray/tune/analysis/experiment_analysis.py @@ -220,34 +220,35 @@ def get_best_trial(self, metric, mode="max", scope="all"): Args: metric (str): Key for trial info to order on. mode (str): One of [min, max]. - scope (str): One of [all, last, avg]. If `scope=last`, only look at - each trial's final step for `metric`, and compare across - trials based on `mode=[min,max]`. If `scope=avg`, consider the - simple average over all steps for `metric` and compare across - trials based on `mode=[min,max]`. If `scope=all`, find each - trial's min/max score for `metric` based on `mode`, and - compare trials based on `mode=[min,max]`. + scope (str): One of [all, last, avg, last-5-avg, last-10-avg]. + If `scope=last`, only look at each trial's final step for + `metric`, and compare across trials based on `mode=[min,max]`. + If `scope=avg`, consider the simple average over all steps + for `metric` and compare across trials based on + `mode=[min,max]`. If `scope=last-5-avg` or `scope=last-10-avg`, + consider the simple average over the last 5 or 10 steps for + `metric` and compare across trials based on `mode=[min,max]`. + If `scope=all`, find each trial's min/max score for `metric` + based on `mode`, and compare trials based on `mode=[min,max]`. """ if mode not in ["max", "min"]: raise ValueError( "ExperimentAnalysis: attempting to get best trial for " "metric {} for mode {} not in [\"max\", \"min\"]".format( metric, mode)) - if scope not in ["all", "last", "avg"]: + if scope not in ["all", "last", "avg", "last-5-avg", "last-10-avg"]: raise ValueError( "ExperimentAnalysis: attempting to get best trial for " - "metric {} for scope {} not in [\"all\", \"last\", \"avg\"]". - format(metric, scope)) + "metric {} for scope {} not in [\"all\", \"last\", \"avg\", " + "\"last-5-avg\", \"last-10-avg\"]".format(metric, scope)) best_trial = None best_metric_score = None for trial in self.trials: if metric not in trial.metric_analysis: continue - if scope == "last": - metric_score = trial.metric_analysis[metric]["last"] - elif scope == "avg": - metric_score = trial.metric_analysis[metric]["avg"] + if scope in ["last", "avg", "last-5-avg", "last-10-avg"]: + metric_score = trial.metric_analysis[metric][scope] else: metric_score = trial.metric_analysis[metric][mode] @@ -273,13 +274,16 @@ def get_best_config(self, metric, mode="max", scope="all"): Args: metric (str): Key for trial info to order on. mode (str): One of [min, max]. - scope (str): One of [all, last, avg]. If `scope=last`, only look at - each trial's final step for `metric`, and compare across - trials based on `mode=[min,max]`. If `scope=avg`, consider the - simple average over all steps for `metric` and compare across - trials based on `mode=[min,max]`. If `scope=all`, find each - trial's min/max score for `metric` based on `mode`, and - compare trials based on `mode=[min,max]`. + scope (str): One of [all, last, avg, last-5-avg, last-10-avg]. + If `scope=last`, only look at each trial's final step for + `metric`, and compare across trials based on `mode=[min,max]`. + If `scope=avg`, consider the simple average over all steps + for `metric` and compare across trials based on + `mode=[min,max]`. If `scope=last-5-avg` or `scope=last-10-avg`, + consider the simple average over the last 5 or 10 steps for + `metric` and compare across trials based on `mode=[min,max]`. + If `scope=all`, find each trial's min/max score for `metric` + based on `mode`, and compare trials based on `mode=[min,max]`. """ best_trial = self.get_best_trial(metric, mode, scope) return best_trial.config if best_trial else None @@ -292,13 +296,16 @@ def get_best_logdir(self, metric, mode="max", scope="all"): Args: metric (str): Key for trial info to order on. mode (str): One of [min, max]. - scope (str): One of [all, last, avg]. If `scope=last`, only look at - each trial's final step for `metric`, and compare across - trials based on `mode=[min,max]`. If `scope=avg`, consider the - simple average over all steps for `metric` and compare across - trials based on `mode=[min,max]`. If `scope=all`, find each - trial's min/max score for `metric` based on `mode`, and - compare trials based on `mode=[min,max]`. + scope (str): One of [all, last, avg, last-5-avg, last-10-avg]. + If `scope=last`, only look at each trial's final step for + `metric`, and compare across trials based on `mode=[min,max]`. + If `scope=avg`, consider the simple average over all steps + for `metric` and compare across trials based on + `mode=[min,max]`. If `scope=last-5-avg` or `scope=last-10-avg`, + consider the simple average over the last 5 or 10 steps for + `metric` and compare across trials based on `mode=[min,max]`. + If `scope=all`, find each trial's min/max score for `metric` + based on `mode`, and compare trials based on `mode=[min,max]`. """ best_trial = self.get_best_trial(metric, mode, scope) return best_trial.logdir if best_trial else None diff --git a/python/ray/tune/trial.py b/python/ray/tune/trial.py --- a/python/ray/tune/trial.py +++ b/python/ray/tune/trial.py @@ -1,4 +1,5 @@ import ray.cloudpickle as cloudpickle +from collections import deque import copy from datetime import datetime import logging @@ -214,9 +215,14 @@ def __init__(self, self.last_result = {} self.last_update_time = -float("inf") - # stores in memory max/min/avg/last result for each metric by trial + # stores in memory max/min/avg/last-n-avg/last result for each + # metric by trial self.metric_analysis = {} + # keep a moving average over these last n steps + self.n_steps = [5, 10] + self.metric_n_steps = {} + self.export_formats = export_formats self.status = Trial.PENDING self.start_time = None @@ -470,6 +476,7 @@ def update_last_result(self, result, terminate=False): self.last_result = result self.last_update_time = time.time() self.result_logger.on_result(self.last_result) + for metric, value in flatten_dict(result).items(): if isinstance(value, Number): if metric not in self.metric_analysis: @@ -479,6 +486,13 @@ def update_last_result(self, result, terminate=False): "avg": value, "last": value } + self.metric_n_steps[metric] = {} + for n in self.n_steps: + key = "last-{:d}-avg".format(n) + self.metric_analysis[metric][key] = value + # Store n as string for correct restore. + self.metric_n_steps[metric][str(n)] = deque( + [value], maxlen=n) else: step = result["training_iteration"] or 1 self.metric_analysis[metric]["max"] = max( @@ -490,6 +504,13 @@ def update_last_result(self, result, terminate=False): (step - 1) * self.metric_analysis[metric]["avg"]) self.metric_analysis[metric]["last"] = value + for n in self.n_steps: + key = "last-{:d}-avg".format(n) + self.metric_n_steps[metric][str(n)].append(value) + self.metric_analysis[metric][key] = sum( + self.metric_n_steps[metric][str(n)]) / len( + self.metric_n_steps[metric][str(n)]) + def get_trainable_cls(self): return get_trainable_cls(self.trainable_name)
diff --git a/python/ray/tune/tests/test_experiment_analysis_mem.py b/python/ray/tune/tests/test_experiment_analysis_mem.py --- a/python/ray/tune/tests/test_experiment_analysis_mem.py +++ b/python/ray/tune/tests/test_experiment_analysis_mem.py @@ -14,11 +14,11 @@ class ExperimentAnalysisInMemorySuite(unittest.TestCase): def setUp(self): class MockTrainable(Trainable): scores_dict = { - 0: [5, 4, 0], - 1: [4, 3, 1], - 2: [2, 1, 8], - 3: [9, 7, 6], - 4: [7, 5, 3] + 0: [5, 4, 4, 4, 4, 4, 4, 4, 0], + 1: [4, 3, 3, 3, 3, 3, 3, 3, 1], + 2: [2, 1, 1, 1, 1, 1, 1, 1, 8], + 3: [9, 7, 7, 7, 7, 7, 7, 7, 6], + 4: [7, 5, 5, 5, 5, 5, 5, 5, 3] } def _setup(self, config): @@ -53,7 +53,7 @@ def testCompareTrials(self): self.MockTrainable, name="analysis_exp", local_dir=self.test_dir, - stop={"training_iteration": 3}, + stop={"training_iteration": len(scores[0])}, num_samples=1, config={"id": grid_search(list(range(5)))}) @@ -67,12 +67,33 @@ def testCompareTrials(self): "avg").metric_analysis["score"]["avg"] min_avg = ea.get_best_trial("score", "min", "avg").metric_analysis["score"]["avg"] + max_avg_5 = ea.get_best_trial( + "score", "max", + "last-5-avg").metric_analysis["score"]["last-5-avg"] + min_avg_5 = ea.get_best_trial( + "score", "min", + "last-5-avg").metric_analysis["score"]["last-5-avg"] + max_avg_10 = ea.get_best_trial( + "score", "max", + "last-10-avg").metric_analysis["score"]["last-10-avg"] + min_avg_10 = ea.get_best_trial( + "score", "min", + "last-10-avg").metric_analysis["score"]["last-10-avg"] self.assertEqual(max_all, max(scores_all)) self.assertEqual(min_all, min(scores_all)) self.assertEqual(max_last, max(scores_last)) + self.assertNotEqual(max_last, max(scores_all)) + self.assertAlmostEqual(max_avg, max(np.mean(scores, axis=1))) self.assertAlmostEqual(min_avg, min(np.mean(scores, axis=1))) - self.assertNotEqual(max_last, max(scores_all)) + + self.assertAlmostEqual(max_avg_5, max(np.mean(scores[:, -5:], axis=1))) + self.assertAlmostEqual(min_avg_5, min(np.mean(scores[:, -5:], axis=1))) + + self.assertAlmostEqual(max_avg_10, max( + np.mean(scores[:, -10:], axis=1))) + self.assertAlmostEqual(min_avg_10, min( + np.mean(scores[:, -10:], axis=1))) class AnalysisSuite(unittest.TestCase):
[tune] Ea scope avg n step <!-- Thank you for your contribution! Please review https://github.com/ray-project/ray/blob/master/CONTRIBUTING.rst before opening a pull request. --> Continuing #8147, this adds `last-n-avg`, specifically `last-5-avg` and `last-10-avg` scopes to the collected experiment metrics. These _n_ are currently hardcoded since a variable number of steps would need an extension of the API for trials and the experiment runner. <!-- Please give a short summary of the change and the problem this solves. --> Closes #8147. <!-- For example: "Closes #1234" --> ## Checks - [X] I've run `scripts/format.sh` to lint the changes in this PR. - [X] I've included any doc changes needed for https://docs.ray.io/en/latest/. - [X] I've made sure the tests are passing. Note that there might be a few flaky tests, see the recent failure rates at https://ray-travis-tracker.herokuapp.com/. - Testing Strategy - [X] Unit tests - [ ] Release tests - [ ] This PR is not tested (please justify below)
Can one of the admins verify this patch? Test FAILed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Ray-PRB/25977/ Test FAILed. Looks like tune tests are failing? There seems to be an error when restoring a run where the `self.metric_n_steps` don't get initialized: https://github.com/ray-project/ray/runs/681419121#step:3:2390 I'll fix this tomorrow. The error was in restoring the n-step metrics. The dict keys were restored as strings, while they were originally created as integers. We should talk about whether this is expected behavior. For the meantime, we're now converting the integers to strings for the dict keys. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Ray-PRB/26084/ Test PASSed. ah can we fix up the PR? looks like the merge history is broken I rebased the PR against the current master and resolved any merge conflicts. I hope it works now. Alternatively, I could re-create the changes on a fresh branch to avoid the messy commit history. Test PASSed. Refer to this link for build results (access rights to CI server needed): https://amplab.cs.berkeley.edu/jenkins//job/Ray-PRB/26574/ Test PASSed.
2020-06-02T11:09:37
ray-project/ray
8,770
ray-project__ray-8770
[ "8769" ]
9410e5884d81d45cb64dc31410c91686c4cbeb4e
diff --git a/rllib/agents/a3c/a3c.py b/rllib/agents/a3c/a3c.py --- a/rllib/agents/a3c/a3c.py +++ b/rllib/agents/a3c/a3c.py @@ -54,6 +54,10 @@ def get_policy_class(config): def validate_config(config): if config["entropy_coeff"] < 0: raise DeprecationWarning("entropy_coeff must be >= 0") + if config["sample_async"] and config["framework"] == "torch": + config["sample_async"] = False + logger.warning("`sample_async=True` is not supported for PyTorch! " + "Multithreading can lead to crashes.") def execution_plan(workers, config): diff --git a/rllib/agents/ars/ars.py b/rllib/agents/ars/ars.py --- a/rllib/agents/ars/ars.py +++ b/rllib/agents/ars/ars.py @@ -11,8 +11,8 @@ from ray.rllib.agents import Trainer, with_common_config from ray.rllib.agents.ars.ars_tf_policy import ARSTFPolicy -from ray.rllib.agents.es import optimizers -from ray.rllib.agents.es import utils +from ray.rllib.agents.es import optimizers, utils +from ray.rllib.agents.es.es import validate_config from ray.rllib.agents.es.es_tf_policy import rollout from ray.rllib.env.env_context import EnvContext from ray.rllib.policy.sample_batch import DEFAULT_POLICY_ID @@ -179,6 +179,7 @@ class ARSTrainer(Trainer): @override(Trainer) def _init(self, config, env_creator): + validate_config(config) env_context = EnvContext(config["env_config"] or {}, worker_index=0) env = env_creator(env_context) diff --git a/rllib/agents/es/es.py b/rllib/agents/es/es.py --- a/rllib/agents/es/es.py +++ b/rllib/agents/es/es.py @@ -169,6 +169,11 @@ def get_policy_class(config): return policy_cls +def validate_config(config): + if config["num_workers"] <= 0: + raise ValueError("`num_workers` must be > 0 for ES!") + + class ESTrainer(Trainer): """Large-scale implementation of Evolution Strategies in Ray.""" @@ -177,6 +182,7 @@ class ESTrainer(Trainer): @override(Trainer) def _init(self, config, env_creator): + validate_config(config) env_context = EnvContext(config["env_config"] or {}, worker_index=0) env = env_creator(env_context) policy_cls = get_policy_class(config)
diff --git a/rllib/tests/test_rollout.py b/rllib/tests/test_rollout.py --- a/rllib/tests/test_rollout.py +++ b/rllib/tests/test_rollout.py @@ -6,9 +6,13 @@ from ray.rllib.utils.test_utils import framework_iterator -def rollout_test(algo, env="CartPole-v0"): +def rollout_test(algo, env="CartPole-v0", test_episode_rollout=False): + extra_config = "" + if algo == "ES": + extra_config = ",\"episodes_per_batch\": 1,\"train_batch_size\": 10, "\ + "\"noise_size\": 250000" - for fw in framework_iterator(frameworks=("torch", "tf")): + for fw in framework_iterator(frameworks=("tf", "torch")): fw_ = ", \"framework\": \"{}\"".format(fw) tmp_dir = os.popen("mktemp -d").read()[:-1] @@ -22,8 +26,8 @@ def rollout_test(algo, env="CartPole-v0"): os.path.exists(rllib_dir))) os.system("python {}/train.py --local-dir={} --run={} " "--checkpoint-freq=1 ".format(rllib_dir, tmp_dir, algo) + - "--config='{" + - "\"num_workers\": 0, \"num_gpus\": 0{}".format(fw_) + + "--config='{" + "\"num_workers\": 1, \"num_gpus\": 0{}{}". + format(fw_, extra_config) + ", \"model\": {\"fcnet_hiddens\": [10]}" "}' --stop='{\"training_iteration\": 1, " "\"timesteps_per_iter\": 5, " @@ -44,12 +48,13 @@ def rollout_test(algo, env="CartPole-v0"): print("rollout output (10 steps) exists!".format(checkpoint_path)) # Test rolling out 1 episode. - os.popen("python {}/rollout.py --run={} \"{}\" --episodes=1 " - "--out=\"{}/rollouts_1episode.pkl\" --no-render".format( - rllib_dir, algo, checkpoint_path, tmp_dir)).read() - if not os.path.exists(tmp_dir + "/rollouts_1episode.pkl"): - sys.exit(1) - print("rollout output (1 ep) exists!".format(checkpoint_path)) + if test_episode_rollout: + os.popen("python {}/rollout.py --run={} \"{}\" --episodes=1 " + "--out=\"{}/rollouts_1episode.pkl\" --no-render".format( + rllib_dir, algo, checkpoint_path, tmp_dir)).read() + if not os.path.exists(tmp_dir + "/rollouts_1episode.pkl"): + sys.exit(1) + print("rollout output (1 ep) exists!".format(checkpoint_path)) # Cleanup. os.popen("rm -rf \"{}\"".format(tmp_dir)).read() @@ -72,13 +77,10 @@ def test_es(self): rollout_test("ES") def test_impala(self): - rollout_test("IMPALA", env="Pong-ram-v4") - - def test_pg(self): - rollout_test("PG") + rollout_test("IMPALA", env="CartPole-v0") def test_ppo(self): - rollout_test("PPO", env="Pendulum-v0") + rollout_test("PPO", env="CartPole-v0", test_episode_rollout=True) def test_sac(self): rollout_test("SAC", env="Pendulum-v0") diff --git a/rllib/tests/test_supported_multi_agent.py b/rllib/tests/test_supported_multi_agent.py --- a/rllib/tests/test_supported_multi_agent.py +++ b/rllib/tests/test_supported_multi_agent.py @@ -14,23 +14,25 @@ def check_support_multiagent(alg, config): register_env("multi_agent_cartpole", lambda _: MultiAgentCartPole({"num_agents": 2})) config["log_level"] = "ERROR" - for _ in framework_iterator(config, frameworks=("tf", "torch")): + for _ in framework_iterator(config, frameworks=("torch", "tf")): if alg in ["DDPG", "APEX_DDPG", "SAC"]: a = get_agent_class(alg)( config=config, env="multi_agent_mountaincar") else: a = get_agent_class(alg)(config=config, env="multi_agent_cartpole") try: - a.train() + print(a.train()) finally: a.stop() -class ModelSupportedSpaces(unittest.TestCase): - def setUp(self): - ray.init(num_cpus=4, ignore_reinit_error=True) +class TestSupportedMultiAgent(unittest.TestCase): + @classmethod + def setUpClass(cls) -> None: + ray.init(num_cpus=4) - def tearDown(self): + @classmethod + def tearDownClass(cls) -> None: ray.shutdown() def test_a3c_multiagent(self): @@ -45,10 +47,11 @@ def test_apex_multiagent(self): check_support_multiagent( "APEX", { "num_workers": 2, - "timesteps_per_iteration": 1000, + "timesteps_per_iteration": 100, "num_gpus": 0, + "buffer_size": 1000, "min_iter_time_s": 1, - "learning_starts": 1000, + "learning_starts": 10, "target_network_update_freq": 100, }) @@ -56,10 +59,11 @@ def test_apex_ddpg_multiagent(self): check_support_multiagent( "APEX_DDPG", { "num_workers": 2, - "timesteps_per_iteration": 1000, + "timesteps_per_iteration": 100, + "buffer_size": 1000, "num_gpus": 0, "min_iter_time_s": 1, - "learning_starts": 1000, + "learning_starts": 10, "target_network_update_freq": 100, "use_state_preprocessor": True, }) @@ -68,12 +72,16 @@ def test_ddpg_multiagent(self): check_support_multiagent( "DDPG", { "timesteps_per_iteration": 1, + "buffer_size": 1000, "use_state_preprocessor": True, "learning_starts": 500, }) def test_dqn_multiagent(self): - check_support_multiagent("DQN", {"timesteps_per_iteration": 1}) + check_support_multiagent("DQN", { + "timesteps_per_iteration": 1, + "buffer_size": 1000, + }) def test_impala_multiagent(self): check_support_multiagent("IMPALA", {"num_gpus": 0}) @@ -94,6 +102,7 @@ def test_ppo_multiagent(self): def test_sac_multiagent(self): check_support_multiagent("SAC", { "num_workers": 0, + "buffer_size": 1000, "normalize_actions": False, }) diff --git a/rllib/tests/test_supported_spaces.py b/rllib/tests/test_supported_spaces.py --- a/rllib/tests/test_supported_spaces.py +++ b/rllib/tests/test_supported_spaces.py @@ -48,7 +48,7 @@ } -def check_support(alg, config, check_bounds=False, tfe=False): +def check_support(alg, config, train=True, check_bounds=False, tfe=False): config["log_level"] = "ERROR" def _do_check(alg, config, a_name, o_name): @@ -83,7 +83,8 @@ def _do_check(alg, config, a_name, o_name): assert isinstance(a.get_policy().model, TorchFCNetV2) else: assert isinstance(a.get_policy().model, FCNetV2) - a.train() + if train: + a.train() except UnsupportedSpaceException: stat = "unsupported" finally: @@ -99,19 +100,22 @@ def _do_check(alg, config, a_name, o_name): if tfe: frameworks += ("tfe", ) for _ in framework_iterator(config, frameworks=frameworks): - # Check all action spaces. + # Check all action spaces (using a discrete obs-space). for a_name, action_space in ACTION_SPACES_TO_TEST.items(): _do_check(alg, config, a_name, "discrete") - # Check all obs spaces. + # Check all obs spaces (using a supported action-space). for o_name, obs_space in OBSERVATION_SPACES_TO_TEST.items(): - _do_check(alg, config, "discrete", o_name) + a_name = "discrete" if alg not in ["DDPG", "SAC"] else "vector" + _do_check(alg, config, a_name, o_name) -class ModelSupportedSpaces(unittest.TestCase): - def setUp(self): - ray.init(num_cpus=4, ignore_reinit_error=True, local_mode=True) +class TestSupportedSpaces(unittest.TestCase): + @classmethod + def setUpClass(cls) -> None: + ray.init(num_cpus=4) - def tearDown(self): + @classmethod + def tearDownClass(cls) -> None: ray.shutdown() def test_a3c(self): @@ -119,7 +123,7 @@ def test_a3c(self): check_support("A3C", config, check_bounds=True) def test_appo(self): - check_support("APPO", {"num_gpus": 0, "vtrace": False}) + check_support("APPO", {"num_gpus": 0, "vtrace": False}, train=False) check_support("APPO", {"num_gpus": 0, "vtrace": True}) def test_ars(self): @@ -138,12 +142,13 @@ def test_ddpg(self): "ou_base_scale": 100.0 }, "timesteps_per_iteration": 1, + "buffer_size": 1000, "use_state_preprocessor": True, }, check_bounds=True) def test_dqn(self): - config = {"timesteps_per_iteration": 1} + config = {"timesteps_per_iteration": 1, "buffer_size": 1000} check_support("DQN", config, tfe=True) def test_es(self): @@ -170,10 +175,10 @@ def test_ppo(self): def test_pg(self): config = {"num_workers": 1, "optimizer": {}} - check_support("PG", config, check_bounds=True, tfe=True) + check_support("PG", config, train=False, check_bounds=True, tfe=True) def test_sac(self): - check_support("SAC", {}, check_bounds=True) + check_support("SAC", {"buffer_size": 1000}, check_bounds=True) if __name__ == "__main__":
[rllib] test_supported_spaces consistent OOM in travis <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? The second-to-last rllib travis build has been consistently failing since: https://github.com/ray-project/ray/pull/8520/files was merged. ``` E ray.memory_monitor.RayOutOfMemoryError: More than 95% of the memory on node travis-job-04150411-53df-4c8c-9bc4-2c378e131a11 is used (7.4 / 7.79 GB). The top 10 memory consumers are: E E PID MEM COMMAND E 30884 3.23GiB ray::RolloutWorker E 26855 0.91GiB bazel(ray) -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=/home/travis/.cache/bazel/_bazel_travis/ E 29471 0.46GiB /home/travis/build/ray-project/ray/python/ray/core/src/ray/thirdparty/redis/src/redis-server *:10263 E 29501 0.33GiB ray::ARSTrainer.train() ``` See the build history here: https://travis-ci.com/github/ray-project/ray/builds
This is particularly odd since ARS shouldn't be creating any RolloutWorker actors at all. Maybe it's something that got leaked from another test or ARS is a red herring. Are we sure it's not a core bug? It's not just this one test. All tests at the end of the alphabet fail due to this memory problem: test_rollout, test_rollout_worker, test_supported_multi_agent, test_supported_spaces. These are the last 4 in the test dir. They all run fine locally. All I know is it's consistently failing after that particular PR, and never had an issue before. Maybe try reverting it and see if that fixes it? yeah. This PR in question added many tests for pytorch that were previously only run for tf. Maybe this just made stuff go OOM (+ spillage between tests?). Will check.
2020-06-03T20:51:29
ray-project/ray
8,771
ray-project__ray-8771
[ "7900" ]
a24d117c689840ce1c67dadb78a7511a234a4bb2
diff --git a/python/ray/tune/syncer.py b/python/ray/tune/syncer.py --- a/python/ray/tune/syncer.py +++ b/python/ray/tune/syncer.py @@ -5,6 +5,7 @@ from shlex import quote +from ray import ray_constants from ray import services from ray.tune.cluster_info import get_ssh_key, get_ssh_user from ray.tune.sync_client import (CommandBasedClient, get_sync_client, @@ -12,7 +13,13 @@ logger = logging.getLogger(__name__) -SYNC_PERIOD = 300 +# Syncing period for syncing local checkpoints to cloud. +# In env variable is not set, sync happens every 300 seconds. +CLOUD_SYNC_PERIOD = ray_constants.env_integer( + key="TUNE_CLOUD_SYNC_S", default=300) + +# Syncing period for syncing worker logs to driver. +NODE_SYNC_PERIOD = 300 _log_sync_warned = False _syncers = {} @@ -70,12 +77,23 @@ def __init__(self, local_dir, remote_dir, sync_client=NOOP): self.last_sync_down_time = float("-inf") self.sync_client = sync_client - def sync_up_if_needed(self): - if time.time() - self.last_sync_up_time > SYNC_PERIOD: + def sync_up_if_needed(self, sync_period): + """Syncs up if time since last sync up is greather than sync_period. + + Arguments: + sync_period (int): Time period between subsequent syncs. + """ + + if time.time() - self.last_sync_up_time > sync_period: self.sync_up() - def sync_down_if_needed(self): - if time.time() - self.last_sync_down_time > SYNC_PERIOD: + def sync_down_if_needed(self, sync_period): + """Syncs down if time since last sync down is greather than sync_period. + + Arguments: + sync_period (int): Time period between subsequent syncs. + """ + if time.time() - self.last_sync_down_time > sync_period: self.sync_down() def sync_up(self): @@ -131,6 +149,19 @@ def _remote_path(self): return self._remote_dir +class CloudSyncer(Syncer): + """Syncer for syncing files to/from the cloud.""" + + def __init__(self, local_dir, remote_dir, sync_client): + super(CloudSyncer, self).__init__(local_dir, remote_dir, sync_client) + + def sync_up_if_needed(self): + return super(CloudSyncer, self).sync_up_if_needed(CLOUD_SYNC_PERIOD) + + def sync_down_if_needed(self): + return super(CloudSyncer, self).sync_down_if_needed(CLOUD_SYNC_PERIOD) + + class NodeSyncer(Syncer): """Syncer for syncing files to/from a remote dir to a local dir.""" @@ -158,12 +189,12 @@ def has_remote_target(self): def sync_up_if_needed(self): if not self.has_remote_target(): return True - return super(NodeSyncer, self).sync_up_if_needed() + return super(NodeSyncer, self).sync_up_if_needed(NODE_SYNC_PERIOD) def sync_down_if_needed(self): if not self.has_remote_target(): return True - return super(NodeSyncer, self).sync_down_if_needed() + return super(NodeSyncer, self).sync_down_if_needed(NODE_SYNC_PERIOD) def sync_up_to_new_location(self, worker_ip): if worker_ip != self.worker_ip: @@ -226,16 +257,16 @@ def get_cloud_syncer(local_dir, remote_dir=None, sync_function=None): return _syncers[key] if not remote_dir: - _syncers[key] = Syncer(local_dir, remote_dir, NOOP) + _syncers[key] = CloudSyncer(local_dir, remote_dir, NOOP) return _syncers[key] client = get_sync_client(sync_function) if client: - _syncers[key] = Syncer(local_dir, remote_dir, client) + _syncers[key] = CloudSyncer(local_dir, remote_dir, client) return _syncers[key] sync_client = get_cloud_sync_client(remote_dir) - _syncers[key] = Syncer(local_dir, remote_dir, sync_client) + _syncers[key] = CloudSyncer(local_dir, remote_dir, sync_client) return _syncers[key] diff --git a/python/ray/tune/tune.py b/python/ray/tune/tune.py --- a/python/ray/tune/tune.py +++ b/python/ray/tune/tune.py @@ -144,7 +144,9 @@ def run(run_or_experiment, from upload_dir. If string, then it must be a string template that includes `{source}` and `{target}` for the syncer to run. If not provided, the sync command defaults to standard S3 or gsutil sync - commands. + commands. By default local_dir is synced to remote_dir every 300 + seconds. To change this, set the TUNE_CLOUD_SYNC_S + environment variable in the driver machine. sync_to_driver (func|str|bool): Function for syncing trial logdir from remote node to local. If string, then it must be a string template that includes `{source}` and `{target}` for the syncer to run.
diff --git a/python/ray/tune/tests/test_cluster.py b/python/ray/tune/tests/test_cluster.py --- a/python/ray/tune/tests/test_cluster.py +++ b/python/ray/tune/tests/test_cluster.py @@ -18,7 +18,7 @@ from ray.tune.ray_trial_executor import RayTrialExecutor from ray.tune.resources import Resources from ray.tune.suggest import BasicVariantGenerator -from ray.tune.syncer import Syncer +from ray.tune.syncer import CloudSyncer from ray.tune.trainable import TrainableUtil from ray.tune.trial import Trial from ray.tune.trial_runner import TrialRunner @@ -521,7 +521,7 @@ def test_cluster_down_full(start_connected_cluster, tmpdir, trainable_id): mock_get_client = "ray.tune.trial_runner.get_cloud_syncer" with patch(mock_get_client) as mock_get_cloud_syncer: - mock_syncer = Syncer(local_dir, upload_dir, mock_storage_client()) + mock_syncer = CloudSyncer(local_dir, upload_dir, mock_storage_client()) mock_get_cloud_syncer.return_value = mock_syncer tune.run_experiments(all_experiments, raise_on_failed_trial=False) diff --git a/python/ray/tune/tests/test_sync.py b/python/ray/tune/tests/test_sync.py --- a/python/ray/tune/tests/test_sync.py +++ b/python/ray/tune/tests/test_sync.py @@ -3,6 +3,7 @@ import shutil import sys import tempfile +import time import unittest from unittest.mock import patch @@ -151,6 +152,38 @@ def sync_func(local, remote): shutil.rmtree(tmpdir) shutil.rmtree(tmpdir2) + @patch("ray.tune.sync_client.S3_PREFIX", "test") + def testCloudSyncPeriod(self): + """Tests that changing CLOUD_SYNC_PERIOD affects syncing frequency.""" + tmpdir = tempfile.mkdtemp() + + def trainable(config): + for i in range(10): + time.sleep(1) + tune.report(score=i) + + mock = unittest.mock.Mock() + + def counter(local, remote): + mock() + + tune.syncer.CLOUD_SYNC_PERIOD = 1 + [trial] = tune.run( + trainable, + name="foo", + max_failures=0, + local_dir=tmpdir, + upload_dir="test", + sync_to_cloud=counter, + stop={ + "training_iteration": 10 + }, + global_checkpoint_period=0.5, + ).trials + + self.assertEqual(mock.call_count, 12) + shutil.rmtree(tmpdir) + def testClusterSyncFunction(self): def sync_func_driver(source, target): assert ":" in source, "Source {} not a remote path.".format(source)
[tune] parametrize sync-to-cloud period As a background, our project interacts with tune (0.8.3) via `tune.run_experiments(<dict_from_flatfile>)`. We noticed that sometimes our trial checkpoints were not promptly syncing to our S3 account. Without hardcoding an S3 upload in our `tune.Trainable._save()` function, I uncovered 2 hardcoded parameters that sometimes prevented sync-to-cloud. For now here's our monkey-patch fix, although I propose tune developers to kindly consider native parametrization support for the future please :) ```python def set_tune_checkpoint_period(checkpoint_period_sec): """ Control delay between syncs to cloud (AWS S3), from hardcoded defaults. - ray.tune.tune.run(global_checkpoint_period): 10 seconds - ray.tune.syncer.SYNC_PERIOD: 300 seconds """ import inspect # noqa: E402 import ray.tune # noqa: E402 argspecs = inspect.getfullargspec(ray.tune.tune.run) arg_idx = argspecs.args.index("global_checkpoint_period") mandatory_arg_len = len(argspecs.args) - len(argspecs.defaults) new_defaults = list(argspecs.defaults) new_defaults[arg_idx - mandatory_arg_len] = checkpoint_period_sec ray.tune.tune.run.__defaults__ = tuple(new_defaults) import ray.tune.syncer # noqa: E402 ray.tune.syncer.SYNC_PERIOD = checkpoint_period_sec set_tune_checkpoint_period(1) ```
2020-06-03T21:31:13
ray-project/ray
8,781
ray-project__ray-8781
[ "8739" ]
41072fbcc8b3aea9fe726e960ca695f6365e9e40
diff --git a/rllib/models/tf/tf_action_dist.py b/rllib/models/tf/tf_action_dist.py --- a/rllib/models/tf/tf_action_dist.py +++ b/rllib/models/tf/tf_action_dist.py @@ -230,7 +230,7 @@ def deterministic_sample(self): @override(ActionDistribution) def logp(self, x): return -0.5 * tf.reduce_sum( - tf.square((x - self.mean) / self.std), axis=1) - \ + tf.square((tf.to_float(x) - self.mean) / self.std), axis=1) - \ 0.5 * np.log(2.0 * np.pi) * tf.to_float(tf.shape(x)[1]) - \ tf.reduce_sum(self.log_std, axis=1) diff --git a/rllib/policy/eager_tf_policy.py b/rllib/policy/eager_tf_policy.py --- a/rllib/policy/eager_tf_policy.py +++ b/rllib/policy/eager_tf_policy.py @@ -5,6 +5,7 @@ import functools import logging import numpy as np +from gym.spaces import Tuple, Dict from ray.util.debug import log_once from ray.rllib.models.catalog import ModelCatalog @@ -586,10 +587,17 @@ def _initialize_loss_with_dummy_batch(self): SampleBatch.NEXT_OBS: np.array( [self.observation_space.sample()]), SampleBatch.DONES: np.array([False], dtype=np.bool), - SampleBatch.ACTIONS: tf.nest.map_structure( - lambda c: np.array([c]), self.action_space.sample()), SampleBatch.REWARDS: np.array([0], dtype=np.float32), } + if isinstance(self.action_space, Tuple) or isinstance( + self.action_space, Dict): + dummy_batch[SampleBatch.ACTIONS] = [ + flatten_to_single_ndarray(self.action_space.sample()) + ] + else: + dummy_batch[SampleBatch.ACTIONS] = tf.nest.map_structure( + lambda c: np.array([c]), self.action_space.sample()) + if obs_include_prev_action_reward: dummy_batch.update({ SampleBatch.PREV_ACTIONS: dummy_batch[SampleBatch.ACTIONS],
diff --git a/rllib/tests/test_supported_spaces.py b/rllib/tests/test_supported_spaces.py --- a/rllib/tests/test_supported_spaces.py +++ b/rllib/tests/test_supported_spaces.py @@ -48,7 +48,7 @@ } -def check_support(alg, config, check_bounds=False): +def check_support(alg, config, check_bounds=False, tfe=False): config["log_level"] = "ERROR" def _do_check(alg, config, a_name, o_name): @@ -95,7 +95,10 @@ def _do_check(alg, config, a_name, o_name): pass print(stat) - for _ in framework_iterator(config, frameworks=("tf", "torch")): + frameworks = ("tf", "torch") + if tfe: + frameworks += ("tfe", ) + for _ in framework_iterator(config, frameworks=frameworks): # Check all action spaces. for a_name, action_space in ACTION_SPACES_TO_TEST.items(): _do_check(alg, config, a_name, "discrete") @@ -141,7 +144,7 @@ def test_ddpg(self): def test_dqn(self): config = {"timesteps_per_iteration": 1} - check_support("DQN", config) + check_support("DQN", config, tfe=True) def test_es(self): check_support( @@ -163,11 +166,11 @@ def test_ppo(self): "rollout_fragment_length": 10, "sgd_minibatch_size": 1, } - check_support("PPO", config, check_bounds=True) + check_support("PPO", config, check_bounds=True, tfe=True) def test_pg(self): config = {"num_workers": 1, "optimizer": {}} - check_support("PG", config, check_bounds=True) + check_support("PG", config, check_bounds=True, tfe=True) def test_sac(self): check_support("SAC", {}, check_bounds=True)
[rllib] Crash when using hybrid action space env and eager mode. <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? RLlib crashes on environment with **hybrid action space** (discrete and continuous) and **eager mode enabled** It works well if eager mode is turned off. _Python3.7 tensorflow==2.2.0 ray==0.9.0.dev0 (used the nightly version) Ubuntu 18.04_ ### Script to reproduce Modified from `ray/rllib/examples/custom_env.py` adding `Box` to the action space to make it hybrid ``` import gym import numpy as np import ray from gym.spaces import Discrete, Box from ray import tune from ray.rllib.utils.framework import try_import_tf tf = try_import_tf() class SimpleCorridor(gym.Env): """Example of a custom env in which you have to walk down a corridor. You can configure the length of the corridor via the env config.""" def __init__(self, config): self.end_pos = config["corridor_length"] self.cur_pos = 0 self.action_space = gym.spaces.Tuple([Discrete(2), Box(0.0, 1.0, shape=(1,))]) # Hybrid action space self.observation_space = Box( 0.0, self.end_pos, shape=(1,), dtype=np.float32) def reset(self): self.cur_pos = 0 return [self.cur_pos] def step(self, action): action = action[0] assert action in [0, 1], action if action == 0 and self.cur_pos > 0: self.cur_pos -= 1 elif action == 1: self.cur_pos += 1 done = self.cur_pos >= self.end_pos return [self.cur_pos], 1.0 if done else -0.1, done, {} if __name__ == "__main__": ray.init() config = { "env": SimpleCorridor, # or "corridor" if registered above "env_config": { "corridor_length": 5, }, "model": { }, "eager": True, # If True crashes "vf_share_layers": True, "lr": 1e-2, "num_workers": 1, "framework": "tf", } results = tune.run("PPO", config=config) ray.shutdown() ``` Outputs following error ``` File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/agents/trainer_template.py", line 90, in __init__ Trainer.__init__(self, config, env, logger_creator) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/agents/trainer.py", line 444, in __init__ super().__init__(config, logger_creator) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/tune/trainable.py", line 174, in __init__ self._setup(copy.deepcopy(self.config)) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/agents/trainer.py", line 613, in _setup self._init(self.config, self.env_creator) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/agents/trainer_template.py", line 115, in _init self.config["num_workers"]) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/agents/trainer.py", line 684, in _make_workers logdir=self.logdir) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/evaluation/worker_set.py", line 59, in __init__ RolloutWorker, env_creator, policy, 0, self._local_config) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/evaluation/worker_set.py", line 282, in _make_worker extra_python_environs=extra_python_environs) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/evaluation/rollout_worker.py", line 393, in __init__ policy_dict, policy_config) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/evaluation/rollout_worker.py", line 930, in _build_policy_map policy_map[name] = cls(obs_space, act_space, merged_conf) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/policy/eager_tf_policy.py", line 258, in __init__ self._initialize_loss_with_dummy_batch() File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/policy/eager_tf_policy.py", line 626, in _initialize_loss_with_dummy_batch explore=False) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/policy/eager_tf_policy.py", line 60, in _func return func(*args, **kwargs) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/policy/eager_tf_policy.py", line 68, in _func out = func(*args, **kwargs) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/ray/rllib/policy/eager_tf_policy.py", line 335, in compute_actions tf.convert_to_tensor(prev_action_batch) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/tensorflow/python/framework/ops.py", line 1217, in convert_to_tensor_v1 return convert_to_tensor_v2(value, dtype, preferred_dtype, name) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/tensorflow/python/framework/ops.py", line 1283, in convert_to_tensor_v2 as_ref=False) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/tensorflow/python/framework/ops.py", line 1341, in convert_to_tensor ret = conversion_func(value, dtype=dtype, name=name, as_ref=as_ref) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/tensorflow/python/ops/array_ops.py", line 1455, in _autopacking_conversion_function return _autopacking_helper(v, dtype, name or "packed") File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/tensorflow/python/ops/array_ops.py", line 1361, in _autopacking_helper return gen_array_ops.pack(list_or_tuple, name=name) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/tensorflow/python/ops/gen_array_ops.py", line 6337, in pack _ops.raise_from_not_ok_status(e, name) File "/home/coac/anaconda3/envs/rllib/lib/python3.7/site-packages/tensorflow/python/framework/ops.py", line 6653, in raise_from_not_ok_status six.raise_from(core._status_to_exception(e.code, message), None) File "<string>", line 3, in raise_from tensorflow.python.framework.errors_impl.InvalidArgumentError: cannot compute Pack as input #1(zero-based) was expected to be a int64 tensor but is a float tensor [Op:Pack] name: packed ``` If we cannot run your script, we cannot fix your issue. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
I think this is a matter of missing tf.cast() to float32 somewhere. Unfortunately, eager mode has slightly different type conversions than graph mode. Can you try finding the right place to add this conversion? Sure, I will do a PR
2020-06-04T13:52:01
ray-project/ray
8,782
ray-project__ray-8782
[ "8689" ]
41072fbcc8b3aea9fe726e960ca695f6365e9e40
diff --git a/python/ray/dashboard/dashboard.py b/python/ray/dashboard/dashboard.py --- a/python/ray/dashboard/dashboard.py +++ b/python/ray/dashboard/dashboard.py @@ -875,7 +875,7 @@ def clean_trials(self, trial_details): # round all floats for key in float_keys: - details[key] = round(details[key], 3) + details[key] = round(details[key], 12) # group together config attributes for key in config_keys:
[tune] Tune dashboard : use scientific notation When displaying parameters in the tune dashboard, they are displayer using standard float format with only 2 significant digits. For example, if your learning rate is lower than 1e-2, it appears as zero. Similarly, if your experiments return a metric whose value are in a narrow range, the displayed digits are all equal and you cannot know which one is better (for example, for an accuracy between 0 and 100% treated as a float, values displayed are between 0.00 and 1.00, with a resolution corresponding to 1%, which is really not much).
2020-06-04T15:23:32
ray-project/ray
8,802
ray-project__ray-8802
[ "8711" ]
d78757623df968ef62e12070e29e609cee470de5
diff --git a/python/ray/tune/progress_reporter.py b/python/ray/tune/progress_reporter.py --- a/python/ray/tune/progress_reporter.py +++ b/python/ray/tune/progress_reporter.py @@ -56,6 +56,11 @@ class TuneReporterBase(ProgressReporter): include in progress table. If this is a dict, the keys should be metric names and the values should be the displayed names. If this is a list, the metric name is used directly. + parameter_columns (dict[str, str]|list[str]): Names of parameters to + include in progress table. If this is a dict, the keys should + be parameter names and the values should be the displayed names. + If this is a list, the parameter name is used directly. If empty, + defaults to all available parameters. max_progress_rows (int): Maximum number of rows to print in the progress table. The progress table describes the progress of each trial. Defaults to 20. @@ -78,10 +83,12 @@ class TuneReporterBase(ProgressReporter): def __init__(self, metric_columns=None, + parameter_columns=None, max_progress_rows=20, max_error_rows=20, max_report_frequency=5): self._metric_columns = metric_columns or self.DEFAULT_COLUMNS + self._parameter_columns = parameter_columns or [] self._max_progress_rows = max_progress_rows self._max_error_rows = max_error_rows @@ -117,6 +124,29 @@ def add_metric_column(self, metric, representation=None): "of metric columns.") self._metric_columns.append(metric) + def add_parameter_column(self, parameter, representation=None): + """Adds a parameter to the existing columns. + + Args: + parameter (str): Parameter to add. This must be a parameter + specified in the configuration. + representation (str): Representation to use in table. Defaults to + `parameter`. + """ + if parameter in self._parameter_columns: + raise ValueError("Column {} already exists.".format(parameter)) + + if isinstance(self._parameter_columns, Mapping): + representation = representation or parameter + self._parameter_columns[parameter] = representation + else: + if representation is not None and representation != parameter: + raise ValueError( + "`representation` cannot differ from `parameter` " + "if this reporter was initialized with a list " + "of metric columns.") + self._parameter_columns.append(parameter) + def _progress_str(self, trials, done, *sys_info, fmt="psql", delim="\n"): """Returns full progress string. @@ -142,6 +172,7 @@ def _progress_str(self, trials, done, *sys_info, fmt="psql", delim="\n"): trial_progress_str( trials, metric_columns=self._metric_columns, + parameter_columns=self._parameter_columns, fmt=fmt, max_rows=max_progress)) messages.append(trial_errors_str(trials, fmt=fmt, max_rows=max_error)) @@ -157,6 +188,11 @@ class JupyterNotebookReporter(TuneReporterBase): include in progress table. If this is a dict, the keys should be metric names and the values should be the displayed names. If this is a list, the metric name is used directly. + parameter_columns (dict[str, str]|list[str]): Names of parameters to + include in progress table. If this is a dict, the keys should + be parameter names and the values should be the displayed names. + If this is a list, the parameter name is used directly. If empty, + defaults to all available parameters. max_progress_rows (int): Maximum number of rows to print in the progress table. The progress table describes the progress of each trial. Defaults to 20. @@ -170,12 +206,13 @@ class JupyterNotebookReporter(TuneReporterBase): def __init__(self, overwrite, metric_columns=None, + parameter_columns=None, max_progress_rows=20, max_error_rows=20, max_report_frequency=5): - super(JupyterNotebookReporter, - self).__init__(metric_columns, max_progress_rows, max_error_rows, - max_report_frequency) + super(JupyterNotebookReporter, self).__init__( + metric_columns, parameter_columns, max_progress_rows, + max_error_rows, max_report_frequency) self._overwrite = overwrite def report(self, trials, done, *sys_info): @@ -196,6 +233,11 @@ class CLIReporter(TuneReporterBase): include in progress table. If this is a dict, the keys should be metric names and the values should be the displayed names. If this is a list, the metric name is used directly. + parameter_columns (dict[str, str]|list[str]): Names of parameters to + include in progress table. If this is a dict, the keys should + be parameter names and the values should be the displayed names. + If this is a list, the parameter name is used directly. If empty, + defaults to all available parameters. max_progress_rows (int): Maximum number of rows to print in the progress table. The progress table describes the progress of each trial. Defaults to 20. @@ -208,12 +250,14 @@ class CLIReporter(TuneReporterBase): def __init__(self, metric_columns=None, + parameter_columns=None, max_progress_rows=20, max_error_rows=20, max_report_frequency=5): - super(CLIReporter, self).__init__(metric_columns, max_progress_rows, - max_error_rows, max_report_frequency) + super(CLIReporter, self).__init__(metric_columns, parameter_columns, + max_progress_rows, max_error_rows, + max_report_frequency) def report(self, trials, done, *sys_info): print(self._progress_str(trials, done, *sys_info)) @@ -241,7 +285,11 @@ def memory_debug_str(): "(or ray[debug]) to resolve)") -def trial_progress_str(trials, metric_columns, fmt="psql", max_rows=None): +def trial_progress_str(trials, + metric_columns, + parameter_columns=None, + fmt="psql", + max_rows=None): """Returns a human readable message for printing to the console. This contains a table where each row represents a trial, its parameters @@ -253,6 +301,11 @@ def trial_progress_str(trials, metric_columns, fmt="psql", max_rows=None): If this is a dict, the keys are metric names and the values are the names to use in the message. If this is a list, the metric name is used in the message directly. + parameter_columns (dict[str, str]|list[str]): Names of parameters to + include. If this is a dict, the keys are parameter names and the + values are the names to use in the message. If this is a list, + the parameter name is used in the message directly. If this is + empty, all parameters are used in the message. fmt (str): Output format (see tablefmt in tabulate API). max_rows (int): Maximum number of rows in the trial table. Defaults to unlimited. @@ -297,22 +350,39 @@ def trial_progress_str(trials, metric_columns, fmt="psql", max_rows=None): # Pre-process trials to figure out what columns to show. if isinstance(metric_columns, Mapping): - keys = list(metric_columns.keys()) + metric_keys = list(metric_columns.keys()) else: - keys = metric_columns - keys = [ - k for k in keys if any( + metric_keys = metric_columns + metric_keys = [ + k for k in metric_keys if any( t.last_result.get(k) is not None for t in trials) ] + + if not parameter_columns: + parameter_keys = sorted( + set().union(*[t.evaluated_params for t in trials])) + elif isinstance(parameter_columns, Mapping): + parameter_keys = list(parameter_columns.keys()) + else: + parameter_keys = parameter_columns + # Build trial rows. - params = sorted(set().union(*[t.evaluated_params for t in trials])) - trial_table = [_get_trial_info(trial, params, keys) for trial in trials] + trial_table = [ + _get_trial_info(trial, parameter_keys, metric_keys) for trial in trials + ] # Format column headings if isinstance(metric_columns, Mapping): - formatted_columns = [metric_columns[k] for k in keys] + formatted_metric_columns = [metric_columns[k] for k in metric_keys] + else: + formatted_metric_columns = metric_keys + if isinstance(parameter_columns, Mapping): + formatted_parameter_columns = [ + parameter_columns[k] for k in parameter_keys + ] else: - formatted_columns = keys - columns = (["Trial name", "status", "loc"] + params + formatted_columns) + formatted_parameter_columns = parameter_keys + columns = (["Trial name", "status", "loc"] + formatted_parameter_columns + + formatted_metric_columns) # Tabulate. messages.append( tabulate(trial_table, headers=columns, tablefmt=fmt, showindex=False))
diff --git a/python/ray/tune/tests/test_progress_reporter.py b/python/ray/tune/tests/test_progress_reporter.py --- a/python/ray/tune/tests/test_progress_reporter.py +++ b/python/ray/tune/tests/test_progress_reporter.py @@ -12,13 +12,13 @@ EXPECTED_RESULT_1 = """Result logdir: /foo Number of trials: 5 (1 PENDING, 3 RUNNING, 1 TERMINATED) -+--------------+------------+-------+-----+-----+ -| Trial name | status | loc | a | b | -|--------------+------------+-------+-----+-----| -| 00001 | PENDING | here | 1 | 2 | -| 00002 | RUNNING | here | 2 | 4 | -| 00000 | TERMINATED | here | 0 | 0 | -+--------------+------------+-------+-----+-----+ ++--------------+------------+-------+-----+-----+------------+ +| Trial name | status | loc | a | b | metric_1 | +|--------------+------------+-------+-----+-----+------------| +| 00001 | PENDING | here | 1 | 2 | 0.5 | +| 00002 | RUNNING | here | 2 | 4 | 1 | +| 00000 | TERMINATED | here | 0 | 0 | 0 | ++--------------+------------+-------+-----+-----+------------+ ... 2 more trials not shown (2 RUNNING)""" EXPECTED_RESULT_2 = """Result logdir: /foo @@ -33,6 +33,17 @@ | 00004 | RUNNING | here | 4 | 8 | +--------------+------------+-------+-----+-----+""" +EXPECTED_RESULT_3 = """Result logdir: /foo +Number of trials: 5 (1 PENDING, 3 RUNNING, 1 TERMINATED) ++--------------+------------+-------+-----+------------+------------+ +| Trial name | status | loc | A | Metric 1 | Metric 2 | +|--------------+------------+-------+-----+------------+------------| +| 00001 | PENDING | here | 1 | 0.5 | 0.25 | +| 00002 | RUNNING | here | 2 | 1 | 0.5 | +| 00000 | TERMINATED | here | 0 | 0 | 0 | ++--------------+------------+-------+-----+------------+------------+ +... 2 more trials not shown (2 RUNNING)""" + END_TO_END_COMMAND = """ import ray from ray import tune @@ -125,6 +136,42 @@ def f(config): | f_xxxxx_00029 | TERMINATED | | | | 9 | +---------------+------------+-------+-----+-----+-----+""" +EXPECTED_END_TO_END_AC = """Number of trials: 30 (30 TERMINATED) ++---------------+------------+-------+-----+-----+-----+ +| Trial name | status | loc | a | b | c | +|---------------+------------+-------+-----+-----+-----| +| f_xxxxx_00000 | TERMINATED | | 0 | | | +| f_xxxxx_00001 | TERMINATED | | 1 | | | +| f_xxxxx_00002 | TERMINATED | | 2 | | | +| f_xxxxx_00003 | TERMINATED | | 3 | | | +| f_xxxxx_00004 | TERMINATED | | 4 | | | +| f_xxxxx_00005 | TERMINATED | | 5 | | | +| f_xxxxx_00006 | TERMINATED | | 6 | | | +| f_xxxxx_00007 | TERMINATED | | 7 | | | +| f_xxxxx_00008 | TERMINATED | | 8 | | | +| f_xxxxx_00009 | TERMINATED | | 9 | | | +| f_xxxxx_00010 | TERMINATED | | | 0 | | +| f_xxxxx_00011 | TERMINATED | | | 1 | | +| f_xxxxx_00012 | TERMINATED | | | 2 | | +| f_xxxxx_00013 | TERMINATED | | | 3 | | +| f_xxxxx_00014 | TERMINATED | | | 4 | | +| f_xxxxx_00015 | TERMINATED | | | 5 | | +| f_xxxxx_00016 | TERMINATED | | | 6 | | +| f_xxxxx_00017 | TERMINATED | | | 7 | | +| f_xxxxx_00018 | TERMINATED | | | 8 | | +| f_xxxxx_00019 | TERMINATED | | | 9 | | +| f_xxxxx_00020 | TERMINATED | | | | 0 | +| f_xxxxx_00021 | TERMINATED | | | | 1 | +| f_xxxxx_00022 | TERMINATED | | | | 2 | +| f_xxxxx_00023 | TERMINATED | | | | 3 | +| f_xxxxx_00024 | TERMINATED | | | | 4 | +| f_xxxxx_00025 | TERMINATED | | | | 5 | +| f_xxxxx_00026 | TERMINATED | | | | 6 | +| f_xxxxx_00027 | TERMINATED | | | | 7 | +| f_xxxxx_00028 | TERMINATED | | | | 8 | +| f_xxxxx_00029 | TERMINATED | | | | 9 | ++---------------+------------+-------+-----+-----+-----+""" + class ProgressReporterTest(unittest.TestCase): def mock_trial(self, status, i): @@ -202,17 +249,38 @@ def testProgressStr(self): t.location = "here" t.config = {"a": i, "b": i * 2} t.evaluated_params = t.config - t.last_result = {"config": {"a": i, "b": i * 2}} + t.last_result = { + "config": { + "a": i, + "b": i * 2 + }, + "metric_1": i / 2, + "metric_2": i / 4 + } t.__str__ = lambda self: self.trial_id trials.append(t) - prog1 = trial_progress_str(trials, ["a", "b"], fmt="psql", max_rows=3) + # One metric, all parameters + prog1 = trial_progress_str( + trials, ["metric_1"], None, fmt="psql", max_rows=3) print(prog1) assert prog1 == EXPECTED_RESULT_1 - prog2 = trial_progress_str( - trials, ["a", "b"], fmt="psql", max_rows=None) + + # No metric, all parameters + prog2 = trial_progress_str(trials, [], None, fmt="psql", max_rows=None) print(prog2) assert prog2 == EXPECTED_RESULT_2 + # Both metrics, one parameter, all with custom representation + prog3 = trial_progress_str( + trials, { + "metric_1": "Metric 1", + "metric_2": "Metric 2" + }, {"a": "A"}, + fmt="psql", + max_rows=3) + print(prog3) + assert prog3 == EXPECTED_RESULT_3 + def testEndToEndReporting(self): try: os.environ["_TEST_TUNE_TRIAL_UUID"] = "xxxxx"
[tune] Configure parameters in output table We want to be able to configure the hyperparameters printed during the experiment execution. cc @krfricke
2020-06-05T13:33:47
ray-project/ray
8,842
ray-project__ray-8842
[ "8841" ]
3388864768fa198f27496824407f9aee16256c2f
diff --git a/python/ray/async_compat.py b/python/ray/async_compat.py --- a/python/ray/async_compat.py +++ b/python/ray/async_compat.py @@ -75,6 +75,11 @@ def done_callback(future): # Result from direct call. assert isinstance(result, AsyncGetResponse), result if result.plasma_fallback_id is None: + # If this future has result set already, we just need to + # skip the set result/exception procedure. + if user_future.done(): + return + if isinstance(result.result, ray.exceptions.RayTaskError): ray.worker.last_task_error_raise_time = time.time() user_future.set_exception(
diff --git a/python/ray/tests/test_asyncio.py b/python/ray/tests/test_asyncio.py --- a/python/ray/tests/test_asyncio.py +++ b/python/ray/tests/test_asyncio.py @@ -5,6 +5,7 @@ import sys import ray +from ray.test_utils import SignalActor def test_asyncio_actor(ray_start_regular_shared): @@ -177,6 +178,26 @@ async def plasma_get(self, plasma_object): assert ray.get(getter.plasma_get.remote([plasma_object])) == 2 [email protected] +async def test_asyncio_double_await(ray_start_regular_shared): + # This is a regression test for + # https://github.com/ray-project/ray/issues/8841 + + signal = SignalActor.remote() + waiting = signal.wait.remote() + + future = waiting.as_future() + with pytest.raises(asyncio.TimeoutError): + await asyncio.wait_for(future, timeout=0.1) + assert future.cancelled() + + # We are explicitly waiting multiple times here to test asyncio state + # override. + await signal.send.remote() + await waiting + await waiting + + if __name__ == "__main__": import pytest sys.exit(pytest.main(["-v", __file__]))
[Asyncio] InvalidStateError when multiple awaits on same oid <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? *Ray version and other system information (Python version, TensorFlow version, OS):* ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): If we cannot run your script, we cannot fix your issue. ```python import ray import time ray.init() @ray.remote def f(): time.sleep(5) oid = f.remote() await asyncio.wait_for(oid, timeout=1) await asyncio.wait_for(oid, timeout=1) ``` Output ``` Exception in callback get_async.<locals>.done_callback(<Future finis... result=None)>) at /Users/simonmo/Desktop/ray/ray/python/ray/async_compat.py:65 handle: <Handle get_async.<locals>.done_callback(<Future finis... result=None)>) at /Users/simonmo/Desktop/ray/ray/python/ray/async_compat.py:65> Traceback (most recent call last): File "/Users/simonmo/miniconda3/lib/python3.6/asyncio/events.py", line 145, in _run self._callback(*self._args) File "/Users/simonmo/Desktop/ray/ray/python/ray/async_compat.py", line 83, in done_callback user_future.set_result(result.result) asyncio.base_futures.InvalidStateError: invalid state ``` - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
Oh nice. This must be related to @allenyin55 's issue right?
2020-06-08T20:06:07
ray-project/ray
8,886
ray-project__ray-8886
[ "8878" ]
cf53b351471716e7bfa71d36368ebea9b0e219c5
diff --git a/python/ray/scripts/scripts.py b/python/ray/scripts/scripts.py --- a/python/ray/scripts/scripts.py +++ b/python/ray/scripts/scripts.py @@ -60,6 +60,7 @@ def check_no_existing_redis_clients(node_ip_address, redis_client): default=ray_constants.LOGGER_FORMAT, type=str, help=ray_constants.LOGGER_FORMAT_HELP) [email protected]_option() def cli(logging_level, logging_format): level = logging.getLevelName(logging_level.upper()) ray.utils.setup_logger(level, logging_format)
[CLI] version option ### Describe your feature request `ray --version` should output the version. Judging from the output of `ray --help`, it looks like Click is used. Version flags should be easy in click; see https://click.palletsprojects.com/en/7.x/api/#click.version_option.
This wouldn't be hard to implement; @sumanthratna would you be willing to push a quick PR? just tag me and I'll help you get it merged! Sure! Is [python/ray/scripts/scripts.py](https://github.com/ray-project/ray/edit/master/python/ray/scripts/scripts.py) the only file I need to change?
2020-06-10T20:50:05
ray-project/ray
8,898
ray-project__ray-8898
[ "8889", "8889" ]
9166e2208574f61eea3454b28ca0f2e0a45186e8
diff --git a/rllib/evaluation/sampler.py b/rllib/evaluation/sampler.py --- a/rllib/evaluation/sampler.py +++ b/rllib/evaluation/sampler.py @@ -888,15 +888,18 @@ def _process_policy_eval_results(*, to_eval, eval_results, active_episodes, pi_info_cols["state_out_{}".format(f_i)] = column policy = _get_or_raise(policies, policy_id) - # Clip if necessary (while action components are still batched). - if clip_actions: - actions = clip_action(actions, policy.action_space_struct) # Split action-component batches into single action rows. actions = unbatch(actions) for i, action in enumerate(actions): env_id = eval_data[i].env_id agent_id = eval_data[i].agent_id - actions_to_send[env_id][agent_id] = action + # Clip if necessary. + if clip_actions: + clipped_action = clip_action(action, + policy.action_space_struct) + else: + clipped_action = action + actions_to_send[env_id][agent_id] = clipped_action episode = active_episodes[env_id] episode._set_rnn_state(agent_id, [c[i] for c in rnn_out_cols]) episode._set_last_pi_info( diff --git a/rllib/tuned_examples/debug_learning_failure_git_bisect.py b/rllib/tuned_examples/debug_learning_failure_git_bisect.py new file mode 100644 --- /dev/null +++ b/rllib/tuned_examples/debug_learning_failure_git_bisect.py @@ -0,0 +1,88 @@ +"""Example of testing, whether RLlib can still learn with a certain config. + +Can be used with git bisect to find the faulty commit responsible for a +learning failure. Produces an error if the given reward is not reached within +the stopping criteria (training iters or timesteps) allowing git bisect to +properly analyze and find the faulty commit. + +Run as follows using a simple command line config: +$ python debug_learning_failure_git_bisect.py --config '{...}' + --env CartPole-v0 --run PPO --stop-reward=180 --stop-iters=100 + +With a yaml file: +$ python debug_learning_failure_git_bisect.py -f [yaml file] --stop-reward=180 + --stop-iters=100 + +Within git bisect: +$ git bisect start +$ git bisect bad +$ git bisect good [some previous commit we know was good] +$ git bisect run python debug_learning_failure_git_bisect.py [... options] +""" +import argparse +import json +import yaml + +import ray +from ray import tune +from ray.rllib.utils import try_import_tf +from ray.rllib.utils.test_utils import check_learning_achieved + +tf = try_import_tf() + +parser = argparse.ArgumentParser() +parser.add_argument("--run", type=str, default=None) +parser.add_argument("--torch", action="store_true") +parser.add_argument("--stop-iters", type=int, default=None) +parser.add_argument("--stop-timesteps", type=int, default=None) +parser.add_argument("--stop-reward", type=float, default=None) +parser.add_argument("-f", type=str, default=None) +parser.add_argument("--config", type=str, default=None) +parser.add_argument("--env", type=str, default=None) + +if __name__ == "__main__": + run = None + + args = parser.parse_args() + + # Explicit yaml config file. + if args.f: + with open(args.f, "r") as fp: + experiment_config = yaml.load(fp) + experiment_config = experiment_config[next( + iter(experiment_config))] + config = experiment_config.get("config", {}) + config["env"] = experiment_config.get("env") + run = experiment_config.pop("run") + # JSON string on command line. + else: + config = json.loads(args.config) + assert args.env + config["env"] = args.env + + # Explicit run. + if args.run: + run = args.run + + # Explicit --torch framework. + if args.torch: + config["framework"] = "torch" + # Framework not specified in config, try to infer it. + if "framework" not in config: + config["framework"] = "torch" if args.torch else "tf" + + ray.init() + + stop = {} + if args.stop_iters: + stop["training_iteration"] = args.stop_iters + if args.stop_timesteps: + stop["timesteps_total"] = args.stop_timesteps + if args.stop_reward: + stop["episode_reward_mean"] = args.stop_reward + + results = tune.run(run, stop=stop, config=config) + + check_learning_achieved(results, args.stop_reward) + + ray.shutdown()
[RLlib] PPO not learning in complex cont. action environments. <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> PPO is currently not learning e.g. HalfCheetah-v2 using the `tuned_examples/ppo/halfcheetah-ppo.yaml` config (on neither tf nor torch). ray-0.9.0dev tf2.2.0 torch1.5.0 The problem does not occur on ray/rllib 0.8.5. *Ray version and other system information (Python version, TensorFlow version, OS):* ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): If we cannot run your script, we cannot fix your issue. - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html). [RLlib] PPO not learning in complex cont. action environments. <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> PPO is currently not learning e.g. HalfCheetah-v2 using the `tuned_examples/ppo/halfcheetah-ppo.yaml` config (on neither tf nor torch). ray-0.9.0dev tf2.2.0 torch1.5.0 The problem does not occur on ray/rllib 0.8.5. *Ray version and other system information (Python version, TensorFlow version, OS):* ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): If we cannot run your script, we cannot fix your issue. - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
<img width="849" alt="Screen Shot 2020-06-10 at 2 27 42 PM" src="https://user-images.githubusercontent.com/20640975/84320506-a1ce9a00-ab26-11ea-9dd1-1ed216d14682.png"> <img width="880" alt="Screen Shot 2020-06-10 at 2 27 27 PM" src="https://user-images.githubusercontent.com/20640975/84320512-a430f400-ab26-11ea-997b-cd9ae4d82422.png"> In addition, `pong-ppo.yaml` and `pong-impala-vectorized.yaml` are not learning either. <img width="849" alt="Screen Shot 2020-06-10 at 2 27 42 PM" src="https://user-images.githubusercontent.com/20640975/84320506-a1ce9a00-ab26-11ea-9dd1-1ed216d14682.png"> <img width="880" alt="Screen Shot 2020-06-10 at 2 27 27 PM" src="https://user-images.githubusercontent.com/20640975/84320512-a430f400-ab26-11ea-997b-cd9ae4d82422.png"> In addition, `pong-ppo.yaml` and `pong-impala-vectorized.yaml` are not learning either.
2020-06-11T09:09:31
ray-project/ray
8,933
ray-project__ray-8933
[ "8919", "8919" ]
19cc1ae781483ff509176a3f3ad3ff09a7084f8b
diff --git a/rllib/train.py b/rllib/train.py --- a/rllib/train.py +++ b/rllib/train.py @@ -149,6 +149,7 @@ def run(args, parser): args.experiment_name: { # i.e. log to ~/ray_results/default "run": args.run, "checkpoint_freq": args.checkpoint_freq, + "checkpoint_at_end": args.checkpoint_at_end, "keep_checkpoints_num": args.keep_checkpoints_num, "checkpoint_score_attr": args.checkpoint_score_attr, "local_dir": args.local_dir,
diff --git a/rllib/tests/test_checkpoint_restore.py b/rllib/tests/test_checkpoint_restore.py --- a/rllib/tests/test_checkpoint_restore.py +++ b/rllib/tests/test_checkpoint_restore.py @@ -124,6 +124,9 @@ def ckpt_restore_test(alg_name, tfe=False): if abs(a1 - a2) > .1: raise AssertionError("algo={} [a1={} a2={}]".format( alg_name, a1, a2)) + # Stop both Trainers. + alg1.stop() + alg2.stop() class TestCheckpointRestore(unittest.TestCase):
[rllib] rllib train ... --checkpoint-at-end flag is ignored <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? Ray 0.8.5 Python 3.7 MacOS Catalina PyTorch 1.4.0 No checkpoint is produced when the `--checkpoint-at-end` flag is used. Adding `--checkpoint-freq 10` does cause checkpoints to be saved. ### Reproduction (REQUIRED) ```shell rllib train --run PPO --env CartPole-v0 --stop='{"training_iteration": 25}' --ray-address auto --checkpoint-at-end ``` No checkpoint directory exists under `~/ray_results/default/PPO...` Add the `--checkpoint-freq 10` flag: ```shell rllib train --run PPO --env CartPole-v0 --stop='{"training_iteration": 25}' --ray-address auto --checkpoint-at-end --checkpoint-freq 10 ``` Now there are `checkpoint_10` and `checkpoint_20` directories, but not a `checkpoint_25` at the end. Could the choice of stop criteria, `training_iteration`, have something to do with it? > **Note:** Being persnickety, it bugs me that the final directory path for the checkpoints is `.../checkpoint_20/checkpoint-20` (underscore vs. dash). How about one or the other? - [x] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html). [rllib] rllib train ... --checkpoint-at-end flag is ignored <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? Ray 0.8.5 Python 3.7 MacOS Catalina PyTorch 1.4.0 No checkpoint is produced when the `--checkpoint-at-end` flag is used. Adding `--checkpoint-freq 10` does cause checkpoints to be saved. ### Reproduction (REQUIRED) ```shell rllib train --run PPO --env CartPole-v0 --stop='{"training_iteration": 25}' --ray-address auto --checkpoint-at-end ``` No checkpoint directory exists under `~/ray_results/default/PPO...` Add the `--checkpoint-freq 10` flag: ```shell rllib train --run PPO --env CartPole-v0 --stop='{"training_iteration": 25}' --ray-address auto --checkpoint-at-end --checkpoint-freq 10 ``` Now there are `checkpoint_10` and `checkpoint_20` directories, but not a `checkpoint_25` at the end. Could the choice of stop criteria, `training_iteration`, have something to do with it? > **Note:** Being persnickety, it bugs me that the final directory path for the checkpoints is `.../checkpoint_20/checkpoint-20` (underscore vs. dash). How about one or the other? - [x] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-06-14T13:52:06
ray-project/ray
8,953
ray-project__ray-8953
[ "4029" ]
4bc1d7c043cca03a743fbeb98b0434fecafa7001
diff --git a/python/ray/log_monitor.py b/python/ray/log_monitor.py --- a/python/ray/log_monitor.py +++ b/python/ray/log_monitor.py @@ -5,6 +5,7 @@ import logging import os import platform +import re import shutil import time import traceback @@ -18,19 +19,24 @@ # entry/init points. logger = logging.getLogger(__name__) +# First group is worker id. Second group is job id. +JOB_LOG_PATTERN = re.compile(".*worker-([0-9a-f]{40})-(\d+)") + class LogFileInfo: def __init__(self, filename=None, size_when_last_opened=None, file_position=None, - file_handle=None): + file_handle=None, + job_id=None): assert (filename is not None and size_when_last_opened is not None and file_position is not None) self.filename = filename self.size_when_last_opened = size_when_last_opened self.file_position = file_position self.file_handle = file_handle + self.job_id = job_id self.worker_pid = None @@ -116,13 +122,20 @@ def update_log_filenames(self): for file_path in log_file_paths + raylet_err_paths: if os.path.isfile( file_path) and file_path not in self.log_filenames: + job_match = JOB_LOG_PATTERN.match(file_path) + if job_match: + job_id = job_match.group(2) + else: + job_id = None + self.log_filenames.add(file_path) self.closed_file_infos.append( LogFileInfo( filename=file_path, size_when_last_opened=0, file_position=0, - file_handle=None)) + file_handle=None, + job_id=job_id)) log_filename = os.path.basename(file_path) logger.info("Beginning to track file {}".format(log_filename)) @@ -231,6 +244,7 @@ def check_log_files_and_publish_updates(self): json.dumps({ "ip": self.ip, "pid": file_info.worker_pid, + "job": file_info.job_id, "lines": lines_to_publish })) anything_published = True diff --git a/python/ray/utils.py b/python/ray/utils.py --- a/python/ray/utils.py +++ b/python/ray/utils.py @@ -403,6 +403,7 @@ def create_and_init_new_worker_log(path, worker_pid): Args: path (str): The name/path of the file to be opened. + worker_pid (int): The pid of the worker process. Returns: A file-like object which can be written to. diff --git a/python/ray/worker.py b/python/ray/worker.py --- a/python/ray/worker.py +++ b/python/ray/worker.py @@ -953,13 +953,15 @@ def stderr_setter(x): return stdout_path, stderr_path -def print_logs(redis_client, threads_stopped): +def print_logs(redis_client, threads_stopped, job_id): """Prints log messages from workers on all of the nodes. Args: redis_client: A client to the primary Redis shard. threads_stopped (threading.Event): A threading event used to signal to the thread that it should exit. + job_id (JobID): The id of the driver's job + """ pubsub_client = redis_client.pubsub(ignore_subscribe_messages=True) pubsub_client.subscribe(ray.gcs_utils.LOG_FILE_CHANNEL) @@ -981,9 +983,20 @@ def print_logs(redis_client, threads_stopped): threads_stopped.wait(timeout=0.01) continue num_consecutive_messages_received += 1 + if (num_consecutive_messages_received % 100 == 0 + and num_consecutive_messages_received > 0): + logger.warning( + "The driver may not be able to keep up with the " + "stdout/stderr of the workers. To avoid forwarding logs " + "to the driver, use 'ray.init(log_to_driver=False)'.") data = json.loads(ray.utils.decode(msg["data"])) + # Don't show logs from other drivers. + if data["job"] and ray.utils.binary_to_hex( + job_id.binary()) != data["job"]: + continue + def color_for(data): if data["pid"] == "raylet": return colorama.Fore.YELLOW @@ -1001,12 +1014,6 @@ def color_for(data): colorama.Style.DIM, color_for(data), data["pid"], data["ip"], colorama.Style.RESET_ALL, line)) - if (num_consecutive_messages_received % 100 == 0 - and num_consecutive_messages_received > 0): - logger.warning( - "The driver may not be able to keep up with the " - "stdout/stderr of the workers. To avoid forwarding logs " - "to the driver, use 'ray.init(log_to_driver=False)'.") except (OSError, redis.exceptions.ConnectionError) as e: logger.error("print_logs: {}".format(e)) finally: @@ -1310,7 +1317,7 @@ def connect(node, worker.logger_thread = threading.Thread( target=print_logs, name="ray_print_logs", - args=(worker.redis_client, worker.threads_stopped)) + args=(worker.redis_client, worker.threads_stopped, job_id)) worker.logger_thread.daemon = True worker.logger_thread.start()
diff --git a/python/ray/test_utils.py b/python/ray/test_utils.py --- a/python/ray/test_utils.py +++ b/python/ray/test_utils.py @@ -130,7 +130,9 @@ def run_string_as_driver_nonblocking(driver_script): f.write(driver_script.encode("ascii")) f.flush() return subprocess.Popen( - [sys.executable, f.name], stdout=subprocess.PIPE) + [sys.executable, f.name], + stdout=subprocess.PIPE, + stderr=subprocess.PIPE) def flat_errors(): diff --git a/python/ray/tests/test_multi_node.py b/python/ray/tests/test_multi_node.py --- a/python/ray/tests/test_multi_node.py +++ b/python/ray/tests/test_multi_node.py @@ -5,13 +5,9 @@ import ray from ray.test_utils import ( - RayTestTimeoutException, - run_string_as_driver, - run_string_as_driver_nonblocking, - wait_for_children_of_pid, - wait_for_children_of_pid_to_exit, - kill_process_by_name, -) + RayTestTimeoutException, run_string_as_driver, + run_string_as_driver_nonblocking, wait_for_children_of_pid, + wait_for_children_of_pid_to_exit, kill_process_by_name, Semaphore) def test_error_isolation(call_ray_start): @@ -634,6 +630,77 @@ def f(): ray.get(f.remote()) +def test_multi_driver_logging(ray_start_regular): + address_info = ray_start_regular + address = address_info["redis_address"] + + # ray.init(address=address) + driver1_wait = Semaphore.options(name="driver1_wait").remote(value=0) + driver2_wait = Semaphore.options(name="driver2_wait").remote(value=0) + main_wait = Semaphore.options(name="main_wait").remote(value=0) + + # Params are address, semaphore name, output1, output2 + driver_script_template = """ +import ray +import sys +from ray.test_utils import Semaphore + [email protected](num_cpus=0) +def remote_print(s, file=None): + print(s, file=file) + +ray.init(address="{}") + +driver_wait = ray.get_actor("{}") +main_wait = ray.get_actor("main_wait") + +ray.get(main_wait.release.remote()) +ray.get(driver_wait.acquire.remote()) + +s1 = "{}" +ray.get(remote_print.remote(s1)) + +ray.get(main_wait.release.remote()) +ray.get(driver_wait.acquire.remote()) + +s2 = "{}" +ray.get(remote_print.remote(s2)) + +ray.get(main_wait.release.remote()) + """ + + p1 = run_string_as_driver_nonblocking( + driver_script_template.format(address, "driver1_wait", "1", "2")) + p2 = run_string_as_driver_nonblocking( + driver_script_template.format(address, "driver2_wait", "3", "4")) + + ray.get(main_wait.acquire.remote()) + ray.get(main_wait.acquire.remote()) + # At this point both of the other drivers are fully initialized. + + ray.get(driver1_wait.release.remote()) + ray.get(driver2_wait.release.remote()) + + # At this point driver1 should receive '1' and driver2 '3' + ray.get(main_wait.acquire.remote()) + ray.get(main_wait.acquire.remote()) + + ray.get(driver1_wait.release.remote()) + ray.get(driver2_wait.release.remote()) + + # At this point driver1 should receive '2' and driver2 '4' + ray.get(main_wait.acquire.remote()) + ray.get(main_wait.acquire.remote()) + + driver1_out = p1.stdout.read().decode("ascii").split("\n") + driver2_out = p2.stdout.read().decode("ascii").split("\n") + + assert driver1_out[0][-1] == "1" + assert driver1_out[1][-1] == "2" + assert driver2_out[0][-1] == "3" + assert driver2_out[1][-1] == "4" + + if __name__ == "__main__": import pytest import sys
Stdout/stderr from all drivers are streamed to all other drivers. As of https://github.com/ray-project/ray/pull/3892, the stdout/stderr of all drivers is streamed to all other drivers. An individual driver can opt out by calling `ray.init(..., log_to_driver=False)`. To reproduce this, do the following. 1. Start Ray. ``` ray start --head --redis-port=6379 ``` 2. Start a driver that prints stuff. ```python import ray import sys import time ray.init(redis_address='localhost:6379') @ray.remote def f(): while True: print('hi') sys.stdout.flush() time.sleep(0.1) f.remote() ``` 3. Start another driver, and see that it receives the stdout from the first driver. ```python import ray ray.init(redis_address='localhost:6379') ``` ### Some options to address this. 1. Workers can belong to specific drivers (this means logs will be complicated for cross driver applications, e.g., named actors). 2. Workers can redirect there stdout/stderr on a per-task basis. 3. Workers can print extra annotations in their stdout/stderr which the log monitor can parse and use to choose which drivers to publish to.
I like option 3. For option 2 what kind of api were you thinking about? Like an additional keyword in ray.remote decoration? I am already making use of this feature a little and find it handy to include timestamp and function name to each stdout. Not saying these should be added by default but giving that as an example. Option 2 would not require the user to do anything. Instead, there would one log file per worker/driver combination, and whenever a worker executes a task for a given driver, it logs to the file for that worker/driver combination. This would all be done under the hood. 2 or 3 sound fine to me then. I am not doing cross driver work yet so that isn't an issue for me. If this isn't going to be fixed soon, can we at least document it somewhere prominently? This was very confusing to me for a while: now I know what's causing it, it's largely a non-issue, as I can look at output from my other logging methods. Thanks @AdamGleave for bringing this up. I'll take a look at fixing it this week.
2020-06-15T23:03:05
ray-project/ray
8,964
ray-project__ray-8964
[ "8713" ]
508149b3c3004ee5ef8b02bd797e26ceaac896b2
diff --git a/python/ray/util/iter.py b/python/ray/util/iter.py --- a/python/ray/util/iter.py +++ b/python/ray/util/iter.py @@ -203,9 +203,9 @@ def for_each(self, fn: Callable[[T], U], max_concurrency=1, `max_concurrency` should be used to achieve a high degree of parallelism without the overhead of increasing the number of shards - (which are actor based). This provides the semantic guarantee that - `fn(x_i)` will _begin_ executing before `fn(x_{i+1})` (but not - necessarily finish first) + (which are actor based). If `max_concurrency` is not 1, this function + provides no semantic guarantees on the output order. + Results will be returned as soon as they are ready. A performance note: When executing concurrently, this function maintains its own internal buffer. If `num_async` is `n` and @@ -234,6 +234,7 @@ def for_each(self, fn: Callable[[T], U], max_concurrency=1, ... [0, 2, 4, 8] """ + assert max_concurrency >= 0, "max_concurrency must be non-negative." return self._with_transform( lambda local_it: local_it.for_each(fn, max_concurrency, resources), ".for_each()") @@ -765,23 +766,13 @@ def apply_foreach(it): if isinstance(item, _NextValueNotReady): yield item else: - finished, remaining = ray.wait(cur, timeout=0) - if max_concurrency and len( - remaining) >= max_concurrency: - ray.wait(cur, num_returns=(len(finished) + 1)) + if max_concurrency and len(cur) >= max_concurrency: + finished, cur = ray.wait(cur) + yield from ray.get(finished) cur.append(remote_fn(item)) - - while len(cur) > 0: - to_yield = cur[0] - finished, remaining = ray.wait( - [to_yield], timeout=0) - if finished: - cur.pop(0) - yield ray.get(to_yield) - else: - break - - yield from ray.get(cur) + while cur: + finished, cur = ray.wait(cur) + yield from ray.get(finished) if hasattr(fn, LocalIterator.ON_FETCH_START_HOOK_NAME): unwrapped = apply_foreach
diff --git a/python/ray/test_utils.py b/python/ray/test_utils.py --- a/python/ray/test_utils.py +++ b/python/ray/test_utils.py @@ -248,7 +248,7 @@ def __init__(self, value=1): self._sema = asyncio.Semaphore(value=value) async def acquire(self): - self._sema.acquire() + await self._sema.acquire() async def release(self): self._sema.release() diff --git a/python/ray/tests/test_iter.py b/python/ray/tests/test_iter.py --- a/python/ray/tests/test_iter.py +++ b/python/ray/tests/test_iter.py @@ -159,7 +159,7 @@ def test_for_each(ray_start_regular_shared): assert list(it.gather_sync()) == [0, 4, 2, 6] -def test_for_each_concur(ray_start_regular_shared): +def test_for_each_concur_async(ray_start_regular_shared): main_wait = Semaphore.remote(value=0) test_wait = Semaphore.remote(value=0) @@ -169,15 +169,18 @@ def task(x): ray.get(test_wait.acquire.remote()) return i + 10 - @ray.remote(num_cpus=0.1) + @ray.remote(num_cpus=0.01) def to_list(it): return list(it) it = from_items( [(i, main_wait, test_wait) for i in range(8)], num_shards=2) - it = it.for_each(task, max_concurrency=2, resources={"num_cpus": 0.1}) + it = it.for_each(task, max_concurrency=2, resources={"num_cpus": 0.01}) + + list_promise = to_list.remote(it.gather_async()) for i in range(4): + assert i in [0, 1, 2, 3] ray.get(main_wait.acquire.remote()) # There should be exactly 4 tasks executing at this point. @@ -189,12 +192,49 @@ def to_list(it): assert ray.get(main_wait.locked.remote()) is True, "Too much parallelism" # Finish everything and make sure the output matches a regular iterator. - for i in range(3): + for i in range(7): + ray.get(test_wait.release.remote()) + + assert repr( + it) == "ParallelIterator[from_items[tuple, 8, shards=2].for_each()]" + result_list = ray.get(list_promise) + assert set(result_list) == set(range(10, 18)) + + +def test_for_each_concur_sync(ray_start_regular_shared): + main_wait = Semaphore.remote(value=0) + test_wait = Semaphore.remote(value=0) + + def task(x): + i, main_wait, test_wait = x + ray.get(main_wait.release.remote()) + ray.get(test_wait.acquire.remote()) + return i + 10 + + @ray.remote(num_cpus=0.01) + def to_list(it): + return list(it) + + it = from_items( + [(i, main_wait, test_wait) for i in range(8)], num_shards=2) + it = it.for_each(task, max_concurrency=2, resources={"num_cpus": 0.01}) + + list_promise = to_list.remote(it.gather_sync()) + + for i in range(4): + assert i in [0, 1, 2, 3] + ray.get(main_wait.acquire.remote()) + + # There should be exactly 4 tasks executing at this point. + assert ray.get(main_wait.locked.remote()) is True, "Too much parallelism" + + for i in range(8): ray.get(test_wait.release.remote()) assert repr( it) == "ParallelIterator[from_items[tuple, 8, shards=2].for_each()]" - assert ray.get(to_list.remote(it.gather_sync())) == list(range(10, 18)) + result_list = ray.get(list_promise) + assert set(result_list) == set(range(10, 18)) def test_combine(ray_start_regular_shared):
[ParallelIterator] for_each_concur leaks memory <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? `for_each_concur` appears to leak memory. After iterating over items, they appear to remain pinned in memory. ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): ``` import ray from ray.util.iter import from_items import numpy as np ray.init() def expand(x): print("Expanding...") return np.array([x for _ in range(2**26)]) def reduce(x): return np.sum(x)/x.shape[0] it = from_items(list(range(1000))) it = it.for_each(expand, max_concurrency=2) for x in it.gather_async(): print(reduce(x)) ``` If we cannot run your script, we cannot fix your issue. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
@wuisawesome Would you mind taking over this issue? For P1 tasks, each assignees will be responsible for resolving them within a month.
2020-06-16T06:42:02
ray-project/ray
9,020
ray-project__ray-9020
[ "7519" ]
c79a49488cde8d11919ab678e409374327695a53
diff --git a/python/ray/autoscaler/updater.py b/python/ray/autoscaler/updater.py --- a/python/ray/autoscaler/updater.py +++ b/python/ray/autoscaler/updater.py @@ -147,6 +147,44 @@ def remote_shell_command_str(self): self.node_id) +class SSHOptions: + def __init__(self, ssh_key, control_path=None, **kwargs): + self.ssh_key = ssh_key + self.arg_dict = { + # Supresses initial fingerprint verification. + "StrictHostKeyChecking": "no", + # SSH IP and fingerprint pairs no longer added to known_hosts. + # This is to remove a "REMOTE HOST IDENTIFICATION HAS CHANGED" + # warning if a new node has the same IP as a previously + # deleted node, because the fingerprints will not match in + # that case. + "UserKnownHostsFile": os.devnull, + # Try fewer extraneous key pairs. + "IdentitiesOnly": "yes", + # Abort if port forwarding fails (instead of just printing to + # stderr). + "ExitOnForwardFailure": "yes", + # Quickly kill the connection if network connection breaks (as + # opposed to hanging/blocking). + "ServerAliveInterval": 5, + "ServerAliveCountMax": 3 + } + if control_path: + self.arg_dict.update({ + "ControlMaster": "auto", + "ControlPath": "{}/%C".format(control_path), + "ControlPersist": "10s", + }) + self.arg_dict.update(kwargs) + + def to_ssh_options_list(self, *, timeout=60): + self.arg_dict["ConnectTimeout"] = "{}s".format(timeout) + return ["-i", self.ssh_key] + [ + x for y in (["-o", "{}={}".format(k, v)] + for k, v in self.arg_dict.items()) for x in y + ] + + class SSHCommandRunner: def __init__(self, log_prefix, node_id, provider, auth_config, cluster_name, process_runner, use_internal_ip): @@ -166,36 +204,8 @@ def __init__(self, log_prefix, node_id, provider, auth_config, self.ssh_user = auth_config["ssh_user"] self.ssh_control_path = ssh_control_path self.ssh_ip = None - - def get_default_ssh_options(self, connect_timeout): - OPTS = [ - ("ConnectTimeout", "{}s".format(connect_timeout)), - # Supresses initial fingerprint verification. - ("StrictHostKeyChecking", "no"), - # SSH IP and fingerprint pairs no longer added to known_hosts. - # This is to remove a "REMOTE HOST IDENTIFICATION HAS CHANGED" - # warning if a new node has the same IP as a previously - # deleted node, because the fingerprints will not match in - # that case. - ("UserKnownHostsFile", os.devnull), - ("ControlMaster", "auto"), - ("ControlPath", "{}/%C".format(self.ssh_control_path)), - ("ControlPersist", "10s"), - # Try fewer extraneous key pairs. - ("IdentitiesOnly", "yes"), - # Abort if port forwarding fails (instead of just printing to - # stderr). - ("ExitOnForwardFailure", "yes"), - # Quickly kill the connection if network connection breaks (as - # opposed to hanging/blocking). - ("ServerAliveInterval", 5), - ("ServerAliveCountMax", 3), - ] - - return ["-i", self.ssh_private_key] + [ - x for y in (["-o", "{}={}".format(k, v)] for k, v in OPTS) - for x in y - ] + self.base_ssh_options = SSHOptions(self.ssh_private_key, + self.ssh_control_path) def get_node_ip(self): if self.use_internal_ip: @@ -241,7 +251,14 @@ def run(self, exit_on_fail=False, port_forward=None, with_output=False, + ssh_options_override=None, **kwargs): + ssh_options = ssh_options_override or self.base_ssh_options + + assert isinstance( + ssh_options, SSHOptions + ), "ssh_options must be of type SSHOptions, got {}".format( + type(ssh_options)) self.set_ssh_ip_if_required() @@ -255,7 +272,7 @@ def run(self, "{} -> localhost:{}".format(local, remote)) ssh += ["-L", "{}:localhost:{}".format(remote, local)] - final_cmd = ssh + self.get_default_ssh_options(timeout) + [ + final_cmd = ssh + ssh_options.to_ssh_options_list(timeout=timeout) + [ "{}@{}".format(self.ssh_user, self.ssh_ip) ] if cmd: @@ -286,16 +303,20 @@ def run_rsync_up(self, source, target): self.set_ssh_ip_if_required() self.process_runner.check_call([ "rsync", "--rsh", - " ".join(["ssh"] + self.get_default_ssh_options(120)), "-avz", - source, "{}@{}:{}".format(self.ssh_user, self.ssh_ip, target) + " ".join(["ssh"] + + self.base_ssh_options.to_ssh_options_list(timeout=120)), + "-avz", source, "{}@{}:{}".format(self.ssh_user, self.ssh_ip, + target) ]) def run_rsync_down(self, source, target): self.set_ssh_ip_if_required() self.process_runner.check_call([ "rsync", "--rsh", - " ".join(["ssh"] + self.get_default_ssh_options(120)), "-avz", - "{}@{}:{}".format(self.ssh_user, self.ssh_ip, source), target + " ".join(["ssh"] + + self.base_ssh_options.to_ssh_options_list(timeout=120)), + "-avz", "{}@{}:{}".format(self.ssh_user, self.ssh_ip, + source), target ]) def remote_shell_command_str(self): @@ -309,6 +330,7 @@ def __init__(self, docker_config, **common_args): self.docker_name = docker_config["container_name"] self.docker_config = docker_config self.home_dir = None + self.check_docker_installed() self.shutdown = False def run(self, @@ -318,6 +340,7 @@ def run(self, port_forward=None, with_output=False, run_env=True, + ssh_options_override=None, **kwargs): if run_env == "auto": run_env = "host" if cmd.find("docker") == 0 else "docker" @@ -335,7 +358,23 @@ def run(self, timeout=timeout, exit_on_fail=exit_on_fail, port_forward=None, - with_output=False) + with_output=False, + ssh_options_override=ssh_options_override) + + def check_docker_installed(self): + try: + self.ssh_command_runner.run("command -v docker") + return + except Exception: + install_commands = [ + "curl -fsSL https://get.docker.com -o get-docker.sh", + "sudo sh get-docker.sh", "sudo usermod -aG docker $USER", + "sudo systemctl restart docker -f" + ] + logger.error( + "Docker not installed. You can install Docker by adding the " + "following commands to 'initialization_commands':\n" + + "\n".join(install_commands)) def shutdown_after_next_cmd(self): self.shutdown = True @@ -422,6 +461,7 @@ def __init__(self, self.setup_commands = setup_commands self.ray_start_commands = ray_start_commands self.runtime_hash = runtime_hash + self.auth_config = auth_config def run(self): logger.info(self.log_prefix + @@ -516,7 +556,10 @@ def do_update(self): self.log_prefix + "Initialization commands", show_status=True): for cmd in self.initialization_commands: - self.cmd_runner.run(cmd) + self.cmd_runner.run( + cmd, + ssh_options_override=SSHOptions( + self.auth_config.get("ssh_private_key"))) with LogTimer( self.log_prefix + "Setup commands", show_status=True):
[autoscaler] How to install docker within the setup yaml file? I want to install docker and run my ray cluster on it. This is what my initialization commands look like: ``` initialization_commands: [ "sudo apt update", "sudo apt install docker.io -y", "sudo usermod -a -G docker ubuntu", "sudo service docker restart", # Docker should be restarted here, but I do not know how "docker login -u $DOCKER_USER -p $DOCKER_PW", ] ``` Docker needs to be restarted, or `newgrp docker` has to be run, but I do not know how to do it in this context, because any command that does this like `newgrp` opens up a subshell and breaks the functionality here. How do I do this? Here is a stackoverflow question I posted to the same effect: https://stackoverflow.com/questions/60609535/how-to-install-and-use-docker-in-one-shell-script
Hi, @preeth1. Do you solve the problem?
2020-06-18T21:35:50
ray-project/ray
9,057
ray-project__ray-9057
[ "9056" ]
981f67bfb01725f1e153c881f81082fcba0ac223
diff --git a/python/ray/tune/config_parser.py b/python/ray/tune/config_parser.py --- a/python/ray/tune/config_parser.py +++ b/python/ray/tune/config_parser.py @@ -73,13 +73,13 @@ def make_parser(parser_creator=None, **kwargs): help="Whether to checkpoint at the end of the experiment. " "Default is False.") parser.add_argument( - "--no-sync-on-checkpoint", + "--sync-on-checkpoint", action="store_true", - help="Disable sync-down of trial checkpoint, which is enabled by " - "default to guarantee recoverability. If set, checkpoint syncing from " - "worker to driver is asynchronous. Set this only if synchronous " - "checkpointing is too slow and trial restoration failures can be " - "tolerated") + help="Enable sync-down of trial checkpoint to guarantee " + "recoverability. If unset, checkpoint syncing from worker " + "to driver is asynchronous, so unset this only if synchronous " + "checkpointing is too slow and trial restoration failures " + "can be tolerated.") parser.add_argument( "--keep-checkpoints-num", default=None, @@ -182,7 +182,7 @@ def create_trial_from_spec(spec, output_path, parser, **trial_kwargs): remote_checkpoint_dir=spec.get("remote_checkpoint_dir"), checkpoint_freq=args.checkpoint_freq, checkpoint_at_end=args.checkpoint_at_end, - sync_on_checkpoint=not args.no_sync_on_checkpoint, + sync_on_checkpoint=args.sync_on_checkpoint, keep_checkpoints_num=args.keep_checkpoints_num, checkpoint_score_attr=args.checkpoint_score_attr, export_formats=spec.get("export_formats", []),
[tune] sync_on_checkpoint ignored when list of experiments passed to tune.run ### What is the problem? When passing a list of predefined Experiments to tune via the `run` function, the `sync_on_checkpoint` flag is ignored. **Reason for this behaviour:** The `make_parser` function takes in `no_sync_on_checkpoint`; it's set to False by default. https://github.com/ray-project/ray/blob/981f67bfb01725f1e153c881f81082fcba0ac223/python/ray/tune/config_parser.py#L75-L77 The create_trial_from_spec function gets a `sync_on_checkpoint` from the Experiment spec, but that key is never used by the parser. So the following line makes the `sync_on_checkpoint` of the returned `Trial` object `True` by default, thus ignoring the value of `sync_on_checkpoint` in the incoming Experiment spec. https://github.com/ray-project/ray/blob/981f67bfb01725f1e153c881f81082fcba0ac223/python/ray/tune/config_parser.py#L185 *Ray Version: All versions starting from 0.7.7 when the feature was released.* ### Reproduction (REQUIRED) ```python3 import ray from ray import tune from ray.tune import Experiment def func(config): pass sync_true_obj = Experiment("test_sync", func, sync_on_checkpoint=True) sync_false_obj = Experiment("test_sync", func, sync_on_checkpoint=False) ls = [sync_true_obj, sync_false_obj] ray.init(local_mode=True) analysis = tune.run(ls) ``` To check the error, add the following lines just before the `return Trial(` line in `config_parser.py` ```python3 print(args) print(args.no_sync_on_checkpoint) ``` https://github.com/ray-project/ray/blob/981f67bfb01725f1e153c881f81082fcba0ac223/python/ray/tune/config_parser.py#L173 - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-06-20T12:52:33
ray-project/ray
9,108
ray-project__ray-9108
[ "1221" ]
f77c638d6dc0ba4533b94fa704e46cf478479aec
diff --git a/python/ray/worker.py b/python/ray/worker.py --- a/python/ray/worker.py +++ b/python/ray/worker.py @@ -399,6 +399,10 @@ def get_gpu_ids(): assigned_ids = [ global_worker.original_gpu_ids[gpu_id] for gpu_id in assigned_ids ] + # Give all GPUs in local_mode. + if global_worker.mode == LOCAL_MODE: + max_gpus = global_worker.node.get_resource_spec().num_gpus + return global_worker.original_gpu_ids[:max_gpus] return assigned_ids
diff --git a/python/ray/tests/test_advanced_2.py b/python/ray/tests/test_advanced_2.py --- a/python/ray/tests/test_advanced_2.py +++ b/python/ray/tests/test_advanced_2.py @@ -661,6 +661,26 @@ def g(): ray.get([g.remote() for _ in range(100)]) +def test_local_mode_gpus(save_gpu_ids_shutdown_only): + allowed_gpu_ids = [4, 5, 6, 7, 8] + os.environ["CUDA_VISIBLE_DEVICES"] = ",".join( + [str(i) for i in allowed_gpu_ids]) + + from importlib import reload + reload(ray.worker) + + ray.init(num_gpus=3, local_mode=True) + + @ray.remote + def f(): + gpu_ids = ray.get_gpu_ids() + assert len(gpu_ids) == 3 + for gpu in gpu_ids: + assert gpu in allowed_gpu_ids + + ray.get([f.remote() for _ in range(100)]) + + def test_blocking_tasks(ray_start_regular): @ray.remote def f(i, j):
`ray.get_gpu_ids()` doesn't work with local_mode <!-- General questions should be asked on the mailing list [email protected]. Before submitting an issue, please fill out the following form. --> ### System information - **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**: Ubuntu 16.04 - **Ray installed from (source or binary)**: binary - **Ray version**: 0.2.1 - **Python version**: 2.7.12 - **Exact command to reproduce**: <!-- You can obtain the Ray version with python -c "import ray; print(ray.__version__)" --> ### Describe the problem `ray.get_gpu_ids()` crashes when `ray.init` was called with `PYTHON_MODE`.
Thanks for reporting this! I don't know if this is related, but I have the same issue. Just testing ray on my local pc with 2 gpus in an interactive python REPL. ``` import ray ray.init(num_gpus=2) ray.get_gpu_ids() ``` This is what I get: ``` Python 3.6.6 |Anaconda, Inc.| (default, Jun 28 2018, 17:14:51) [GCC 7.2.0] on linux Type "help", "copyright", "credits" or "license" for more information. >>> import ray >>> ray.init(num_gpus=2) Process STDOUT and STDERR is being redirected to /tmp/raylogs/. Waiting for redis server at 127.0.0.1:45001 to respond... Waiting for redis server at 127.0.0.1:61294 to respond... Starting local scheduler with the following resources: {'GPU': 2, 'CPU': 16}. Failed to start the UI, you may need to run 'pip install jupyter'. {'node_ip_address': '137.43.130.68', 'redis_address': '137.43.130.68:45001', 'object_store_addresses': [ObjectStoreAddress(name='/tmp/plasma_store10462356', manager_name='/tmp/plasma_manager30881352', manager_port=55957)], 'local_scheduler_socket_names': ['/tmp/scheduler26420348'], 'raylet_socket_names': [], 'webui_url': None} >>> ray.get_gpu_ids() [] >>> ``` Hi, `ray.get_gpu_ids()` doesn't return all of the GPUs on the machine. It is intended to be used within a task and returns the GPU IDs that are reserved for that task. Instead, try ``` import ray ray.init(num_gpus=2) @ray.remote(num_gpus=1) def f(): print(ray.get_gpu_ids()) ``` That said, you probably don't need to actually call this method because Ray will automatically set the `CUDA_VISIBLE_DEVICES` environment variable. Hi @robertnishihara thanks for the clarification, it makes things a bit more clear. If you don't mind me asking, I'm facing another problem. I am trying to run the following example from the tutorials. [ray_example.zip](https://github.com/ray-project/ray/files/2396322/ray_example.zip) The goal is to have that simple example parallelize across both my gpus on my local pc, but I keep getting the following error: ``` WARNING: Serializing objects of type <class 'tensorflow.contrib.learn.python.learn.datasets.base.Datasets'> by expanding them as dictionaries of their fields. This behavior may be incorrect in some cases. WARNING: Serializing objects of type <class 'tensorflow.contrib.learn.python.learn.datasets.mnist.DataSet'> by expanding them as dictionaries of their fields. This behavior may be incorrect in some cases. 2018-09-19 10:05:19.118840: I tensorflow/core/platform/cpu_feature_guard.cc:141] Your CPU supports instructions that this TensorFlow binary was not compiled to use: SSE4.1 SSE4.2 AVX AVX2 AVX512F FMA 2018-09-19 10:05:19.356415: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1405] Found device 0 with properties: name: GeForce GTX 1080 Ti major: 6 minor: 1 memoryClockRate(GHz): 1.7085 pciBusID: 0000:17:00.0 totalMemory: 10.92GiB freeMemory: 10.56GiB 2018-09-19 10:05:19.356444: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1484] Adding visible gpu devices: 0 2018-09-19 10:05:20.263083: I tensorflow/core/common_runtime/gpu/gpu_device.cc:965] Device interconnect StreamExecutor with strength 1 edge matrix: 2018-09-19 10:05:20.263122: I tensorflow/core/common_runtime/gpu/gpu_device.cc:971] 0 2018-09-19 10:05:20.263128: I tensorflow/core/common_runtime/gpu/gpu_device.cc:984] 0: N 2018-09-19 10:05:20.263598: I tensorflow/core/common_runtime/gpu/gpu_device.cc:1097] Created TensorFlow device (/job:localhost/replica:0/task:0/device:GPU:0 with 10215 MB memory) -> physical GPU (device: 0, name: GeForce GTX 1080 Ti, pci bus id: 0000:17:00.0, compute capability: 6.1) Accuracy is 0.8903999924659729. WARNING: Logging before InitGoogleLogging() is written to STDERR I0919 10:05:21.059509 4924 local_scheduler.cc:177] Killed worker pid 5344 which hadn't started yet. ``` Any ideas on why is that happening? It looks like the application is completing successfully since it prints `Accuracy is 0.8903999924659729.`. I suspect the error message is an error that happens accidentally when Ray is shutting down, which we should probably fix. @robertnishihara Oh cool! thank you for your input. Could you please help me clarify if the following Ray notation utilizes both gpus i.e. parallel training? > import ray ray.init(num_gpus=2) @ray.remote(num_gpus=1) def f(): print(ray.get_gpu_ids()) Or should we instead use: > import ray ray.init(num_gpus=2) @ray.remote(num_gpus=2) def f(): print(ray.get_gpu_ids()) for parallel training on both gpus? One last remark is that I tried to modify the `train` method in the above example that I had attached in order to print the accuracy after every step: from this: ``` def train(self, num_steps): for step in range(num_steps): batch_xs, batch_ys = self.mnist.train.next_batch(100) self.sess.run( self.train_step, feed_dict={self.x: batch_xs, self.y_: batch_ys} ) ``` to this one: ``` def train(self, num_steps): for step in range(num_steps): batch_xs, batch_ys = self.mnist.train.next_batch(100) self.sess.run( self.train_step, feed_dict={self.x: batch_xs, self.y_: batch_ys} ) print("Step: {}, Accuracy is {}:".format( step, ray.get(self.get_accuracy.remote()) ) ) ``` But it doesn't print anything, am I using the wrong notation regarding Ray?
2020-06-23T18:53:37
ray-project/ray
9,110
ray-project__ray-9110
[ "8483" ]
7904235517b479f2f4b4216ba901882e35863ccc
diff --git a/rllib/contrib/maddpg/maddpg.py b/rllib/contrib/maddpg/maddpg.py --- a/rllib/contrib/maddpg/maddpg.py +++ b/rllib/contrib/maddpg/maddpg.py @@ -22,6 +22,9 @@ # yapf: disable # __sphinx_doc_begin__ DEFAULT_CONFIG = with_common_config({ + # === Framework to run the algorithm === + "framework": "tf", + # === Settings for each individual policy === # ID of the agent controlled by this policy "agent_id": None, diff --git a/rllib/contrib/maddpg/maddpg_policy.py b/rllib/contrib/maddpg/maddpg_policy.py --- a/rllib/contrib/maddpg/maddpg_policy.py +++ b/rllib/contrib/maddpg/maddpg_policy.py @@ -74,12 +74,14 @@ def _make_continuous_space(space): "Space {} is not supported.".format(space)) obs_space_n = [ - _make_continuous_space(space) for _, (_, space, _, _) in - sorted(config["multiagent"]["policies"].items()) + _make_continuous_space(space) + for _, (_, space, _, + _) in config["multiagent"]["policies"].items() ] act_space_n = [ - _make_continuous_space(space) for _, (_, _, space, _) in - sorted(config["multiagent"]["policies"].items()) + _make_continuous_space(space) + for _, (_, _, space, + _) in config["multiagent"]["policies"].items() ] # _____ Placeholders
[rllib][bug] Some policy-names lead to seemingly unrelated errors ### Some policy-names lead to strange errors The following very simple multi-agent script fails because of the names given to the policies ```python from gym.spaces import Discrete, Box import numpy as np import ray from ray import tune from ray.rllib.env.multi_agent_env import MultiAgentEnv class ErrorGame(MultiAgentEnv): def __init__(self, env_config): pass def reset(self): obs = self._obs() return obs def step(self, action_dict): rewards = {0: 0, 1: 0} obs = self._obs() return obs, rewards, {"__all__": False}, {} def _obs(self): return {0: np.array([0.2, 0.2], dtype=np.float32), 1: np.array([0.2], dtype=np.float32)} if __name__ == "__main__": obs1 = Box(low=np.array([0] * 2), high=np.array([1] * 2)) obs2 = Box(low=np.array([0] * 1), high=np.array([1] * 1)) act1 = Discrete(1) act2 = Discrete(1) # helper_act = Discrete(1) # -- POLICIES-- name1 = "seeker" name2 = "helper" policies = { name1: (None, obs1, act1, {"agent_id": 0}), name2: (None, obs2, act2, {"agent_id": 1}) } def policy_name_mapping(x): return {0:name1, 1:name2}[x] config = { "learning_starts": 100, "evaluation_interval": 10000, "multiagent": { "policies": policies, "policy_mapping_fn": policy_name_mapping, }, } ray.init( num_cpus=2, # object_store_memory=200 * 1024 * 1024, ) tune.run( "contrib/MADDPG", stop={ "timesteps_total": 5000, }, config=dict(config, **{ "env": ErrorGame, }), checkpoint_at_end=True, checkpoint_freq=10000 # restore="checkpoint_50000/checkpoint-50000" ) ``` This will raise the error ` ValueError: Cannot feed value of shape (1, 2) for Tensor 'foo/obs_0:0', which has shape '(?, 1)'` it took me a long time to find that this error will not come up if i change the names of my policies If i replace the lines: ```python name1 = "seeker" name2 = "helper" ``` just with for example ```python name1 = "seeker1" name2 = "seeker2" ``` I can not really find a patter when it works and when not: Here are some working and not working examples: **working**: name1 = "seeker1" | name2 = "seeker2" name1 = "pol" | name2 = "pool" name1 = "ags" | name2 = "asdgasdg" **not-working**: name1 = "seeker" | name2 = "helper" name1 = "rom" | name2 = "rem" name1 = "foo" | name2 = "bar" I can't imagine that's intentional. But if there is some pattern behind it that I don't recognize it would be useful to get a more sophisticated error message Also, as far as I can judge now, the error seems to occur only when the observation_size of the two policies have different sizes. So my guess is that the policies might get mixed up at some point? And also i only got this error when using `"contrib/MADDPG"` but did not test very much with other setups. *Ray version and other system information (Python version, TensorFlow version, OS):* software | version --- | --- ray |0.8.4 python | 3.6.9 tensorflow | 1.14.0 OS | Ubuntu (running in a VM on a Windows OS) Release 18.04
2020-06-23T20:43:33
ray-project/ray
9,133
ray-project__ray-9133
[ "9131" ]
80bcbe20c70dd284e0d93ea94fe97d1b34b81433
diff --git a/python/ray/autoscaler/updater.py b/python/ray/autoscaler/updater.py --- a/python/ray/autoscaler/updater.py +++ b/python/ray/autoscaler/updater.py @@ -169,7 +169,14 @@ def __init__(self, log_prefix, node_id, provider, auth_config, def get_default_ssh_options(self, connect_timeout): OPTS = [ ("ConnectTimeout", "{}s".format(connect_timeout)), + # Supresses initial fingerprint verification. ("StrictHostKeyChecking", "no"), + # SSH IP and fingerprint pairs no longer added to known_hosts. + # This is to remove a "REMOTE HOST IDENTIFICATION HAS CHANGED" + # warning if a new node has the same IP as a previously + # deleted node, because the fingerprints will not match in + # that case. + ("UserKnownHostsFile", os.devnull), ("ControlMaster", "auto"), ("ControlPath", "{}/%C".format(self.ssh_control_path)), ("ControlPersist", "10s"),
Ray should remove ssh fingerprints from known_hosts when tearing down a cluster <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### Describe your feature request When Ray kills a node or tears down a cluster, it should remove the ssh fingerprints for the corresponding IP addresses from ~/.ssh/known_hosts. GCP often reuses IP addresses, so if you delete and reprovision a node, it is common for GCP to give you the same IP. Because it's a new VM, the fingerprint is going to be different than the last node's fingerprint. This triggers `WARNING: REMOTE HOST IDENTIFICATION HAS CHANGED!`, which ruins the current action. This issue probably hasn't been noticed until now because AWS rarely gives you the same IP. Currently I've been fixing this by manually deleting entries from known_hosts, but it would be nice if this happened automatically (e.g. by executing `ssh-keygen -R $DEST`). This should also happen if Ray notices that the cluster has been deleted through the GCP/AWS console.
cc @thomasdesr This seems to be that scary error that I've seen sometimes?
2020-06-24T23:36:39
ray-project/ray
9,141
ray-project__ray-9141
[ "9128" ]
536795ef7983b2224742c768e3e62284b5daddb6
diff --git a/python/ray/tune/trial_runner.py b/python/ray/tune/trial_runner.py --- a/python/ray/tune/trial_runner.py +++ b/python/ray/tune/trial_runner.py @@ -276,7 +276,7 @@ def checkpoint(self, force=False): with open(tmp_file_name, "w") as f: json.dump(runner_state, f, indent=2, cls=_TuneFunctionEncoder) - os.rename(tmp_file_name, self.checkpoint_file) + os.replace(tmp_file_name, self.checkpoint_file) if force: self._syncer.sync_up() else:
[tune][rllib] Windows: FileExistsError when running rllib tune job <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? My script (available below) runs fine on Linux but does not fully run yet on Windows. I can't tell exactly what went wrong, but the two major errors I see is: `TypeError: len() of unsized object` `FileExistsError: [WinError 183] Cannot create a file when that file already exists` Full error log: https://gist.github.com/juliusfrost/acbca090259610a176847e7026dd6d30 ### Reproduction Run on Windows OS Install pytorch Install the latest `ray` and `rllib` versions Install `atari-py` if necessary Download `train_a2c.py` from https://gist.github.com/juliusfrost/61b8be67d33b9bc9ab1faf7ada9d2ae3 Run `python train_a2c.py BreakoutNoFrameskip-v4` if you don't have a gpu add `--gpus 0`
#9114 Seems related cc @mehrdadn Thanks for the report! I added this to the list.
2020-06-25T12:30:08
ray-project/ray
9,142
ray-project__ray-9142
[ "8507" ]
536795ef7983b2224742c768e3e62284b5daddb6
diff --git a/rllib/examples/models/custom_loss_model.py b/rllib/examples/models/custom_loss_model.py --- a/rllib/examples/models/custom_loss_model.py +++ b/rllib/examples/models/custom_loss_model.py @@ -121,6 +121,8 @@ def __init__(self, obs_space, action_space, num_outputs, model_config, nn.Module.__init__(self) self.input_files = input_files + # Create a new input reader per worker. + self.reader = JsonReader(self.input_files) self.fcnet = TorchFC( self.obs_space, self.action_space, @@ -135,13 +137,30 @@ def forward(self, input_dict, state, seq_lens): @override(ModelV2) def custom_loss(self, policy_loss, loss_inputs): - # Create a new input reader per worker. - reader = JsonReader(self.input_files) - input_ops = reader.tf_input_ops() + """Calculates a custom loss on top of the given policy_loss(es). + + Args: + policy_loss (List[TensorType]): The list of already calculated + policy losses (as many as there are optimizers). + loss_inputs (TensorStruct): Struct of np.ndarrays holding the + entire train batch. + + Returns: + List[TensorType]: The altered list of policy losses. In case the + custom loss should have its own optimizer, make sure the + returned list is one larger than the incoming policy_loss list. + In case you simply want to mix in the custom loss into the + already calculated policy losses, return a list of altered + policy losses (as done in this example below). + """ + # Get the next batch from our input files. + batch = self.reader.next() # Define a secondary loss by building a graph copy with weight sharing. obs = restore_original_dimensions( - tf.cast(input_ops["obs"], tf.float32), self.obs_space) + torch.from_numpy(batch["obs"]).float(), + self.obs_space, + tensorlib="torch") logits, _ = self.forward({"obs": obs}, [], None) # You can also add self-supervised losses easily by referencing tensors @@ -155,11 +174,15 @@ def custom_loss(self, policy_loss, loss_inputs): action_dist = TorchCategorical(logits, self.model_config) self.policy_loss = policy_loss self.imitation_loss = torch.mean( - -action_dist.logp(input_ops["actions"])) - return policy_loss + 10 * self.imitation_loss + -action_dist.logp(torch.from_numpy(batch["actions"]))) + + # Add the imitation loss to each already calculated policy loss term. + # Alternatively (if custom loss has its own optimizer): + # return policy_loss + [10 * self.imitation_loss] + return [l + 10 * self.imitation_loss for l in policy_loss] def custom_stats(self): return { - "policy_loss": self.policy_loss, + "policy_loss": torch.mean(self.policy_loss), "imitation_loss": self.imitation_loss, } diff --git a/rllib/offline/json_reader.py b/rllib/offline/json_reader.py --- a/rllib/offline/json_reader.py +++ b/rllib/offline/json_reader.py @@ -45,7 +45,7 @@ def __init__(self, inputs, ioctx=None): logger.warning( "Treating input directory as glob pattern: {}".format( inputs)) - if urlparse(inputs).scheme: + if urlparse(inputs).scheme not in ["d", ""]: raise ValueError( "Don't know how to glob over `{}`, ".format(inputs) + "please specify a list of files to read instead.") diff --git a/rllib/policy/torch_policy.py b/rllib/policy/torch_policy.py --- a/rllib/policy/torch_policy.py +++ b/rllib/policy/torch_policy.py @@ -240,6 +240,9 @@ def learn_on_batch(self, postprocessed_batch): train_batch = self._lazy_tensor_dict(postprocessed_batch) loss_out = force_list( self._loss(self, self.model, self.dist_class, train_batch)) + # Call Model's custom-loss with Policy loss outputs and train_batch. + if self.model: + loss_out = self.model.custom_loss(loss_out, train_batch) assert len(loss_out) == len(self._optimizers) # assert not any(torch.isnan(l) for l in loss_out)
[rllib] Pytorch missing custom loss <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? Running `rllib/examples/custom_loss.py` example, the tensorflow implementation displays the custom loss, but Pytorch does not. During the process, I had to update the `rllib/examples/models/custom_loss_model.py` models method name from `custom_stats `to `metrics`. *Ray version and other system information (Python version, TensorFlow version, OS):* Ray version: latest wheel Python: 3.6.8 TF: 2.1 OS: RHEL 7.7 ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): Use existing example script in Rllib If we cannot run your script, we cannot fix your issue. - [x ] I have verified my script runs in a clean environment and reproduces the issue. - [x ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
#8193 Sorry, this may have slipped through the cracks. I'll take a look and provide the missing example. Got it, `ModelV2.custom_loss()` is not supported for PyTorch yet at all (ignored). Will fix this. Yes, `metrics` is the new name (ModelV2) for the old `custom_stats` (Model(V1)). I'm looking closer at the current tf policies, and I'm actually struggling to see where `model.custom_loss` has been implemented in tf IMPALA either... Can you highlight the part of the code where this is done for my benefit? In the case of IMPALA, I know it calls the `model.custom_loss` method, as I can see the custom loss metrics I've produced, but what is not clear is that it is actually making it into the gradient calculations... `build_vtrace_loss` seems to just call the class `VTraceLoss`, where I don't see any references to the output of` model.custom_loss`. and dynamic_tf_policy seems to just return the output of this loss function... so I'm struggling to see where it comes into play. Looks like it is only used once [here](https://github.com/ray-project/ray/blob/master/rllib/policy/tf_policy.py#L215) and from there on it's accessed through `self._loss`. Thanks for that, not sure how that eluded me… it looks like on further review my issue on the TF side was using 1.x style tf.losses.get_regularization_loss rather than model.losses … once I switched everything is working for TF models. On May 28, 2020, at 4:50 PM, Jan Blumenkamp <[email protected]> wrote: > > Looks like it is only used once here and from there on it's accessed through self._loss. > > — > You are receiving this because you authored the thread. > Reply to this email directly, view it on GitHub, or unsubscribe. > > Got it, `ModelV2.custom_loss()` is not supported for PyTorch yet at all (ignored). Will fix this. > > Yes, `metrics` is the new name (ModelV2) for the old `custom_stats` (Model(V1)). Hi! Is there any update on this? If update is not expected soon, I might give it a shot with tensorflow side. Many thanks. Hey, @cedros23 . Taking another look right now.
2020-06-25T13:03:46
ray-project/ray
9,227
ray-project__ray-9227
[ "8784" ]
f1173d55e051467aa0ec4fccc2f7ff3bbe06add0
diff --git a/python/ray/serve/metric/client.py b/python/ray/serve/metric/client.py --- a/python/ray/serve/metric/client.py +++ b/python/ray/serve/metric/client.py @@ -1,11 +1,8 @@ import asyncio -from typing import Dict, Optional, Tuple +from typing import Dict, Optional, Tuple, List -from ray.serve.metric.types import ( - MetricType, - convert_event_type_to_class, - MetricMetadata, -) +from ray.serve.metric.types import (MetricType, convert_event_type_to_class, + MetricMetadata, MetricRecord) from ray.serve.utils import _get_logger from ray.serve.constants import METRIC_PUSH_INTERVAL_S @@ -29,8 +26,8 @@ def __init__( self.exporter = metric_exporter_actor self.default_labels = default_labels or dict() - self.registered_metrics: Dict[str, MetricMetadata] = dict() - self.metric_records = [] + self.registered_metrics: Dict[int, MetricMetadata] = dict() + self.metric_records: List[MetricRecord] = [] assert asyncio.get_event_loop().is_running() self.push_task = asyncio.get_event_loop().create_task( @@ -101,10 +98,6 @@ def _new_metric( description: str, label_names: Tuple[str] = (), ): - if name in self.registered_metrics: - raise ValueError( - "Metric with name {} is already registered.".format(name)) - if not isinstance(label_names, tuple): raise ValueError("label_names need to be a tuple, it is {}".format( type(label_names))) @@ -116,11 +109,17 @@ def _new_metric( label_names=label_names, default_labels=self.default_labels.copy(), ) + + key = hash(metric_metadata) + if key in self.registered_metrics: + raise ValueError("Metric named {} and associated metadata " + "is already registered.".format(name)) + self.registered_metrics[key] = metric_metadata + metric_class = convert_event_type_to_class(metric_type) metric_object = metric_class( - client=self, name=name, label_names=label_names) + client=self, key=key, label_names=label_names) - self.registered_metrics[name] = metric_metadata return metric_object async def _push_to_exporter_once(self): diff --git a/python/ray/serve/metric/exporter.py b/python/ray/serve/metric/exporter.py --- a/python/ray/serve/metric/exporter.py +++ b/python/ray/serve/metric/exporter.py @@ -1,4 +1,4 @@ -from typing import Dict +from typing import Dict, Any from collections import Counter, namedtuple import ray @@ -46,15 +46,15 @@ def __init__(self, exporter_class: ExporterInterface): # TODO(simon): Add support for initializer args and kwargs. self.exporter = exporter_class() - # Stores the mapping metric_name -> MetricMetadata + # Stores the all recorded MetricMetadata (keyed by hash key) # This field is tolerant to failures since each client will always push # an updated copy of the metadata for each ingest call. - self.metric_metadata = dict() + self.metric_metadata: Dict[int, MetricMetadata] = dict() logger.debug("Initialized with metric exporter of type {}".format( type(self.exporter))) - def ingest(self, metric_metadata: Dict[str, MetricMetadata], + def ingest(self, metric_metadata: Dict[int, MetricMetadata], batch: MetricBatch): self.metric_metadata.update(metric_metadata) self.exporter.export(self.metric_metadata, batch) @@ -70,9 +70,10 @@ def __init__(self): # Keep track of latest observation of measures self.latest_measures: Dict[namedtuple, float] = dict() - def export(self, metric_metadata, metric_batch): + def export(self, metric_metadata: Dict[int, MetricMetadata], + metric_batch: MetricBatch): for record in metric_batch: - metadata = metric_metadata[record.name] + metadata = metric_metadata[record.key] metric_key = make_metric_namedtuple(metadata, record) if metadata.type == MetricType.COUNTER: self.counters[metric_key] += record.value @@ -88,7 +89,8 @@ def inspect_metrics(self): for info_tuple, value in metrics_to_collect.items(): # Represent the metric type as a human readable name info_tuple = info_tuple._replace(type=str(info_tuple.type)) - items.append({"info": info_tuple._asdict(), "value": value}) + # _asdict returns OrderedDict, we just need to return regular dict + items.append({"info": dict(info_tuple._asdict()), "value": value}) return items @@ -103,7 +105,7 @@ def __init__(self): } self.prom_generate_latest = generate_latest - self.metrics_cache = dict() + self.metrics_cache: Dict[str, Any] = dict() self.default_labels = dict() self.registry = CollectorRegistry() @@ -112,10 +114,12 @@ def inspect_metrics(self): def export(self, metric_metadata, metric_batch): self._process_metric_metadata(metric_metadata) - self._process_batch(metric_batch) + self._process_batch(metric_metadata, metric_batch) def _process_metric_metadata(self, metric_metadata): - for name, metric_metadata in metric_metadata.items(): + for key, metric_metadata in metric_metadata.items(): + name = metric_metadata.name + if name not in self.metrics_cache: constructor = self.metric_type_to_prom_type[ metric_metadata.type] @@ -124,7 +128,7 @@ def _process_metric_metadata(self, metric_metadata): label_names = tuple( default_labels.keys()) + metric_metadata.label_names metric_object = constructor( - metric_metadata.name, + name, metric_metadata.description, labelnames=label_names, registry=self.registry, @@ -132,14 +136,17 @@ def _process_metric_metadata(self, metric_metadata): self.metrics_cache[name] = (metric_object, metric_metadata.type) - self.default_labels[name] = default_labels - def _process_batch(self, batch): - for name, labels, value in batch: - assert name in self.metrics_cache, ( - "Metrics {} was not registered.".format(name)) + self.default_labels[key] = metric_metadata.default_labels + + def _process_batch(self, metric_metadata, batch): + for key, labels, value in batch: + if key not in metric_metadata: + continue + + name = metric_metadata[key].name metric, metric_type = self.metrics_cache[name] - default_labels = self.default_labels[name] + default_labels = self.default_labels[key] merged_labels = {**default_labels, **labels} if metric_type == MetricType.COUNTER: metric.labels(**merged_labels).inc(value) diff --git a/python/ray/serve/metric/types.py b/python/ray/serve/metric/types.py --- a/python/ray/serve/metric/types.py +++ b/python/ray/serve/metric/types.py @@ -2,27 +2,53 @@ from typing import Tuple, List, Dict, Optional from collections import namedtuple + +class MetricType(enum.IntEnum): + COUNTER = 1 + MEASURE = 2 + + # We split the information about a metric into two parts: the MetricMetadata # and MetricRecord. Metadata is declared at creation time and include names # and the label names. The label values will be supplied at observation time. -MetricMetadata = namedtuple( - "MetricMetadata", - ["name", "type", "description", "label_names", "default_labels"]) -MetricRecord = namedtuple("MetricRecord", ["name", "labels", "value"]) +class MetricMetadata: + def __init__(self, name: str, type: MetricType, description: str, + label_names: Tuple[str], default_labels: Dict[str, str]): + self.name = name + self.type = type + self.description = description + self.label_names = label_names + self.default_labels = default_labels + + def __eq__(self, value: "MetricMetadata"): + if not isinstance(value, MetricMetadata): + return False + + return (self.name == value.name and self.type == self.type + and self.description == value.description + and self.label_names == value.label_names + and self.default_labels == value.default_labels) + + def __hash__(self): + return hash((self.name, self.type, self.description, self.label_names, + frozenset(self.default_labels.items()))) + + +MetricRecord = namedtuple("MetricRecord", ["key", "labels", "value"]) MetricBatch = List[MetricRecord] class BaseMetric: def __init__(self, client, - name: str, + key: int, label_names: Tuple[str], dynamic_labels: Optional[Dict[str, str]] = None): """Represent a single metric stream Args: client(MetricClient): The client object to push update to. - name(str): The name of the metric. + key(int): The unique hash key of the metric. label_names(Tuple[str]): The names of the labels that must be set before an observation. dynamic_labels(Optional[Dict[str,str]]): A partially preset labels. @@ -30,7 +56,7 @@ def __init__(self, ``metric.labels(a=b).labels(c=d)``. """ self.client = client - self.name = name + self.key = key self.dynamic_labels = dynamic_labels or dict() self.label_names = label_names @@ -56,7 +82,7 @@ def labels(self, **kwargs): "label names. Allowed label names are {}.".format( k, self.label_names)) new_dynamic_labels[k] = str(v) - return type(self)(self.client, self.name, self.label_names, + return type(self)(self.client, self.key, self.label_names, new_dynamic_labels) @@ -67,7 +93,7 @@ def add(self, increment=1): """Increment the counter by some amount. Default is 1""" self.check_all_labels_fulfilled_or_error() self.client.metric_records.append( - MetricRecord(self.name, self.dynamic_labels, increment)) + MetricRecord(self.key, self.dynamic_labels, increment)) class Measure(BaseMetric): @@ -75,12 +101,7 @@ def record(self, value): """Record the given value for the measure""" self.check_all_labels_fulfilled_or_error() self.client.metric_records.append( - MetricRecord(self.name, self.dynamic_labels, value)) - - -class MetricType(enum.IntEnum): - COUNTER = 1 - MEASURE = 2 + MetricRecord(self.key, self.dynamic_labels, value)) def convert_event_type_to_class(event_type: MetricType) -> BaseMetric:
diff --git a/python/ray/serve/tests/test_metric.py b/python/ray/serve/tests/test_metric.py --- a/python/ray/serve/tests/test_metric.py +++ b/python/ray/serve/tests/test_metric.py @@ -46,29 +46,29 @@ async def test_client(): await collector._push_to_exporter_once() - assert exporter.metadata == { - "counter": MetricMetadata( - name="counter", - type=MetricType.COUNTER, - description="", - label_names=("a", "b"), - default_labels={"default": "label"}, - ), - "measure": MetricMetadata( - name="measure", - type=MetricType.MEASURE, - description="", - label_names=(), - default_labels={"default": "label"}, - ) - } - assert exporter.batches == [("counter", { - "a": "1", - "b": "2" - }, 1), ("counter", { - "a": "1", - "b": "3" - }, 42), ("measure", {}, 2)] + assert MetricMetadata( + name="counter", + type=MetricType.COUNTER, + description="", + label_names=("a", "b"), + default_labels={"default": "label"}, + ) in exporter.metadata.values() + assert MetricMetadata( + name="measure", + type=MetricType.MEASURE, + description="", + label_names=(), + default_labels={"default": "label"}, + ) in exporter.metadata.values() + + metric_values = [item.value for item in exporter.batches] + assert set(metric_values) == {1, 42, 2} + + metric_labels = [ + frozenset(item.labels.items()) for item in exporter.batches + ] + assert frozenset(dict(a="1", b="2").items()) in metric_labels + assert frozenset(dict(a="1", b="3").items()) in metric_labels async def test_in_memory_exporter(serve_instance): @@ -140,6 +140,26 @@ async def test_prometheus_exporter(serve_instance): assert fragment in metric_stored +async def test_prometheus_conflicting_labels(serve_instance): + exporter = MetricExporterActor.remote(PrometheusExporter) + + collector_a = MetricClient( + exporter, push_interval=2, default_labels={"default": "a"}) + collector_b = MetricClient( + exporter, push_interval=2, default_labels={"default": "b"}) + + for collector in [collector_a, collector_b]: + counter = collector.new_counter("num") + counter.add() + await collector._push_to_exporter_once() + + metric_stored = (await exporter.inspect_metrics.remote()).decode() + + fragments = ['num_total{default="a"}', 'num_total{default="b"}'] + for fragment in fragments: + assert fragment in metric_stored + + async def test_system_metric_endpoints(serve_instance): def test_error_counter(flask_request): 1 / 0
[Serve] Prometheus Metric Overrides <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? Serve's prometheus exporter is stores a metric name -> default label mapping, this is a bug because there are can different metrics for the same name without different default labels. *Ray version and other system information (Python version, TensorFlow version, OS):* ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): Code ```python import time import pprint import ray from ray import serve from ray.serve.constants import SERVE_MASTER_NAME from ray.serve.metric import MetricClient, PrometheusExporter serve.init(metric_exporter=PrometheusExporter) @ray.remote class AsyncActor: async def run(self, key): actor = ray.get_actor(SERVE_MASTER_NAME) [metric_exporter] = ray.get(actor.get_metric_exporter.remote()) client = MetricClient(metric_exporter, default_labels={"key": key}) counter = client.new_counter("same_name") for _ in range(10): counter.add() await client._push_to_exporter_once() a = AsyncActor.remote() b = AsyncActor.remote() ray.get([a.run.remote("a"), b.run.remote("b")]) print(serve.stat().decode()) ``` Output ``` 2020-06-04 12:07:19,281 INFO resource_spec.py:212 -- Starting Ray with 21.97 GiB memory available for workers and up to 10.99 GiB for objects. You can adjust these settings with ray.init(memory=<bytes>, object_store_memory=<bytes>). 2020-06-04 12:07:19,500 WARNING services.py:923 -- Redis failed to start, retrying now. 2020-06-04 12:07:19,740 INFO services.py:1165 -- View the Ray dashboard at localhost:8265 (pid=61255) 2020-06-04 12:07:21,622 INFO master.py:182 -- Starting metric exporter with name 'SERVE_METRIC_SINK_ACTOR' (pid=61255) 2020-06-04 12:07:21,636 INFO master.py:130 -- Starting router with name 'SERVE_ROUTER_ACTOR' (pid=61255) 2020-06-04 12:07:21,644 INFO master.py:152 -- Starting HTTP proxy with name 'SERVE_PROXY_ACTOR' on node 'node:192.168.31.141' (pid=61267) INFO: Started server process [61267] (pid=61267) INFO: Waiting for application startup. (pid=61267) INFO: Application startup complete. # HELP same_name_total # TYPE same_name_total counter same_name_total{key="a"} 20.0 # TYPE same_name_created gauge same_name_created{key="a"} 1.591297642889063e+09 ``` Expected: ``` same_name_total{key="a"} 10.0 ... same_name_total{key="b"} 10.0 ``` If we cannot run your script, we cannot fix your issue. - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-06-30T23:35:07
ray-project/ray
9,228
ray-project__ray-9228
[ "9251" ]
e0d8b58969728054d1800abc925aa5352882baa8
diff --git a/python/ray/async_compat.py b/python/ray/async_compat.py --- a/python/ray/async_compat.py +++ b/python/ray/async_compat.py @@ -3,8 +3,6 @@ It will raise SyntaxError when importing from Python 2. """ import asyncio -from collections import namedtuple -import time import inspect try: @@ -35,84 +33,13 @@ async def wrapper(*args, **kwargs): return wrapper -# Class encapsulate the get result from direct actor. -# Case 1: plasma_fallback_id=None, result=<Object> -# Case 2: plasma_fallback_id=ObjectID, result=None -AsyncGetResponse = namedtuple("AsyncGetResponse", - ["plasma_fallback_id", "result"]) - - def get_async(object_id): - """Asyncio compatible version of ray.get""" - # Delayed import because raylet import this file and - # it creates circular imports. - from ray.experimental.async_api import init as async_api_init, as_future - from ray.experimental.async_plasma import PlasmaObjectFuture - - assert isinstance(object_id, ray.ObjectID), "Batched get is not supported." - - # Setup - async_api_init() + """C++ Asyncio version of ray.get""" loop = asyncio.get_event_loop() core_worker = ray.worker.global_worker.core_worker - # Here's the callback used to implement async get logic. - # What we want: - # - If direct call, first try to get it from in memory store. - # If the object if promoted to plasma, retry it from plasma API. - # - If not direct call, directly use plasma API to get it. - user_future = loop.create_future() - - # We have three future objects here. - # user_future is directly returned to the user from this function. - # and it will be eventually fulfilled by the final result. - # inner_future is the first attempt to retrieve the object. It can be - # fulfilled by either core_worker.get_async or plasma_api.as_future. - # When inner_future completes, done_callback will be invoked. This - # callback set the final object in user_future if the object hasn't - # been promoted by plasma, otherwise it will retry from plasma. - # retry_plasma_future is only created when we are getting objects that's - # promoted to plasma. It will also invoke the done_callback when it's - # fulfilled. - - def done_callback(future): - result = future.result() - # Result from async plasma, transparently pass it to user future - if isinstance(future, PlasmaObjectFuture): - if isinstance(result, ray.exceptions.RayTaskError): - ray.worker.last_task_error_raise_time = time.time() - user_future.set_exception(result.as_instanceof_cause()) - else: - user_future.set_result(result) - else: - # Result from direct call. - assert isinstance(result, AsyncGetResponse), result - if result.plasma_fallback_id is None: - # If this future has result set already, we just need to - # skip the set result/exception procedure. - if user_future.done(): - return - - if isinstance(result.result, ray.exceptions.RayTaskError): - ray.worker.last_task_error_raise_time = time.time() - user_future.set_exception( - result.result.as_instanceof_cause()) - else: - user_future.set_result(result.result) - else: - # Schedule plasma to async get, use the the same callback. - retry_plasma_future = as_future(result.plasma_fallback_id) - retry_plasma_future.add_done_callback(done_callback) - # A hack to keep reference to the future so it doesn't get GC. - user_future.retry_plasma_future = retry_plasma_future - - inner_future = loop.create_future() - # We must add the done_callback before sending to in_memory_store_get - inner_future.add_done_callback(done_callback) - core_worker.in_memory_store_get_async(object_id, inner_future) - # A hack to keep reference to inner_future so it doesn't get GC. - user_future.inner_future = inner_future + future = loop.create_future() + core_worker.get_async(object_id, future) # A hack to keep a reference to the object ID for ref counting. - user_future.object_id = object_id - - return user_future + future.object_id = object_id + return future diff --git a/python/ray/experimental/async_api.py b/python/ray/experimental/async_api.py deleted file mode 100644 --- a/python/ray/experimental/async_api.py +++ /dev/null @@ -1,45 +0,0 @@ -import asyncio - -import ray -from ray.experimental.async_plasma import PlasmaEventHandler -from ray.services import logger - -handler = None - - -def init(): - """Initialize plasma event handlers for asyncio support.""" - assert ray.is_initialized(), "Please call ray.init before async_api.init" - - global handler - if handler is None: - worker = ray.worker.global_worker - loop = asyncio.get_event_loop() - handler = PlasmaEventHandler(loop, worker) - worker.core_worker.set_plasma_added_callback(handler) - logger.debug("AsyncPlasma Connection Created!") - - -def as_future(object_id): - """Turn an object_id into a Future object. - - Args: - object_id: A Ray object_id. - - Returns: - PlasmaObjectFuture: A future object that waits the object_id. - """ - if handler is None: - init() - return handler.as_future(object_id) - - -def shutdown(): - """Manually shutdown the async API. - - Cancels all related tasks and all the socket transportation. - """ - global handler - if handler is not None: - handler.close() - handler = None diff --git a/python/ray/experimental/async_plasma.py b/python/ray/experimental/async_plasma.py deleted file mode 100644 --- a/python/ray/experimental/async_plasma.py +++ /dev/null @@ -1,70 +0,0 @@ -import asyncio - -import ray -from ray.services import logger -from collections import defaultdict - - -class PlasmaObjectFuture(asyncio.Future): - """This class is a wrapper for a Future on Plasma.""" - pass - - -class PlasmaEventHandler: - """This class is an event handler for Plasma.""" - - def __init__(self, loop, worker): - super().__init__() - self._loop = loop - self._worker = worker - self._waiting_dict = defaultdict(list) - - def _complete_future(self, ray_object_id): - # TODO(ilr): Consider race condition between popping from the - # waiting_dict and as_future appending to the waiting_dict's list. - logger.debug( - "Completing plasma futures for object id {}".format(ray_object_id)) - if ray_object_id not in self._waiting_dict: - return - obj = self._worker.get_objects([ray_object_id], timeout=0)[0] - futures = self._waiting_dict.pop(ray_object_id) - for fut in futures: - try: - fut.set_result(obj) - except asyncio.InvalidStateError: - # Avoid issues where process_notifications - # and check_immediately both get executed - logger.debug("Failed to set result for future {}." - "Most likely already set.".format(fut)) - - def close(self): - """Clean up this handler.""" - for futures in self._waiting_dict.values(): - for fut in futures: - fut.cancel() - - def check_immediately(self, object_id): - ready, _ = ray.wait([object_id], timeout=0) - if ready: - self._complete_future(object_id) - - def as_future(self, object_id, check_ready=True): - """Turn an object_id into a Future object. - - Args: - object_id: A Ray's object_id. - check_ready (bool): If true, check if the object_id is ready. - - Returns: - PlasmaObjectFuture: A future object that waits the object_id. - """ - if not isinstance(object_id, ray.ObjectID): - raise TypeError("Input should be a Ray ObjectID.") - - future = PlasmaObjectFuture(loop=self._loop) - self._waiting_dict[object_id].append(future) - if not self.check_immediately(object_id) and len( - self._waiting_dict[object_id]) == 1: - # Only subscribe once - self._worker.core_worker.subscribe_to_plasma_object(object_id) - return future
diff --git a/ci/asan_tests/run_asan_tests.sh b/ci/asan_tests/run_asan_tests.sh --- a/ci/asan_tests/run_asan_tests.sh +++ b/ci/asan_tests/run_asan_tests.sh @@ -35,7 +35,7 @@ asan_run() { cd "${ROOT_DIR}"/../.. # async plasma test - python -m pytest -v --durations=5 --timeout=300 python/ray/experimental/test/async_test.py + python -m pytest -v --durations=5 --timeout=300 python/ray/tests/test_async.py # Ray tests bazel test --test_tag_filters=-jenkins_only python/ray/serve/... diff --git a/python/ray/tests/BUILD b/python/ray/tests/BUILD --- a/python/ray/tests/BUILD +++ b/python/ray/tests/BUILD @@ -472,3 +472,11 @@ py_test( tags = ["exclusive"], deps = ["//:ray_lib"], ) + +py_test( + name = "test_async", + size = "medium", + srcs = SRCS + ["test_async.py"], + tags = ["exclusive"], + deps = ["//:ray_lib"], +) diff --git a/python/ray/experimental/test/async_test.py b/python/ray/tests/test_async.py similarity index 67% rename from python/ray/experimental/test/async_test.py rename to python/ray/tests/test_async.py --- a/python/ray/experimental/test/async_test.py +++ b/python/ray/tests/test_async.py @@ -1,20 +1,19 @@ import asyncio +import sys import time -import pytest import numpy as np +import pytest + import ray -from ray.experimental import async_api @pytest.fixture def init(): ray.init(num_cpus=4) - async_api.init() asyncio.get_event_loop().set_debug(False) yield - async_api.shutdown() ray.shutdown() @@ -33,7 +32,7 @@ def f(): time.sleep(1) return np.zeros(1024 * 1024, dtype=np.uint8) - future = async_api.as_future(f.remote()) + future = f.remote().as_future() result = asyncio.get_event_loop().run_until_complete(future) assert isinstance(result, np.ndarray) @@ -41,7 +40,7 @@ def f(): def test_gather(init): loop = asyncio.get_event_loop() tasks = gen_tasks() - futures = [async_api.as_future(obj_id) for obj_id in tasks] + futures = [obj_id.as_future() for obj_id in tasks] results = loop.run_until_complete(asyncio.gather(*futures)) assert all(a[0] == b[0] for a, b in zip(results, ray.get(tasks))) @@ -49,7 +48,7 @@ def test_gather(init): def test_wait(init): loop = asyncio.get_event_loop() tasks = gen_tasks() - futures = [async_api.as_future(obj_id) for obj_id in tasks] + futures = [obj_id.as_future() for obj_id in tasks] results, _ = loop.run_until_complete(asyncio.wait(futures)) assert set(results) == set(futures) @@ -57,7 +56,7 @@ def test_wait(init): def test_wait_timeout(init): loop = asyncio.get_event_loop() tasks = gen_tasks(10) - futures = [async_api.as_future(obj_id) for obj_id in tasks] + futures = [obj_id.as_future() for obj_id in tasks] fut = asyncio.wait(futures, timeout=5) results, _ = loop.run_until_complete(fut) assert list(results)[0] == futures[0] @@ -75,12 +74,7 @@ async def g(n): await asyncio.sleep(n * 0.1) return n, np.zeros(1024 * 1024, dtype=np.uint8) - tasks = [ - async_api.as_future(f.remote(1)), - g(2), - async_api.as_future(f.remote(3)), - g(4) - ] + tasks = [f.remote(1).as_future(), g(2), f.remote(3).as_future(), g(4)] results = loop.run_until_complete(asyncio.gather(*tasks)) assert [result[0] for result in results] == [1, 2, 3, 4] @@ -100,11 +94,31 @@ async def _g(_n): return asyncio.ensure_future(_g(n)) - tasks = [ - async_api.as_future(f.remote(0.1)), - g(7), - async_api.as_future(f.remote(5)), - g(2) - ] + tasks = [f.remote(0.1).as_future(), g(7), f.remote(5).as_future(), g(2)] ready, _ = loop.run_until_complete(asyncio.wait(tasks, timeout=4)) assert set(ready) == {tasks[0], tasks[-1]} + + [email protected] [email protected]( + "ray_start_regular_shared", [{ + "object_store_memory": 100 * 1024 * 1024, + }], + indirect=True) +async def test_garbage_collection(ray_start_regular_shared): + # This is a regression test for + # https://github.com/ray-project/ray/issues/9134 + + @ray.remote + def f(): + return np.zeros(40 * 1024 * 1024, dtype=np.uint8) + + for _ in range(10): + await f.remote() + for _ in range(10): + put_id = ray.put(np.zeros(40 * 1024 * 1024, dtype=np.uint8)) + await put_id + + +if __name__ == "__main__": + sys.exit(pytest.main(["-v", __file__])) diff --git a/python/ray/tests/test_asyncio.py b/python/ray/tests/test_asyncio.py --- a/python/ray/tests/test_asyncio.py +++ b/python/ray/tests/test_asyncio.py @@ -1,8 +1,9 @@ # coding: utf-8 import asyncio +import sys import threading + import pytest -import sys import ray from ray.test_utils import SignalActor @@ -113,10 +114,6 @@ async def test_asyncio_get(ray_start_regular_shared, event_loop): asyncio.set_event_loop(loop) loop.set_debug(True) - # This is needed for async plasma - from ray.experimental.async_api import init - init() - # Test Async Plasma @ray.remote def task():
Aysyncio ObjectID becomes None <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? Calling `await` on a task's ObjectID inside an async Actor sometimes results in the following error `AttributeError: 'NoneType' object has no attribute 'call_soon_threadsafe'`. This happens when the task runs after the actor method returns. This does not happen if `asyncio.wait()` is called. Sometimes this issue results in a weird segfault. *Ray version and other system information (Python version, TensorFlow version, OS):* ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): - Change this [line](https://github.com/ray-project/ray/blob/master/python/ray/serve/router.py#L324) to `await worker.handle_request.remote(req)` - Run `pytest -vs serve/tests/test_api.py::test_shadow_traffic` If we cannot run your script, we cannot fix your issue. - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-07-01T00:55:54
ray-project/ray
9,297
ray-project__ray-9297
[ "9268" ]
7a2d7964d8f944bd60c1f03d58d6cc190c7a7015
diff --git a/python/ray/tune/logger.py b/python/ray/tune/logger.py --- a/python/ray/tune/logger.py +++ b/python/ray/tune/logger.py @@ -187,7 +187,7 @@ class TBXLogger(Logger): """ # NoneType is not supported on the last TBX release yet. - VALID_HPARAMS = (str, bool, int, float, list) + VALID_HPARAMS = (str, bool, np.bool8, int, np.integer, float, list) def _init(self): try:
diff --git a/python/ray/tune/tests/test_logger.py b/python/ray/tune/tests/test_logger.py --- a/python/ray/tune/tests/test_logger.py +++ b/python/ray/tune/tests/test_logger.py @@ -2,6 +2,7 @@ import unittest import tempfile import shutil +import numpy as np from ray.tune.logger import JsonLogger, CSVLogger, TBXLogger @@ -46,7 +47,17 @@ def testJSON(self): logger.close() def testTBX(self): - config = {"a": 2, "b": [1, 2], "c": {"c": {"D": 123}}} + config = { + "a": 2, + "b": [1, 2], + "c": { + "c": { + "D": 123 + } + }, + "d": np.int64(1), + "e": np.bool8(True) + } t = Trial(evaluated_params=config, trial_id="tbx") logger = TBXLogger(config=config, logdir=self.test_dir, trial=t) logger.on_result(result(0, 4))
[tune] Parameters from `tune.choice()` do not get logged to TensorBoard when integers ### What is the problem? When providing parameters via `tune.choice()` that include integers, the values are not logged to TensorBoard's HPARAMS section. The issue is that `numpy.random.choice([1, 2, 3])` (for example) returns `numpy.int32`/`numpy.int64` and those types are not included in the `VALID_HPARAMS = (str, bool, int, float, list)` tuple (python/ray/tune/logger.py). Since TensorBoard has no issues with logging `numpy.int32/64`, one simple solution would be to just include those types in the tuple above. Happy to provide a PR if you think this is the way to go. *Ray version and other system information (Python version, TensorFlow version, OS):* ray: 0.8.6 python: 3.7.7 tensorboard: 2.2.2 ubuntu: 20.04 ### Reproduction (REQUIRED) ```python from ray import tune def trainable(config): tune.report(score=config["a"]) config_dict = {"a": tune.choice([1, 2, 3])} tune.run(trainable, config=config_dict, num_samples=1) ``` - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
yes! that'd be great - could you push a PR and ping me?
2020-07-03T17:40:47
ray-project/ray
9,300
ray-project__ray-9300
[ "9166" ]
c0ba337fe06ba6cef3df23dea65f7b59c0291e15
diff --git a/python/ray/resource_spec.py b/python/ray/resource_spec.py --- a/python/ray/resource_spec.py +++ b/python/ray/resource_spec.py @@ -3,6 +3,8 @@ import logging import multiprocessing import os +import subprocess +import sys import ray import ray.ray_constants as ray_constants @@ -229,12 +231,23 @@ def resolve(self, is_head, node_ip_address=None): def _autodetect_num_gpus(): """Attempt to detect the number of GPUs on this machine. - TODO(rkn): This currently assumes Nvidia GPUs and Linux. + TODO(rkn): This currently assumes NVIDIA GPUs on Linux. + TODO(mehrdadn): This currently does not work on macOS. + TODO(mehrdadn): Use a better mechanism for Windows. + + Possibly useful: tensorflow.config.list_physical_devices() Returns: The number of GPUs if any were detected, otherwise 0. """ - proc_gpus_path = "/proc/driver/nvidia/gpus" - if os.path.isdir(proc_gpus_path): - return len(os.listdir(proc_gpus_path)) - return 0 + result = 0 + if sys.platform.startswith("linux"): + proc_gpus_path = "/proc/driver/nvidia/gpus" + if os.path.isdir(proc_gpus_path): + result = len(os.listdir(proc_gpus_path)) + elif sys.platform == "win32": + props = "AdapterCompatibility" + cmdargs = ["WMIC", "PATH", "Win32_VideoController", "GET", props] + lines = subprocess.check_output(cmdargs).splitlines()[1:] + result = len([l.rstrip() for l in lines if l.startswith(b"NVIDIA")]) + return result
[tune][rllib] Windows: GPU not recognized ### What is the problem? I'm getting `ray.tune.error.TuneError: Insufficient cluster resources to launch trial`. I specified a GPU in my config but ray does not recognize my GPU (RTX 2080) and throws an error. I can get passed this by setting `num_gpus: 0` in my config for now. https://gist.github.com/juliusfrost/fa7ebbb8d1dfc66eea0bbc4babcbe5aa ### Reproduction (REQUIRED) ``` git clone https://github.com/juliusfrost/rllib-tune-atari.git cd rllib-tune-atari pip install -r requirements.txt python train.py --algo a2c ``` - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
#9114 @mehrdadn @mehrdadn Did you get a chance to look at this? I updated the script to reproduce to remove external dependencies.
2020-07-04T00:48:38
ray-project/ray
9,381
ray-project__ray-9381
[ "9364" ]
b97b474ae9ad360fff363291476d33cde7bee7a4
diff --git a/python/ray/tune/analysis/experiment_analysis.py b/python/ray/tune/analysis/experiment_analysis.py --- a/python/ray/tune/analysis/experiment_analysis.py +++ b/python/ray/tune/analysis/experiment_analysis.py @@ -65,6 +65,13 @@ def get_best_config(self, metric, mode="max"): mode (str): One of [min, max]. """ rows = self._retrieve_rows(metric=metric, mode=mode) + if not rows: + # only nans encountered when retrieving rows + logger.warning("Not able to retrieve the best config for {} " + "according to the specified metric " + "(only nans encountered).".format( + self._experiment_dir)) + return None all_configs = self.get_all_configs() compare_op = max if mode == "max" else min best_path = compare_op(rows, key=lambda k: rows[k][metric]) @@ -77,11 +84,20 @@ def get_best_logdir(self, metric, mode="max"): metric (str): Key for trial info to order on. mode (str): One of [min, max]. """ + assert mode in ["max", "min"] df = self.dataframe(metric=metric, mode=mode) - if mode == "max": - return df.iloc[df[metric].idxmax()].logdir - elif mode == "min": - return df.iloc[df[metric].idxmin()].logdir + mode_idx = pd.Series.idxmax if mode == "max" else pd.Series.idxmin + try: + return df.iloc[mode_idx(df[metric])].logdir + except KeyError: + # all dirs contains only nan values + # for the specified metric + # -> df is an empty dataframe + logger.warning("Not able to retrieve the best logdir for {} " + "according to the specified metric " + "(only nans encountered).".format( + self._experiment_dir)) + return None def fetch_trial_dataframes(self): fail_count = 0 @@ -161,7 +177,13 @@ def _retrieve_rows(self, metric=None, mode=None): idx = df[metric].idxmin() else: idx = -1 - rows[path] = df.iloc[idx].to_dict() + try: + rows[path] = df.iloc[idx].to_dict() + except TypeError: + # idx is nan + logger.warning( + "Warning: Non-numerical value(s) encountered for {}". + format(path)) return rows
diff --git a/python/ray/tune/tests/test_experiment_analysis.py b/python/ray/tune/tests/test_experiment_analysis.py --- a/python/ray/tune/tests/test_experiment_analysis.py +++ b/python/ray/tune/tests/test_experiment_analysis.py @@ -4,6 +4,7 @@ import random import os import pandas as pd +from numpy import nan import ray from ray.tune import run, sample_from @@ -38,6 +39,21 @@ def run_test_exp(self): "height": sample_from(lambda spec: int(100 * random.random())), }) + def nan_test_exp(self): + nan_ea = run( + lambda x: nan, + name="testing_nan", + local_dir=self.test_dir, + stop={"training_iteration": 1}, + checkpoint_freq=1, + num_samples=self.num_samples, + config={ + "width": sample_from( + lambda spec: 10 + int(90 * random.random())), + "height": sample_from(lambda spec: int(100 * random.random())), + }) + return nan_ea + def testDataframe(self): df = self.ea.dataframe() @@ -58,11 +74,15 @@ def testTrialDataframe(self): def testBestConfig(self): best_config = self.ea.get_best_config(self.metric) - self.assertTrue(isinstance(best_config, dict)) self.assertTrue("width" in best_config) self.assertTrue("height" in best_config) + def testBestConfigNan(self): + nan_ea = self.nan_test_exp() + best_config = nan_ea.get_best_config(self.metric) + self.assertIsNone(best_config) + def testBestLogdir(self): logdir = self.ea.get_best_logdir(self.metric) self.assertTrue(logdir.startswith(self.test_path)) @@ -70,6 +90,11 @@ def testBestLogdir(self): self.assertTrue(logdir2.startswith(self.test_path)) self.assertNotEquals(logdir, logdir2) + def testBestLogdirNan(self): + nan_ea = self.nan_test_exp() + logdir = nan_ea.get_best_logdir(self.metric) + self.assertIsNone(logdir) + def testGetTrialCheckpointsPathsByTrial(self): best_trial = self.ea.get_best_trial(self.metric) checkpoints_metrics = self.ea.get_trial_checkpoints_paths(best_trial)
[tune] Handling exceptions for the analysis module when encountering nans Hi, after running some experiments with tune I was trying to get the best configuration/log dir using `get_best_config` and `get_best_logdir` from the analysis module. Unfortunately, I had some `nan` values in the metric I was measuring, and both methods returned an error related to Pandas. After checking the code, I found that the line causing the error is in the function `_retrieve_rows`: https://github.com/ray-project/ray/blob/4da0e542d54c7977aae5bda5242cab15ab3f344b/python/ray/tune/analysis/experiment_analysis.py#L164 In particular, if there are `nan` values then the variable `idx` will be equal to `nan` and that will cause the following: `TypeError: Cannot index by location index with a non-integer key`. I was wondering if you could add an exception for this case, warning the user that the experiment returned some `nan` values. I had to spend some time understanding what was wrong.
Thank you for opening this issue! Sorry about this - I think we're not handling nans correctly throughout the system. We'll get to this next week, and I'll ping you once we have a patch merged (or feel free to push a PR too). Thanks!
2020-07-09T15:04:41
ray-project/ray
9,386
ray-project__ray-9386
[ "9366" ]
bed1be611e075a1e676c0bc66e26824e728108b2
diff --git a/rllib/agents/dqn/dqn_torch_model.py b/rllib/agents/dqn/dqn_torch_model.py --- a/rllib/agents/dqn/dqn_torch_model.py +++ b/rllib/agents/dqn/dqn_torch_model.py @@ -53,46 +53,45 @@ def __init__( self.dueling = dueling ins = num_outputs - # Dueling case: Build the shared (advantages and value) fc-network. advantage_module = nn.Sequential() - value_module = None - if self.dueling: - value_module = nn.Sequential() - for i, n in enumerate(q_hiddens): - advantage_module.add_module("dueling_A_{}".format(i), - nn.Linear(ins, n)) - value_module.add_module("dueling_V_{}".format(i), + value_module = nn.Sequential() + + # Dueling case: Build the shared (advantages and value) fc-network. + for i, n in enumerate(q_hiddens): + advantage_module.add_module("dueling_A_{}".format(i), nn.Linear(ins, n)) - # Add activations if necessary. - if dueling_activation == "relu": - advantage_module.add_module("dueling_A_act_{}".format(i), - nn.ReLU()) - value_module.add_module("dueling_V_act_{}".format(i), + value_module.add_module("dueling_V_{}".format(i), + nn.Linear(ins, n)) + # Add activations if necessary. + if dueling_activation == "relu": + advantage_module.add_module("dueling_A_act_{}".format(i), nn.ReLU()) - elif dueling_activation == "tanh": - advantage_module.add_module("dueling_A_act_{}".format(i), - nn.Tanh()) - value_module.add_module("dueling_V_act_{}".format(i), + value_module.add_module("dueling_V_act_{}".format(i), + nn.ReLU()) + elif dueling_activation == "tanh": + advantage_module.add_module("dueling_A_act_{}".format(i), nn.Tanh()) + value_module.add_module("dueling_V_act_{}".format(i), + nn.Tanh()) - # Add LayerNorm after each Dense. - if add_layer_norm: - advantage_module.add_module("LayerNorm_A_{}".format(i), - nn.LayerNorm(n)) - value_module.add_module("LayerNorm_V_{}".format(i), + # Add LayerNorm after each Dense. + if add_layer_norm: + advantage_module.add_module("LayerNorm_A_{}".format(i), nn.LayerNorm(n)) - ins = n - # Actual Advantages layer (nodes=num-actions) and - # value layer (nodes=1). + value_module.add_module("LayerNorm_V_{}".format(i), + nn.LayerNorm(n)) + ins = n + + # Actual Advantages layer (nodes=num-actions). + if q_hiddens: advantage_module.add_module("A", nn.Linear(ins, action_space.n)) - value_module.add_module("V", nn.Linear(ins, 1)) - # Non-dueling: - # Q-value layer (use main module's outputs as Q-values). - else: - pass self.advantage_module = advantage_module - self.value_module = value_module + + # Value layer (nodes=1). + if self.dueling: + value_module.add_module("V", nn.Linear(ins, 1)) + self.value_module = value_module def get_advantages_or_q_values(self, model_out): """Returns distributional values for Q(s, a) given a state embedding. diff --git a/rllib/examples/parametric_actions_cartpole.py b/rllib/examples/parametric_actions_cartpole.py --- a/rllib/examples/parametric_actions_cartpole.py +++ b/rllib/examples/parametric_actions_cartpole.py @@ -48,7 +48,7 @@ # TODO(ekl) we need to set these to prevent the masked values # from being further processed in DistributionalQModel, which # would mess up the masking. It is possible to support these if we - # defined a a custom DistributionalQModel that is aware of masking. + # defined a custom DistributionalQModel that is aware of masking. "hiddens": [], "dueling": False, }
[rllib] incorrect model output for DQN with torch and dueling=false <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? The output fo the DQN model is not within the action space. Something is wrong when constructing the torch model when dueling is off. The output dimension of the model is equal to whatever is passed in "fcnet_hiddens" instead of being of the size of the action space. *Ray version and other system information (Python version, TensorFlow version, OS):* - ray==0.9.0.dev0 - python 3.6.10 - mac OS ### Reproduction (REQUIRED) ```python import ray from ray import tune ray.init() config = { "env": "CartPole-v1", "num_workers": 1, "train_batch_size": 128, "learning_starts": 128, "model": {"fcnet_hiddens": [32]}, "dueling": False , "framework": "torch" } tune.run("DQN", name="MWE", config=config, stop={"training_iteration": 100}) ``` - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-07-09T22:12:56
ray-project/ray
9,429
ray-project__ray-9429
[ "9402" ]
e6225bdfa1d534fd4d70a61b70a3197425e79d1f
diff --git a/rllib/examples/custom_loss.py b/rllib/examples/custom_loss.py --- a/rllib/examples/custom_loss.py +++ b/rllib/examples/custom_loss.py @@ -31,7 +31,7 @@ type=str, default=os.path.join( os.path.dirname(os.path.abspath(__file__)), - "../tests/data/cartpole_small")) + "../tests/data/cartpole/small")) if __name__ == "__main__": ray.init()
diff --git a/rllib/agents/marwil/tests/test_marwil.py b/rllib/agents/marwil/tests/test_marwil.py --- a/rllib/agents/marwil/tests/test_marwil.py +++ b/rllib/agents/marwil/tests/test_marwil.py @@ -1,3 +1,5 @@ +import os +from pathlib import Path import unittest import ray @@ -18,19 +20,40 @@ def setUpClass(cls): def tearDownClass(cls): ray.shutdown() - def test_marwil_compilation(self): - """Test whether a MARWILTrainer can be built with all frameworks.""" + def test_marwil_compilation_and_learning_from_offline_file(self): + """Test whether a MARWILTrainer can be built with all frameworks. + + And learns from a historic-data file. + """ + rllib_dir = Path(__file__).parent.parent.parent.parent + print("rllib dir={}".format(rllib_dir)) + data_file = os.path.join(rllib_dir, "tests/data/cartpole/large.json") + print("data_file={} exists={}".format( + data_file, os.path.isfile(data_file))) + config = marwil.DEFAULT_CONFIG.copy() config["num_workers"] = 0 # Run locally. - num_iterations = 2 + config["evaluation_num_workers"] = 1 + config["evaluation_interval"] = 1 + config["evaluation_config"] = {"input": "sampler"} + config["input"] = [data_file] + num_iterations = 300 # Test for all frameworks. for _ in framework_iterator(config): trainer = marwil.MARWILTrainer(config=config, env="CartPole-v0") for i in range(num_iterations): - trainer.train() + eval_results = trainer.train()["evaluation"] + print("iter={} R={}".format( + i, eval_results["episode_reward_mean"])) + # Learn until some reward is reached on an actual live env. + if eval_results["episode_reward_mean"] > 60.0: + print("learnt!") + break + check_compute_single_action( trainer, include_prev_action_reward=True) + trainer.stop() diff --git a/rllib/tests/data/cartpole/large.json b/rllib/tests/data/cartpole/large.json new file mode 100644 --- /dev/null +++ b/rllib/tests/data/cartpole/large.json @@ -0,0 +1,21 @@ +{"type": "SampleBatch", "eps_id": [531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 531374291, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAALLBHb6TDbw8Rph7vL7yuL5YSR2+QxIwvhj7uLwUYJa91M4gvq51wDzIAsW8Bw2/vqdTIL6oYy++1hIBveeDtL2l1SO+/ufGPFBLCL2A/Me+WVYjvjZxLr7ASii9MGTevX7TJr5mz7q+CjAxvQv0Lj5tTC6+qTUtvv8wI72XZgq+QsMxvgI/2TxxQy690Vnhvjk4Mb7HASy+0FFSvU8IJb7mqDS+0oK5vq2FX71ESus9iBQ8vjN2Kr5PHFa9FytHvkt9P741u7i+RgtmvfZjpj3x4Ea+hxkOv29jX71sHbY+JT9Svi7st74CQEK9Hb49PYOaWb7AYCe+h3Q+vX2Zhb5981y+fzy3vsPUU72MA4k81EdkvufyJb4CdlK95GiVvnyZZ76ze7a+1V1qvXAxgbwd5m6+qPgMv5Goa70bQ4Q+NS16vhSltb4cf1a99LpUvaC4gL7swSK+ScBavdSpuL5JWYK+Xd+0vhlMeL3rrq69WveFvvgmDL/bSH+95ipAPoKSi76V3D2/ROlvvQCM6z6xKpO+xZZvv0Q5Sr2k1Ds/FcCcvup4Pb8uHg693hDaPkhUpL4fS2+/Z3TWvNcbNT+l5q2+f0A9vxYXRbxLNNA+lni1vvFBC7+bW3+7ca/hPZUKu77HMT2/pdbduvCkzT7vm8K+sjwLv9vDzzsXEt49uS3IviA2Pb/xaQs8XmTOPj+/z76XSgu/nMCHPC+s5z2XUdW+Z009v0RJmjzDbNI+DOTcvrdrC79Fn9083ZD+Pbd34r6NJLO+x/zxPG6PI77tDOa+G+0evlnR1zzeyeK+x6Pnvh2Ns77VPo88jHwRvhQ7674xygu/aO9vPBrRHz6H0vC+Pcmzvs2JkTyAHwe+CGv0vk8IIL5S1nc8wYTWvrcE9r7XBrS+MhfdO+r4+L10nvm+JFkgvmJrjTvCANO+8Tj7vpQbtL4SqoC7LtHxvRjT/r6a/Qu/ugvOuxGHMT5NNgK/dQi0vjfbOLuLaPi9LwMEvxb1C78466u7LJcuPsTPBr9a+bO+PrnwuiSd/b2AnAi/xh0gvltWjbuyjdW+c2kJv93ts77IV0+8l8sAvhE2C79q3gu/sY54vNnLJj4yAg6/xrmzvr0uQ7ylxQm+S84PvxbFC78HRW+8+w8ePuqZEr8kiLO+jbA8vGZTEr6EZRS/fxcfvo+Da7yI3eC+JzEVv7lXs77Atr2837IavkX8Fr8miQu/NXfWvKVvCT6xxhm/dWM9v9N5wLzlSNY+XZAdv6ZaC789z3e8B7/yPdtZIL9NsrK+W/hQvFcuN75RIyK/U2cdvkXLhbx+ffO+y+wiv/d7sr79tdO8+pZAvre1JL+Ayxy+d4byvMBD+r5pfiW/bRSyvhNOIb1fhlK+TEYnv2nXCr+fJTK9WWaYPSoNKr+5ebG+DA0svU06bb6A0yu/cIcKv3kHP71pdUI9xZguv2jUsL7YIzu9vOOEvnRdML8WMQq/AGdQvU+Qljz/IDO/c/I7v4/lTr3c05Y+SeM2v3m0bb+pwza9xCcSP1akO7/gmTu/nv4HvSlvhz7aZD+/XIgJv3am5Lweex29BSVCv+5gO78R8+q88i97PmbkRb/YOG2/b8LCvHZfBz/6okq/8y87v+I9WLwRNmo+YGFOvz0wCb9NSw28VXWLvccfUb+UGzu/hJsjvOItYz7F3VS/zx4Jv/vRtbsKeZe91JtXvxQNO7+iSua7SSxePodZW78HEwm/oDMwu4aXn71ZF16/AgQ7v5Uri7ugCls+3tRhv4gMCb/3FgI4txGkvY+SZL8TADu/e/HNusiuWT4AUGi/HfJsv4qpLzt0OAE/Kg1tvxUBO79eUVE8fgpaPqDKcL+xGQm/pIuLPPsFm72UiHO/7xw7v4dJfjwYpmM+mUZ3v8Eebb9CkaM8+RQFP6cEfL8IPju/Z734PJEhbz5Vw3+/GE1tvxuADz0SHwk/KEGCv9yvj78XYTs9F9JWP9Yghb8Il22/mQ+APTGqDz8RgYe/hOA7vxwMlz1445M+CGKJv5kTbr/e4KI9uJUaP4LDi7+Jbjy/qZy7PUOqrD7kpY2/wNcKv9hsyT0RK5k9VAmPv/4dPb8Qfcw9zirLPnjtkL/sjwu/7r3cPdd+DD6/UpK/Exe0vppc4j1x/vS9QzmTv4ZaDL88dt09RrhSPpKglL+7rLW+AeTlPWWaUL0diZW/WlwlvvvN4z1ZxZy+8fKVv6mfAj1QQ9c9BtQPvwvelb9zfCi+HEDAPfWtc77gSZa/c8O4vtWAtj3yi6k9XzaXvw6hDr/p5Lk9TO3NPoGjmL++4EC/S17KPTzZOD9FkZq/P08Pv7Hx5z1Akew+JACcv7eeQb+X3vo9icVJP8/vnb8rJhC/lJMNPldRCT/VYJ+/+4O9vtiPGD73npQ+aVOgvwiwNb65gR4+AMZEPbHHoL9wsL++mH0fPoRDxT4OvaG/+SU6vpNhJz7i9hQ+7igFO7Mks7yWAOQ88HQovXD+0DoA/jA+l0PdPFlZpr4ChqU7ppC5vEcIqDy0KcO8o62WO6VLX74LIaQ82u6MPvWceDkSUL68RzrRPJTVNLxslm65oKEvPldrzzzQUpe+e+VRO1mHuz7s/p48YIQUvxN+LDzI+i4+h8f/Oy0VkL5nfGQ8tVW7PgG1DjvYVRK/qzCuPGHVLj7poRe8WHeOvtkpyjzcYrs+ks9yvEvnEr9DEAM98y4vPnRs17zKXJK+BBQRPf7Rwbw8IQO9DV3QutYjDz0yjF++kkIDvQfIjz4qg/o8VW/TvpKC2LzGLxA/a9q2PBCyXr4Tdni8LFGGPs84kzwZYbW8kH8ivLK5D70lmI8800YxPg//Lby8aam+ZvWrPGr5srzYNY28BUUqvQ1hqDyLrjE+aAWUvG7lrb7lzsQ8YN+8Pvyqy7zJUSO/q58APdtNMj6fGBq9sNi0vlbjDj18dqe8GAg3vZu+lL2iNg09tA1cvj37PL3vj1I+3zf3PI6T0b7sIiy9CYH3PmAntDxixFq+KIkEvfULNj6sJpE8rgeUvK7x67yR3P+9wjCOPM7fWb7dNAC9x1QiPkKpVjxFAI28pHDmvCdRE75pBVE8lAFZvr8C/rwuKw8+Q5QLPNkr0L6PGue8G1nYPmcvyznqI1i+U9+hvGQK+D3tQnu7KNPPvnQHjryFo9A+wNJDvPWYV76Ihxa80AzgPUBohLwXos++41zlu/ZfzD6P2ca8SVxXvqrygDoDldU9uU7pvPMeebyfKkk7KZlAvnnM67x2a1e+8G81uq0y2D0IIge9Aq15vMcDujr11D++nmEIvS1uV76TiRi7x6rYPaKdGb16dnm8Ye1dufIfQL7y3Bq9qWRXviXlgbs/B9c9MhgsveGRz74YWPS6FPjKPkJOTb3KTle+4LbGO1NC0z3Dh169i3p5vJwoBTzAHEC+GMdfvfGDV76KXY87BWvcPdkEcb0tsc++MObVO8upzT71H4m9vKxXvvaSbjzDd+M9eMCRvVfYz74FfIk87w3RPiNhor0NIli+w2HMPDm89z1WBqu9tyXQvmEz4Dxxw9c+Mq27vcflWL5cnxI9GMgMPjpaxL1kFo28keIdPRglE77RDsW9JJU1PgodEj0TE9m+acu9vfTDlbxWw94840v2vR2Lvr22rDQ+Lw/LPPr4zr4BUbe9MNmbvP7TiDwqr9S9fRi4veOGW75xoG88iLZGPnLgwL3Pg5+8gZuXPHR3wL2grMG9xpYzPso1iDyf88K+on26vfmgo7y0phM8OMCpvRRPu727Y1y+avvwO7q6WT7eH8S9kibSvhkqPjxj8wE/we/UvXaqXL4tQLI8Nd1fPl/D3b2Jfam8pxHWPOx3ib1RnN69WV1dvkwSyzwDTG8+GHfnvT1zr7zmW/E8SzJRvatX6L25bTE+uv3oPJUrq77NPuG9WWe8PnY3sjx+Lh6/SyzSvciyMD759Rk84Qqjvucay70p8Lm8XXRGO8jdurznCMy9axdfvlGOKDu0qYo+XPXUvSCYurwc4gI8ymWsvDPk1b3EXDA+hPn3OyRVn75A1s69qWq8vKsdsDrOMIS8bMfPvag7MD6c0IU6++Wdvsy6yL2asby8/Keou34ofLxTrMm9GUswPhW+srsSkJ6+FJ/Cvblzu7z32T68r3qZvAWPw704EF++mP1EvC1eij4we8y95624vM7e2LuMo9a8k2fNvQfRMD6cCuq7flWkvvpUxr1YEbe8xzFevN80+rxNP8e9Kx4xPuUzaLycqqe+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 1, 1, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 0, 0, 1, 1, 1, 0, 1, 1, 0, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 0, 0, 1, 1, 1, 0, 1, 1, 0, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAFhJHb5DEjC+GPu4vBRglr3UziC+rnXAPMgCxbwHDb++p1MgvqhjL77WEgG954O0vaXVI77+58Y8UEsIvYD8x75ZViO+NnEuvsBKKL0wZN69ftMmvmbPur4KMDG9C/QuPm1MLr6pNS2+/zAjvZdmCr5CwzG+Aj/ZPHFDLr3RWeG+OTgxvscBLL7QUVK9TwglvuaoNL7Sgrm+rYVfvURK6z2IFDy+M3Yqvk8cVr0XK0e+S30/vjW7uL5GC2a99mOmPfHgRr6HGQ6/b2NfvWwdtj4lP1K+Luy3vgJAQr0dvj09g5pZvsBgJ76HdD69fZmFvn3zXL5/PLe+w9RTvYwDiTzUR2S+5/IlvgJ2Ur3kaJW+fJlnvrN7tr7VXWq9cDGBvB3mbr6o+Ay/kahrvRtDhD41LXq+FKW1vhx/Vr30ulS9oLiAvuzBIr5JwFq91Km4vklZgr5d37S+GUx4veuurr1a94W++CYMv9tIf73mKkA+gpKLvpXcPb9E6W+9AIzrPrEqk77Flm+/RDlKvaTUOz8VwJy+6ng9vy4eDr3eENo+SFSkvh9Lb79ndNa81xs1P6Xmrb5/QD2/FhdFvEs00D6WeLW+8UELv5tbf7txr+E9lQq7vscxPb+l1t268KTNPu+bwr6yPAu/28PPOxcS3j25Lci+IDY9v/FpCzxeZM4+P7/PvpdKC7+cwIc8L6znPZdR1b5nTT2/REmaPMNs0j4M5Ny+t2sLv0Wf3TzdkP49t3fivo0ks77H/PE8bo8jvu0M5r4b7R6+WdHXPN7J4r7Ho+e+HY2zvtU+jzyMfBG+FDvrvjHKC79o7288GtEfPofS8L49ybO+zYmRPIAfB74Ia/S+TwggvlLWdzzBhNa+twT2vtcGtL4yF9076vj4vXSe+b4kWSC+YmuNO8IA077xOPu+lBu0vhKqgLsu0fG9GNP+vpr9C7+6C867EYcxPk02Ar91CLS+N9s4u4to+L0vAwS/FvULvzjrq7ssly4+xM8Gv1r5s74+ufC6JJ39vYCcCL/GHSC+W1aNu7KN1b5zaQm/3e2zvshXT7yXywC+ETYLv2reC7+xjni82csmPjICDr/GubO+vS5DvKXFCb5Lzg+/FsULvwdFb7z7Dx4+6pkSvySIs76NsDy8ZlMSvoRlFL9/Fx++j4NrvIjd4L4nMRW/uVezvsC2vbzfshq+RfwWvyaJC781d9a8pW8JPrHGGb91Yz2/03nAvOVI1j5dkB2/ploLvz3Pd7wHv/I921kgv02ysr5b+FC8Vy43vlEjIr9TZx2+RcuFvH59877L7CK/93uyvv2107z6lkC+t7Ukv4DLHL53hvK8wEP6vml+Jb9tFLK+E04hvV+GUr5MRie/adcKv58lMr1ZZpg9Kg0qv7l5sb4MDSy9TTptvoDTK79whwq/eQc/vWl1Qj3FmC6/aNSwvtgjO72844S+dF0wvxYxCr8AZ1C9T5CWPP8gM79z8ju/j+VOvdzTlj5J4za/ebRtv6nDNr3EJxI/VqQ7v+CZO7+e/ge9KW+HPtpkP79ciAm/dqbkvB57Hb0FJUK/7mA7vxHz6rzyL3s+ZuRFv9g4bb9vwsK8dl8HP/qiSr/zLzu/4j1YvBE2aj5gYU6/PTAJv01LDbxVdYu9xx9Rv5QbO7+EmyO84i1jPsXdVL/PHgm/+9G1uwp5l73Um1e/FA07v6JK5rtJLF4+h1lbvwcTCb+gMzC7hpefvVkXXr8CBDu/lSuLu6AKWz7e1GG/iAwJv/cWAji3EaS9j5JkvxMAO7978c26yK5ZPgBQaL8d8my/iqkvO3Q4AT8qDW2/FQE7v15RUTx+Clo+oMpwv7EZCb+ki4s8+wWbvZSIc7/vHDu/h0l+PBimYz6ZRne/wR5tv0KRozz5FAU/pwR8vwg+O79nvfg8kSFvPlXDf78YTW2/G4APPRIfCT8oQYK/3K+PvxdhOz0X0lY/1iCFvwiXbb+ZD4A9MaoPPxGBh7+E4Du/HAyXPXjjkz4IYom/mRNuv97goj24lRo/gsOLv4luPL+pnLs9Q6qsPuSljb/A1wq/2GzJPRErmT1UCY+//h09vxB9zD3OKss+eO2Qv+yPC7/uvdw9134MPr9Skr8TF7S+mlziPXH+9L1DOZO/hloMvzx23T1GuFI+kqCUv7ustb4B5OU9ZZpQvR2Jlb9aXCW++83jPVnFnL7x8pW/qZ8CPVBD1z0G1A+/C96Vv3N8KL4cQMA99a1zvuBJlr9zw7i+1YC2PfKLqT1fNpe/DqEOv+nkuT1M7c0+gaOYv77gQL9LXso9PNk4P0WRmr8/Tw+/sfHnPUCR7D4kAJy/t55Bv5fe+j2JxUk/z++dvysmEL+Ukw0+V1EJP9Vgn7/7g72+2I8YPveelD5pU6C/CLA1vrmBHj4AxkQ9scegv3Cwv76YfR8+hEPFPg69ob/5JTq+k2EnPuL2FD4wNKK/BrkuPEZcKj7wabu9cP7QOgD+MD6XQ908WVmmvgKGpTumkLm8RwioPLQpw7yjrZY7pUtfvgshpDza7ow+9Zx4ORJQvrxHOtE8lNU0vGyWbrmgoS8+V2vPPNBSl7575VE7WYe7Puz+njxghBS/E34sPMj6Lj6Hx/87LRWQvmd8ZDy1Vbs+AbUOO9hVEr+rMK48YdUuPumhF7xYd46+2SnKPNxiuz6Sz3K8S+cSv0MQAz3zLi8+dGzXvMpckr4EFBE9/tHBvDwhA70NXdC61iMPPTKMX76SQgO9B8iPPiqD+jxVb9O+koLYvMYvED9r2rY8ELJevhN2eLwsUYY+zziTPBlhtbyQfyK8srkPvSWYjzzTRjE+D/8tvLxpqb5m9as8avmyvNg1jbwFRSq9DWGoPIuuMT5oBZS8buWtvuXOxDxg37w+/KrLvMlRI7+rnwA9200yPp8YGr2w2LS+VuMOPXx2p7wYCDe9m76UvaI2DT20DVy+Pfs8ve+PUj7fN/c8jpPRvuwiLL0Jgfc+YCe0PGLEWr4oiQS99Qs2PqwmkTyuB5S8rvHrvJHc/73CMI48zt9Zvt00AL3HVCI+QqlWPEUAjbykcOa8J1ETvmkFUTyUAVm+vwL+vC4rDz5DlAs82SvQvo8a57wbWdg+Zy/LOeojWL5T36G8ZAr4Pe1Ce7so08++dAeOvIWj0D7A0kO89ZhXvoiHFrzQDOA9QGiEvBeiz77jXOW79l/MPo/ZxrxJXFe+qvKAOgOV1T25Tum88x55vJ8qSTspmUC+eczrvHZrV77wbzW6rTLYPQgiB70CrXm8xwO6OvXUP76eYQi9LW5XvpOJGLvHqtg9op0ZvXp2ebxh7V258h9AvvLcGr2pZFe+JeWBuz8H1z0yGCy94ZHPvhhY9LoU+Mo+Qk5NvcpOV77gtsY7U0LTPcOHXr2Lenm8nCgFPMAcQL4Yx1+98YNXvopdjzsFa9w92QRxvS2xz74w5tU7y6nNPvUfib28rFe+9pJuPMN34z14wJG9V9jPvgV8iTzvDdE+I2GivQ0iWL7DYcw8Obz3PVYGq723JdC+YTPgPHHD1z4yrbu9x+VYvlyfEj0YyAw+OlrEvWQWjbyR4h09GCUTvtEOxb0klTU+Ch0SPRMT2b5py7299MOVvFbD3jzjS/a9HYu+vbasND4vD8s8+vjOvgFRt70w2Zu8/tOIPCqv1L19GLi944ZbvnGgbzyItkY+cuDAvc+Dn7yBm5c8dHfAvaCswb3GljM+yjWIPJ/zwr6ifbq9+aCjvLSmEzw4wKm9FE+7vbtjXL5q+/A7urpZPt4fxL2SJtK+GSo+PGPzAT/B79S9dqpcvi1Asjw13V8+X8PdvYl9qbynEdY87HeJvVGc3r1ZXV2+TBLLPANMbz4Yd+e9PXOvvOZb8TxLMlG9q1fovbltMT66/eg8lSurvs0+4b1ZZ7w+djeyPH4uHr9LLNK9yLIwPvn1GTzhCqO+5xrLvSnwubxddEY7yN26vOcIzL1rF1++UY4oO7Spij5c9dS9IJi6vBziAjzKZay8M+TVvcRcMD6E+fc7JFWfvkDWzr2pary8qx2wOs4whLxsx8+9qDswPpzQhTr75Z2+zLrIvZqxvLz8p6i7fih8vFOsyb0ZSzA+Fb6yuxKQnr4Un8K9uXO7vPfZPryvepm8BY/DvTgQX76Y/US8LV6KPjB7zL3nrbi8zt7Yu4yj1ryTZ829B9EwPpwK6rt+VaS++lTGvVgRt7zHMV683zT6vE0/x70rHjE+5TNovJyqp76dKcC99NGzvC/Bqbwv8iC9lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.8949677348136902, 0.2854820489883423, 0.9052359461784363, 0.24984733760356903, 0.9164604544639587, 0.7893484234809875, 0.5310271382331848, 0.18627631664276123, 0.9338834285736084, 0.8487390875816345, 0.40080294013023376, 0.8715166449546814, 0.658420979976654, 0.689258873462677, 0.3081173002719879, 0.8998115062713623, 0.26141199469566345, 0.913749635219574, 0.7835085988044739, 0.5238140225410461, 0.1894565224647522, 0.9329298734664917, 0.8453688025474548, 0.6032693982124329, 0.27685293555259705, 0.893722414970398, 0.2753779888153076, 0.8976496458053589, 0.7433987259864807, 0.5804827213287354, 0.7586576342582703, 0.5516151785850525, 0.7794906497001648, 0.5118468999862671, 0.8042495846748352, 0.538456916809082, 0.24315714836120605, 0.9007118940353394, 0.748458981513977, 0.5794850587844849, 0.27144572138786316, 0.8901560306549072, 0.27271538972854614, 0.8910347819328308, 0.7340371608734131, 0.5822507739067078, 0.7293457388877869, 0.5845097899436951, 0.27595728635787964, 0.8879104852676392, 0.7372824549674988, 0.5629269480705261, 0.7396020293235779, 0.5552812218666077, 0.25820234417915344, 0.8933025002479553, 0.7614201307296753, 0.48816370964050293, 0.7862693667411804, 0.5133485794067383, 0.2381698489189148, 0.897735059261322, 0.2182164043188095, 0.9072064757347107, 0.8073776960372925, 0.4133623540401459, 0.8243219256401062, 0.3771069347858429, 0.8413662314414978, 0.6608332395553589, 0.3812047839164734, 0.8311269283294678, 0.6087785363197327, 0.6787657141685486, 0.4062293469905853, 0.8176510334014893, 0.5982443690299988, 0.6696091890335083, 0.5969210267066956, 0.6648812294006348, 0.5974056124687195, 0.6586463451385498, 0.5995053648948669, 0.6510249972343445, 0.39692065119743347, 0.8157919049263, 0.6219735145568848, 0.6127027869224548, 0.36448463797569275, 0.8313741683959961, 0.337843656539917, 0.15515053272247314, 0.9258947372436523, 0.8647976517677307, 0.25696811079978943, 0.8831287026405334, 0.784720778465271, 0.3644827902317047, 0.8195744752883911, 0.6923335194587708, 0.4548424482345581, 0.7395390868186951, 0.6029321551322937, 0.47271326184272766, 0.6373288035392761, 0.4909321963787079, 0.3159151077270508, 0.16873405873775482, 0.911851704120636, 0.14181312918663025, 0.9227877855300903, 0.882472813129425, 0.8231956362724304, 0.25083988904953003, 0.8550863265991211, 0.7966943383216858, 0.5455768704414368, 0.7930395007133484, 0.43641868233680725, 0.8619645237922668, 0.6047736406326294, 0.24368080496788025, 0.9141198992729187, 0.2317906618118286, 0.9197134375572205, 0.20742760598659515, 0.9279178977012634, 0.8253203630447388, 0.5318312048912048, 0.19953255355358124, 0.9282232522964478, 0.8053845763206482, 0.45844629406929016, 0.8473955392837524, 0.425632119178772, 0.13817322254180908, 0.9456128478050232, 0.8870267868041992, 0.6990184783935547, 0.3420375883579254, 0.8814630508422852, 0.6405285000801086, 0.7468997836112976, 0.6068054437637329, 0.7705122232437134, 0.42884406447410583, 0.8519783020019531, 0.4350297152996063, 0.8528088331222534, 0.42816853523254395, 0.8586954474449158, 0.5914314985275269, 0.7693524956703186, 0.5925441980361938, 0.7694152593612671, 0.5915730595588684, 0.7708016633987427, 0.41154390573501587, 0.8666397929191589, 0.6139383912086487, 0.7519904375076294, 0.378618448972702, 0.8802904486656189, 0.3438532054424286, 0.893405556678772, 0.29963135719299316, 0.9077726602554321, 0.7506021857261658, 0.4040895104408264, 0.8543979525566101, 0.4241299331188202, 0.8495427966117859, 0.5679969787597656, 0.7946709990501404, 0.4594833254814148, 0.836846113204956, 0.541279137134552, 0.19262954592704773, 0.9331842660903931, 0.8330914974212646, 0.45803844928741455, 0.8505011200904846, 0.5852563977241516, 0.23887963593006134, 0.9138653874397278, 0.773739218711853, 0.4239019453525543, 0.8591747879981995, 0.5933403372764587, 0.7661020755767822, 0.5868037343025208, 0.7736600041389465, 0.5719375610351562, 0.7861037254333496, 0.45181483030319214, 0.8443316221237183, 0.5466866493225098, 0.8014476299285889, 0.5193928480148315, 0.8187002539634705], "advantages": [0.21745966374874115, -1.3587554693222046, 0.2890452742576599, -1.3989312648773193, 0.5305141806602478, -1.250394582748413, -2.47453236579895, -1.1006877422332764, 1.2681771516799927, -0.5738205313682556, -2.072775363922119, -0.05673271417617798, -1.6238951683044434, -2.815504789352417, -1.2751059532165527, 0.8723791837692261, -0.6053474545478821, 1.4760985374450684, 0.1886976808309555, -1.1512235403060913, 0.7168366312980652, 2.4015207290649414, 1.5091012716293335, 0.5575752854347229, -0.7179286479949951, -2.1994543075561523, -0.5006898045539856, -2.078684091567993, -0.5276636481285095, 0.3403717279434204, -0.5139458775520325, 0.07105739414691925, -0.6441061496734619, -0.31162527203559875, -0.9114060997962952, -0.7999306321144104, -0.9937840700149536, -0.7982866764068604, -1.1026670932769775, -1.1428343057632446, -1.4890213012695312, -1.4131602048873901, -1.5820198059082031, -1.516117811203003, -1.6305266618728638, -1.5550901889801025, -1.9545233249664307, -1.8626054525375366, -2.2881417274475098, -2.2834410667419434, -2.306039571762085, -2.234785556793213, -2.5949575901031494, -2.542801856994629, -2.884005546569824, -2.8175363540649414, -2.847313165664673, -2.884158134460449, -3.3757777214050293, -3.433220863342285, -3.6368627548217773, -3.441377878189087, -3.580565929412842, -3.222698211669922, -3.444593906402588, -3.7557919025421143, -3.542888879776001, -3.941218376159668, -3.5916621685028076, -4.086845874786377, -5.003336429595947, -6.441424369812012, -5.719454288482666, -5.232696533203125, -6.273830890655518, -7.807031154632568, -7.120766639709473, -6.63323450088501, -7.815937519073486, -7.302456378936768, -8.561629295349121, -8.027447700500488, -9.35847282409668, -8.809571266174316, -10.206872940063477, -11.984430313110352, -11.363447189331055, -10.877972602844238, -12.373501777648926, -14.212943077087402, -13.69473648071289, -15.550026893615723, -17.74781036376953, -17.18828773498535, -16.90947723388672, -18.776329040527344, -18.59465217590332, -18.61871337890625, -20.198728561401367, -20.302513122558594, -20.4595890045166, -21.88591194152832, -22.124610900878906, -22.25788116455078, -22.106473922729492, -23.6201171875, -25.072233200073242, -26.655244827270508, -28.601224899291992, -28.853046417236328, -30.872180938720703, -31.212308883666992, -31.802932739257812, -32.508113861083984, -34.050262451171875, -34.85260772705078, 16.146289825439453, 15.883732795715332, 15.914982795715332, 16.893966674804688, 15.654492378234863, 15.357436180114746, 15.64351749420166, 15.129483222961426, 15.430910110473633, 14.906143188476562, 15.246929168701172, 14.690457344055176, 14.721792221069336, 15.656871795654297, 17.956584930419922, 15.197473526000977, 14.025788307189941, 13.805998802185059, 13.732562065124512, 13.539511680603027, 14.07516098022461, 13.330644607543945, 13.174162864685059, 13.762556076049805, 15.580601692199707, 13.255160331726074, 12.447029113769531, 12.825960159301758, 12.111981391906738, 12.403304100036621, 13.778947830200195, 11.94132137298584, 13.174569129943848, 11.498538970947266, 12.612510681152344, 11.069765090942383, 10.65079116821289, 10.683614730834961, 10.3037109375, 10.296500205993652, 9.956096649169922, 9.908475875854492, 10.686444282531738, 9.485438346862793, 9.20966911315918, 9.094189643859863, 9.749462127685547, 8.668691635131836, 9.283961296081543, 8.241490364074707, 8.84189510345459, 7.812507152557373, 7.534442901611328, 7.739901065826416, 7.127336502075195, 7.329253196716309, 6.721649646759033, 6.588075160980225, 6.28385591506958, 6.476897716522217, 5.866865158081055, 5.7219438552856445, 6.298417091369629, 5.243937969207764, 4.915683746337891, 4.780351161956787, 4.428664684295654, 4.520269393920898, 5.244883060455322, 4.0799174308776855, 3.508349895477295, 3.3825762271881104, 3.0049028396606445, 3.100775718688965, 2.517244577407837, 2.624274969100952, 2.0264055728912354, 2.1551806926727295, 1.5343471765518188, 1.3545794486999512, 1.0044246912002563, 1.1741865873336792, 0.4995787739753723, 0.7142715454101562]} +{"type": "SampleBatch", "eps_id": [1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 1857488294, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAJ0pwL300bO8L8GpvC/yIL3JD8G90PhdvkUxsLzts3w+x/DJvf/frryQwoe8D3xXvZ7Qyr1NLzI+IGGQvEtxs74BsMO9KN+qvA/Nybzl14G9uYrEvW7MMj5AMNS8uz26vtNjvb0+9aS8kOQHvSSBor34Nr69A+xbvpxkDr3cjE8++ALHvSkCnbz1k/u8BVrOvfHLx71/91q+BQsGvUNzOj4pjtC9XZaVvApB7ryiSPe9ok3RvYQQWr60BAG9tYYmPp0G2r13gI68f2TnvAsvD74Evdq9IjFZvk1N/rxhQxM+EG3jvY9D0L5nvea8smPaPk4W9L2EU1i+8NqgvPgdAD59vfy9humBvENbjLzw3TG+xmP9veDJV76y0Ki83ILoPbkCA756pM++3TaWvNedzD78UAu+/TdXvlp5KbwIVM89554PvmRvz74wTQi8dAHIPgrrF74K8la+GMQEuppCwz2PNxy+cCFyvM6MtzqmPUq+C4UcvmL1Vr7eFye7SdbDPaHRIL5z3HG8tgcnuoScSr4GHyG+qbI4PtCMlrsiQ/u+X20dvialcLw2FWy8sk1MvmG6Hb6UoVa+4LqWvHlttT1KBSK+y/povD03iLwa4Va+2E8ivmAeVr6ymKq8zdGePSGYJr5zO2C8EuSdvEnzYr7j3ya+LYhVvvUzwryp74Q9KyUrvoh9zr5skbe83zWzPqFnM75h21S+AnF8vJxCTj11qTe+/0RNvM3wa7xEFH2+Jes3vt4uOz6Gdp68ulkLv8QsNL63aEW8taX3vOz/g77wazS+HbBTvpDxEL0xUM48x6c4vnIdNrxm4Q699pCOvg7iOL4BqVK+6rAlvSeYRjujGD2+hezMvl1xJb2Iu5A+D0tFvj96Ub4eSQ699wO4vJV7Sb6/4xK8MiAQvWncpr6Wqkm+cXRQvtPSKr29WDa94NVNvrXPy75weC69flpwPuf8Vb7XsRe//z0bvcx/Az+cH2K+cznLvipT4rxqT1Y+oUBqvqZ1F78FCcC8X3v8PoVedr782cq+ZntevGrJRT64e36+sbNNvrUwH7zhjNS9dUyBvnKuyr7EMkG860Y+PjBahb4jY02+Rk8EvFBu4r36Z4e+TonKvuKJKLws3zc+93SLvmsuF784Ztu7dh7wPg+Bkb6Aacq+8+cvO5xgMj5pjZW+xioXv1Mdyjuree8+XJmbvgR6yr5KUn48kj01Pgumn74aYU2+wiicPJvS4r3Qs6G+BoC7u24Dijz9csu+0MKhvqLnTb6+0RE8IprLve7Ro74Wase7Z3ziO0pPx77i4aO+5yJOvkgLZbqwYMG9l/GlvtX2yr7SBTW7b7xKPsQAqr5KFU6+ePWcOnK5w71XEKy+6vPKvgAjO7pUO0o+dR+wvhsXTr6NElQ7g2nDvQ0vsr6j+Mq+cwSuOn4LSz5DPra+NihOvv1zrTtud8C9Bk64vgUFy75Cul876y1NPnxdvL7WSE6+ty3zO6/Yur2Tbb6+SBnLvkJjtzsWrFA+cH3Cvpx5Tr4OeB48onGyvQSOxL4hCNq7/+oBPOvkwL51n8S+l7tOvpbmzjk+EKe9sbDGvsng3bu0HaK6Ro6/vnLCxr52uE6+NdwOvI2bp72m08i+uzLLvmutKbzOElU+BuTMvv9wTr7h/Mq72euzvYP0zr4cetG7BEgCvNPUw75FBc++CDxOvgadf7wrEb29OxXRvkjnyr6b7o68hxdIPhkk1b6EVhe/rtVdvD0O9z7MMdu+Tq3KvlvgfrvbED4+gT/fvgd+Tb5AeTm5VMrdvZBN4b4Y6rW7tYkZu29Uzb4eXOG+onRNvqrLKbxSa9+9FWrjvt0BsLvqik28jV/Pvip4477jHk2+eSGpvHw77r1GheW+FZ2gu3kwvLyztdS+H5Llvqp7TL7UIAC9zjcFvpmd575W6Mm+IskKvco3HD5dp+u+1sYWv5ST/Lx/ZN4+Uq/xvuCcSL8uabW8Ppg3P5i1+b7ae3q/GqT/u69bgD9G3QG/kIFIv73GSDw1KjU/3d8Fv1GZFr9sVdg8w3bWPu7iCL/vfcm+J3sOPSndCT7A5gq/2bRLvpuCGT2zZha+f+sLvxgFyr5keg09ECshPqrwDb8gwky+HV8aPbtb/r3C9g6/Dtmwu3wyED1LPM++1f0Ov/TSTb48FN48qjrPvUkFEL8FAsu+L4DNPB+5TD78DBK/XJZOvqRB7jxUjK29axUTvztny75gX+A8ri5ePiEeFb/OaU+++/UBPesZib2eJxa/oNXLviL0+DxvO3E+bzEYv2VSUL6Axg89zwJCvRY8Gb/UT8y+KuULPUUqgz4gRxu/slVRvrDhID00MNG8E1Mcv34OIrwqyh4963OcvgpgHL/+eVK+3MEFPaL5cLpzbR2/rakyvJSuBT129ZC+vnsdv4YlPD4i+tw83LkQv+qKHL/ZxsE+OVqAPElyWb/Ymhq/kIg7PjkDrbrZRA2/bzD4u/I/rzsa5wU9ld1EvCiv9LvbykK+HesEPbTjlD75rDi8ksGQO6C9HD1krgs8Zjo3vOm+Sz5rcB098C6LvhcP7LsdsVk7eCsHPaf8BT3M4em7ybRKPnvZCT1NXX++1UxQu65FGzs81+o87A9cPdcxTbtgzEk+rKTzPGpKa77EbVQ645PIPir/zTxzTwO/gqUNPEr/SD6O6nM8j4dZvjP3TTwoSHo6iU4uPJDlrT1LR048759IPlkhSjxvTVG+PD2HPJg1JDpKJwc8Nrq8PYJXhzwDU0g+kFklPLyqSr7KZKc8cwK/OSz+yDvRkMg9EnSnPF4WSD5ClgQ8tm9FvqZ3xzzq4sc+itCKO1pC974stwM9LOhHPlutsbtsckG+RLUTPWxFOjnavRa87AHRPf24Ez2afEe+z5nqu7KcyD6BwwM9lcfXORd2MTqxbcY9IswDPWU3SD4LXCs7OEZIvpHQEz3O/cc+rPupumKR+b430DM9XjJIPqr4NLzp2Ee+PtRDPW4wDjoo7HS80IrAPZ7fQz3eBke+oB1WvC2Owz6M8zM9lC53OqvrsbtWba49Uwc0PYbZSD4zNXS7Lj5WvroYRD2+Wsg+FJwBvLLJAL8/J2Q95gRJPqU6k7z6AVq+Hjx0PYuKyD49HLa8ut0CvyQpij3xmkk+rO4EvcUDZ76UOZI9D5USO9tpF70gIGg9CFGSPc7zRL5fxRK9T8isPj1wij0Ov1Y7Z0DuvAULCj2Zkoo9IpNLPti66LwOOYm+NLeSPV1KhjsSUgq9vS9/PC7ikj31ekw+bgsJvSE8k74OEJs9PMKlOyyaIL1MPbi7GUWbPZyJTT4WECG9m+6evs59oz1J8so+8X06vf1/Hb8lurM9x8ROPlfkbL2qqay+df+7PTdO+ztNQoS907SBveBPvD0M8j6+ZtqGvXtLVT6YrLQ9SR0cPIWkfL3k1NW9ghC1PfAWPb4UmYK9yzosPjuArT2e88G+5Gp3vV5U4T4Z/J09VEc7vmVdU73OGAQ+Xn6WPQMlwb4MzEi97FfPPsQKhz0Kzzm+Q58nvVgmxz0rOH89oX7AvvinH72v48A+lmtgPe+jOL44ywC9+YCTPSemUT0NyHk8j8n1vCTxa77f5VI93b03vt/EDb26r1c92DJEPbmCv76OdAm9Ph+rPpGOJT2Cvja+yybcvA0s/zz47xY9eo6LPFAM17wTL4q+O1UYPYL3Nb4rogG9C+hrPI3GCT11pb6+NXQAvUEHmD5bi9Y8cAs1vkJC0LzTb7O7xJO5PMU7vr7vJ9G8auOOPq5neTzu+BC/hm6jvCdIED+HXH87utlCv8YuDrwslVk/e5E5vIjeEL+7Ugg85fUNPx+Aubx75b2+LgSfPGBqhz5vRPa870Y0vmtZyjxwV7S8S44JvcI3vr4SvsY8AYWOPp39J71c/jS+RVn0PGZz17tdeDa9Y+KSPH5F8zyGU5S+VgA1vWPdNb6bzsM8Cd9HPO6MQ72ipYw8R87FPKuyi77fJEK9cZE2vjkamTwFEuA83r9QvXaohzx3lZ086MyEvpVkT72THze+6CxmPIMFIT3zCl692JC/vqAOczzXO6w+fLF8vcyLN760pLA8TFpGPUGwhb2R1L++1ZO4PDkXsj7wCJW99jA4vgaR8Tz0V389D2ecvUI1wL6+x/s8nXW6PnvHq71uEjm+QLkbPT2Xpj2eLrO9+3diPCRjIj2a4ku+rZ2yvSU1Or6SExI9LLXYPXEQur3H5FA8qL4aPe+dM77Airm9/0c7vhpgDD2aCwQ+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 0, 0, 0, 1, 1, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 0, 0, 0, 1, 1, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAMkPwb3Q+F2+RTGwvO2zfD7H8Mm9/9+uvJDCh7wPfFe9ntDKvU0vMj4gYZC8S3GzvgGww70o36q8D83JvOXXgb25isS9bswyPkAw1Ly7Pbq+02O9vT71pLyQ5Ae9JIGivfg2vr0D7Fu+nGQOvdyMTz74Ase9KQKdvPWT+7wFWs698cvHvX/3Wr4FCwa9Q3M6PimO0L1dlpW8CkHuvKJI972iTdG9hBBavrQEAb21hiY+nQbavXeAjrx/ZOe8Cy8PvgS92r0iMVm+TU3+vGFDEz4QbeO9j0PQvme95ryyY9o+Thb0vYRTWL7w2qC8+B0APn29/L2G6YG8Q1uMvPDdMb7GY/294MlXvrLQqLzcgug9uQIDvnqkz77dNpa8153MPvxQC779N1e+WnkpvAhUzz3nng++ZG/PvjBNCLx0Acg+CusXvgryVr4YxAS6mkLDPY83HL5wIXK8zoy3OqY9Sr4LhRy+YvVWvt4XJ7tJ1sM9odEgvnPccby2Bye6hJxKvgYfIb6psjg+0IyWuyJD+75fbR2+JqVwvDYVbLyyTUy+YbodvpShVr7gupa8eW21PUoFIr7L+mi8PTeIvBrhVr7YTyK+YB5WvrKYqrzN0Z49IZgmvnM7YLwS5J28SfNivuPfJr4tiFW+9TPCvKnvhD0rJSu+iH3OvmyRt7zfNbM+oWczvmHbVL4CcXy8nEJOPXWpN77/RE28zfBrvEQUfb4l6ze+3i47PoZ2nry6WQu/xCw0vrdoRby1pfe87P+DvvBrNL4dsFO+kPEQvTFQzjzHpzi+ch02vGbhDr32kI6+DuI4vgGpUr7qsCW9J5hGO6MYPb6F7My+XXElvYi7kD4PS0W+P3pRvh5JDr33A7i8lXtJvr/jErwyIBC9adymvpaqSb5xdFC+09Iqvb1YNr3g1U2+tc/LvnB4Lr1+WnA+5/xVvtexF7//PRu9zH8DP5wfYr5zOcu+KlPivGpPVj6hQGq+pnUXvwUJwLxfe/w+hV52vvzZyr5me168aslFPrh7fr6xs02+tTAfvOGM1L11TIG+cq7KvsQyQbzrRj4+MFqFviNjTb5GTwS8UG7ivfpnh75Oicq+4okovCzfNz73dIu+ay4Xvzhm27t2HvA+D4GRvoBpyr7z5y87nGAyPmmNlb7GKhe/Ux3KO6t57z5cmZu+BHrKvkpSfjySPTU+C6afvhphTb7CKJw8m9LivdCzob4GgLu7bgOKPP1yy77QwqG+oudNvr7RETwimsu97tGjvhZqx7tnfOI7Sk/HvuLho77nIk6+SAtlurBgwb2X8aW+1fbKvtIFNbtvvEo+xACqvkoVTr549Zw6crnDvVcQrL7q88q+ACM7ulQ7Sj51H7C+GxdOvo0SVDuDacO9DS+yvqP4yr5zBK46fgtLPkM+tr42KE6+/XOtO253wL0GTri+BQXLvkK6XzvrLU0+fF28vtZITr63LfM7r9i6vZNtvr5IGcu+QmO3OxasUD5wfcK+nHlOvg54HjyicbK9BI7EviEI2rv/6gE86+TAvnWfxL6Xu06+lubOOT4Qp72xsMa+yeDdu7QdorpGjr++csLGvna4Tr413A68jZunvabTyL67Msu+a60pvM4SVT4G5My+/3BOvuH8yrvZ67O9g/TOvhx60bsESAK809TDvkUFz74IPE6+Bp1/vCsRvb07FdG+SOfKvpvujryHF0g+GSTVvoRWF7+u1V28PQ73Pswx275Orcq+W+B+u9sQPj6BP9++B35NvkB5OblUyt29kE3hvhjqtbu1iRm7b1TNvh5c4b6idE2+qsspvFJr370VauO+3QGwu+qKTbyNX8++KnjjvuMeTb55Iam8fDvuvUaF5b4VnaC7eTC8vLO11L4fkuW+qntMvtQgAL3ONwW+mZ3nvlboyb4iyQq9yjccPl2n677Wxha/lJP8vH9k3j5Sr/G+4JxIvy5ptbw+mDc/mLX5vtp7er8apP+7r1uAP0bdAb+QgUi/vcZIPDUqNT/d3wW/UZkWv2xV2DzDdtY+7uIIv+99yb4new49Kd0JPsDmCr/ZtEu+m4IZPbNmFr5/6wu/GAXKvmR6DT0QKyE+qvANvyDCTL4dXxo9u1v+vcL2Dr8O2bC7fDIQPUs8z77V/Q6/9NJNvjwU3jyqOs+9SQUQvwUCy74vgM08H7lMPvwMEr9clk6+pEHuPFSMrb1rFRO/O2fLvmBf4DyuLl4+IR4Vv85pT7779QE96xmJvZ4nFr+g1cu+IvT4PG87cT5vMRi/ZVJQvoDGDz3PAkK9FjwZv9RPzL4q5Qs9RSqDPiBHG7+yVVG+sOEgPTQw0bwTUxy/fg4ivCrKHj3rc5y+CmAcv/55Ur7cwQU9ovlwunNtHb+tqTK8lK4FPXb1kL6+ex2/hiU8PiL63DzcuRC/6oocv9nGwT45WoA8SXJZv9iaGr+QiDs+OQOtutlEDb/Nqhm/5QlEvGhzSry05YS+KK/0u9vKQr4d6wQ9tOOUPvmsOLySwZA7oL0cPWSuCzxmOje86b5LPmtwHT3wLou+Fw/sux2xWTt4Kwc9p/wFPczh6bvJtEo+e9kJPU1df77VTFC7rkUbOzzX6jzsD1w91zFNu2DMST6spPM8akprvsRtVDrjk8g+Kv/NPHNPA7+CpQ08Sv9IPo7qczyPh1m+M/dNPChIejqJTi48kOWtPUtHTjzvn0g+WSFKPG9NUb48PYc8mDUkOkonBzw2urw9gleHPANTSD6QWSU8vKpKvspkpzxzAr85LP7IO9GQyD0SdKc8XhZIPkKWBDy2b0W+pnfHPOrixz6K0Io7WkL3viy3Az0s6Ec+W62xu2xyQb5EtRM9bEU6Odq9FrzsAdE9/bgTPZp8R77Pmeq7spzIPoHDAz2Vx9c5F3YxOrFtxj0izAM9ZTdIPgtcKzs4Rki+kdATPc79xz6s+6m6YpH5vjfQMz1eMkg+qvg0vOnYR74+1EM9bjAOOijsdLzQisA9nt9DPd4GR76gHVa8LY7DPozzMz2ULnc6q+uxu1Ztrj1TBzQ9htlIPjM1dLsuPla+uhhEPb5ayD4UnAG8sskAvz8nZD3mBEk+pTqTvPoBWr4ePHQ9i4rIPj0ctry63QK/JCmKPfGaST6s7gS9xQNnvpQ5kj0PlRI722kXvSAgaD0IUZI9zvNEvl/FEr1PyKw+PXCKPQ6/VjtnQO68BQsKPZmSij0ik0s+2LrovA45ib40t5I9XUqGOxJSCr29L388LuKSPfV6TD5uCwm9ITyTvg4Qmz08wqU7LJogvUw9uLsZRZs9nIlNPhYQIb2b7p6+zn2jPUnyyj7xfTq9/X8dvyW6sz3HxE4+V+RsvaqprL51/7s9N077O01ChL3TtIG94E+8PQzyPr5m2oa9e0tVPpistD1JHRw8haR8veTU1b2CELU98BY9vhSZgr3LOiw+O4CtPZ7zwb7kane9XlThPhn8nT1URzu+ZV1Tvc4YBD5efpY9AyXBvgzMSL3sV88+xAqHPQrPOb5Dnye9WCbHPSs4fz2hfsC++Kcfva/jwD6Wa2A976M4vjjLAL35gJM9J6ZRPQ3IeTyPyfW8JPFrvt/lUj3dvTe+38QNvbqvVz3YMkQ9uYK/vo50Cb0+H6s+kY4lPYK+Nr7LJty8DSz/PPjvFj16jos8UAzXvBMvir47VRg9gvc1viuiAb0L6Gs8jcYJPXWlvr41dAC9QQeYPluL1jxwCzW+QkLQvNNvs7vEk7k8xTu+vu8n0bxq444+rmd5PO74EL+GbqO8J0gQP4dcfzu62UK/xi4OvCyVWT97kTm8iN4Qv7tSCDzl9Q0/H4C5vHvlvb4uBJ88YGqHPm9E9rzvRjS+a1nKPHBXtLxLjgm9wje+vhK+xjwBhY4+nf0nvVz+NL5FWfQ8ZnPXu114Nr1j4pI8fkXzPIZTlL5WADW9Y901vpvOwzwJ30c87oxDvaKljDxHzsU8q7KLvt8kQr1xkTa+ORqZPAUS4Dzev1C9dqiHPHeVnTzozIS+lWRPvZMfN77oLGY8gwUhPfMKXr3YkL++oA5zPNc7rD58sXy9zIs3vrSksDxMWkY9QbCFvZHUv77Vk7g8OReyPvAIlb32MDi+BpHxPPRXfz0PZ5y9QjXAvr7H+zyddbo+e8ervW4SOb5AuRs9PZemPZ4us737d2I8JGMiPZriS76tnbK9JTU6vpITEj0stdg9cRC6vcfkUDyovho9750zvsCKub3/Rzu+GmAMPZoLBD6CCMG9H8nBvmTwFj1LT90+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.5256412625312805, 0.8248687386512756, 0.45793473720550537, 0.8605185151100159, 0.41744932532310486, 0.8761753439903259, 0.6323373913764954, 0.7565934658050537, 0.6648380160331726, 0.7313669919967651, 0.6943894028663635, 0.7043389678001404, 0.72172611951828, 0.3250804841518402, 0.8962647914886475, 0.6675595641136169, 0.7484984993934631, 0.3536776900291443, 0.8895854353904724, 0.35201022028923035, 0.8918532133102417, 0.6602283716201782, 0.7452659606933594, 0.6582269072532654, 0.2524130642414093, 0.917079508304596, 0.7705786228179932, 0.6142973899841309, 0.7849718332290649, 0.5895155668258667, 0.8005152344703674, 0.43997082114219666, 0.8593524694442749, 0.5531554222106934, 0.182420551776886, 0.9353232979774475, 0.8434819579124451, 0.45985931158065796, 0.8621677160263062, 0.5924195647239685, 0.777825653553009, 0.37582361698150635, 0.8886260986328125, 0.6741248965263367, 0.28747785091400146, 0.9027026295661926, 0.2959340810775757, 0.9025185704231262, 0.7082251310348511, 0.6970930099487305, 0.7032329440116882, 0.7009216547012329, 0.30063489079475403, 0.9033343195915222, 0.28241822123527527, 0.9098028540611267, 0.7454308271408081, 0.3646588921546936, 0.8789021968841553, 0.36683428287506104, 0.8802663087844849, 0.6412331461906433, 0.7515718936920166, 0.6350235342979431, 0.755964994430542, 0.6265958547592163, 0.7617502808570862, 0.6158391833305359, 0.7688626646995544, 0.6025813221931458, 0.7772388458251953, 0.41340601444244385, 0.8601136207580566, 0.4070168733596802, 0.8646483421325684, 0.6092738509178162, 0.7653318643569946, 0.38926249742507935, 0.8717235326766968, 0.6342201828956604, 0.2572495937347412, 0.9108331799507141, 0.7513124942779541, 0.37811678647994995, 0.8719948530197144, 0.36002182960510254, 0.880423367023468, 0.3321553170681, 0.8915058970451355, 0.7042906284332275, 0.3347679078578949, 0.11618360131978989, 0.05332870036363602, 0.9659713506698608, 0.9502090811729431, 0.9039193987846375, 0.7499552369117737, 0.5826046466827393, 0.7750289440155029, 0.4584871828556061, 0.81419837474823, 0.5211343765258789, 0.8059515357017517, 0.4875529408454895, 0.8215456604957581, 0.45166710019111633, 0.8367767930030823, 0.4134962558746338, 0.8516069650650024, 0.6266768574714661, 0.687355101108551, 0.6514890193939209, 0.334202378988266, 0.14039455354213715, 0.9345529675483704, 0.8740472197532654, 0.36252230405807495, 0.896816074848175, 0.6860520839691162, 0.7085597515106201, 0.7153002619743347, 0.6816492080688477, 0.7390608787536621, 0.34369027614593506, 0.8874228000640869, 0.658348560333252, 0.752244770526886, 0.6502488255500793, 0.7577680349349976, 0.6442596316337585, 0.7616212964057922, 0.3597574830055237, 0.887228786945343, 0.663818895816803, 0.25934192538261414, 0.9174041748046875, 0.74948650598526, 0.33978766202926636, 0.8942300081253052, 0.6890450119972229, 0.2840145230293274, 0.9105319380760193, 0.719329297542572, 0.3007564842700958, 0.9059088230133057, 0.2674787640571594, 0.916103720664978, 0.7743216156959534, 0.40157946944236755, 0.8711622953414917, 0.5707144737243652, 0.8170074820518494, 0.5215039849281311, 0.8404732346534729, 0.4640161395072937, 0.13695785403251648, 0.9464011192321777, 0.8910295367240906, 0.6980783939361572, 0.6736809611320496, 0.7521014213562012, 0.3941214680671692, 0.8625127077102661, 0.4404342472553253, 0.8474799394607544, 0.47648704051971436, 0.8352437615394592, 0.49761420488357544, 0.8413048982620239, 0.5463384985923767, 0.8017957210540771, 0.43185946345329285, 0.8646782040596008, 0.6081406474113464, 0.7633503675460815, 0.6285426020622253, 0.2485833466053009, 0.0842682495713234, 0.9590526223182678, 0.9267340898513794, 0.808161735534668, 0.5388147830963135, 0.8305384516716003, 0.5077704787254333, 0.825065553188324, 0.5343879461288452, 0.8146411776542664, 0.5551165342330933, 0.8061191439628601, 0.4295116066932678, 0.8729360103607178, 0.3912937045097351, 0.886841893196106, 0.3450402021408081, 0.9012870788574219, 0.70702064037323, 0.6860236525535583, 0.7417863011360168, 0.6472751498222351, 0.22767147421836853], "advantages": [2.3915624618530273, 0.9128128290176392, 1.9939098358154297, 4.373640060424805, 1.7288860082626343, 4.224225997924805, 1.5726784467697144, -0.17270918190479279, 1.3748666048049927, -0.490682989358902, 1.202043056488037, -0.7665486931800842, 1.056379795074463, -0.9943071603775024, -1.921170711517334, -1.345302700996399, 0.5937110781669617, -1.5748848915100098, -2.789512872695923, -1.9502424001693726, -3.2849485874176025, -2.400507926940918, -0.5713468194007874, -2.7596092224121094, -0.9667003154754639, 1.0491036176681519, -1.144873857498169, -3.192415475845337, -1.4253991842269897, -3.391909122467041, -1.6913598775863647, -3.5597715377807617, -5.203500747680664, -3.922785520553589, -2.4654603004455566, -0.8183251023292542, -2.5049383640289307, -4.026360034942627, -2.6796934604644775, -4.064598560333252, -5.498188018798828, -4.320070266723633, -3.22491455078125, -4.361584663391113, -5.5552167892456055, -6.870035648345947, -5.913549423217773, -7.11842679977417, -6.396326541900635, -5.94174337387085, -6.774332046508789, -6.455793857574463, -7.192698001861572, -8.07259464263916, -7.874381065368652, -8.623242378234863, -8.654796600341797, -8.83682918548584, -8.763457298278809, -9.346219062805176, -9.320013999938965, -9.839715957641602, -10.233786582946777, -10.577777862548828, -10.916341781616211, -11.346333503723145, -11.63433837890625, -12.14673137664795, -12.388915061950684, -12.980792999267578, -13.181760787963867, -13.850817680358887, -14.28651237487793, -14.506301879882812, -14.988292694091797, -15.159668922424316, -15.276196479797363, -16.052671432495117, -16.63519287109375, -16.730117797851562, -16.782697677612305, -17.1455135345459, -17.86905860900879, -18.816301345825195, -19.566633224487305, -19.58060646057129, -20.33991050720215, -20.35165023803711, -21.093889236450195, -21.123939514160156, -21.0780029296875, -21.398609161376953, -22.184791564941406, -23.385377883911133, -23.519378662109375, -24.169841766357422, -25.254911422729492, -26.55855941772461, -26.465360641479492, -27.809185028076172, -28.995939254760742, -28.951562881469727, -28.935897827148438, -30.32781982421875, -30.33701515197754, -31.768753051757812, -31.799833297729492, -33.27592086791992, -33.323875427246094, -34.850650787353516, -36.28150939941406, -36.345703125, -37.7971076965332, -38.577781677246094, -37.94478225708008, -39.647216796875, 16.350248336791992, 16.597885131835938, 16.050472259521484, 16.253828048706055, 15.76790714263916, 15.910856246948242, 15.485125541687012, 15.571906089782715, 16.475852966308594, 15.268145561218262, 14.935413360595703, 14.946843147277832, 14.645359992980957, 14.624359130859375, 14.35145378112793, 14.300219535827637, 15.055255889892578, 14.010693550109863, 13.775786399841309, 14.385168075561523, 13.436009407043457, 13.329998016357422, 14.034575462341309, 13.035112380981445, 12.832073211669922, 13.481013298034668, 12.468961715698242, 12.332097053527832, 13.03783893585205, 12.03313159942627, 12.801027297973633, 11.753522872924805, 11.526556968688965, 12.149247169494629, 11.119643211364746, 11.061503410339355, 10.75860595703125, 10.74571704864502, 10.395262718200684, 10.44758415222168, 11.762396812438965, 10.271475791931152, 9.765162467956543, 10.128925323486328, 9.379042625427246, 9.62575912475586, 10.951822280883789, 9.047138214111328, 10.22319221496582, 8.495255470275879, 9.526464462280273, 7.965987205505371, 7.8194427490234375, 7.537895679473877, 8.315401077270508, 7.053107261657715, 7.229898452758789, 6.676485061645508, 7.1954145431518555, 6.253802299499512, 6.625970363616943, 8.238375663757324, 10.621915817260742, 7.401065349578857, 5.3030877113342285, 4.76797342300415, 4.780954837799072, 4.279568672180176, 5.106289386749268, 3.8309929370880127, 4.702719211578369, 3.385199546813965, 4.311069488525391, 2.9430389404296875, 2.835113763809204, 2.4244446754455566, 2.2777273654937744, 1.8650037050247192, 1.7062467336654663, 1.2580653429031372, 2.148122549057007, 0.6627457141876221, 1.50052011013031, 0.04488372802734375]} +{"type": "SampleBatch", "eps_id": [89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 89905220, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAIIIwb0fycG+ZPAWPUtP3T49idC96VE8vj1ZOj13CRs+oxHYvV5hwr5lwEY9F4rqPoue573+sD2+IUdsPcltOT78NO+9TowTPLccez3uLL69jtbuvRATUD5RgXM9Pi67vt+D5r28qu47Zo5VPZJTYL1/N+a9pHVOPtkRUT2HPqm+WvXdvXlnvjua/TU9mwy2vGy43b0cE00+jis0PVzjmb51hNW9sgWVO0yMGz1Bh7k7xVTVvRivQ74JAxw94cKePpMo3b0A22I75mk1PT7i8jxHBN294KtKPq7XNz0NtX6+7OjUvSZlDztGdyM9mZNsPfvR1L3Yb0k+jDIoPaJmY75Ew8y9XjeHOmABFj0Rjao9c7jMvfxNSD7S0xw9nl9KvlS1xL3R0R+4NqMMPTPk2j27tcS9gFRIvqhkFT0gBdI+HLnMvf1riboU/zY9A1UEPhrEzL03DEY+PZVBPSqbGL4Y2MS9aqYau91fNT01+iE+1/DEvSq2RD4pVUI9uTD2vYMSvb1DV3C7LHw4PYCLPz73OL29bVxDPgPPRz0Xi7q9eGi1vTr2o7vPWEA9KsddPvCctb1q9kE+0xZSPWePeb3D2q29HH3EPhMZTT3giqy+rCKevRl7QD68fTE9czDtvKtvlr25z8M+iB4vPRaFnb5zxYa9wTg/PoTqFT1INOy6rj5+vf+bEby5xBU9ArqaPg/5fr3SJj4+U4YuPb49rjzBwm+9DqfCPmJEMD157YO+yZ1QvSjmPD6gKBs9dLVFPSKBQb1MEMI+5RwfPabMbb5JdCK96tUSP8QWDD0bLQO/Bu/mvMqoRD+fOcQ8YUtLv7ckUrxHoBI/uDsEPEjt/L5EsrO6u0rBPpMe7bqVjUu+VH3KOzXOOj7Hjb27fiO/PcUFITzvWME+s2OAu0kATr7nYY48kPI6PnYdArw34Lg9RkusPBtvwT7lEcm7AdNRvmMx6jxPJzs+v60nvP7Jrz2ZEQQ9to3BPm6NC7zlGVe+jQkjPZFtOz6EYlC8UK+jPRQIMj1mtcE+/DE2vDPyXb5iBlE9hNgSP883fbxXTgO//gGAPfrmwT4VpdK8dItmvhyFjz2sYjw+Moj3vIX6cj0uDpc9emcvvBjQ7bx8RK8+652WPaFAPT4uurW8vmYmPd0vnj0ZsMI+OxKvvCeYhL4Tw609z+M9PlyA2by2Ttw8i1u1PX0Jwz5jGNW860+Mvuf1xD22qD4+Yf8AvWMRKTxBlsw9rnTDPvkmAL1KkpW+MDncPdWTPz5rFRi9lkobvPHi4z0W9MM+MdwYvY2VoL4RkPM9japAPriNMr33/Qa9+ET7PZKKxD7hQDW9bpqtvhF/BT4880E+qwdRvdl4eL0XYAk+fjvFPtj/Vb2A7by+wEMRPll1Qz5TOnS9MvO+vYAsFT4C/lq7p917vf5fOD77GhU++TlFPqcdbb1rhwa+xwwZPivqxj7O4He9gSzivqoBIT7c80Y+cgiOvVLJLL5N/CQ+sExsOcfxlL1ub889fP0kPogFST61y5C9SoxavrcCKT505Mg+pImZvUsNB7/aCzE+MiNLPl4lr72xz4S+6xs1PgrYlTtXxbm9zjaoO+QzNT6orUG+g4+5veomiT5CVDE+zFXqO6SWrr0szlS9wXkxPokhP75st7C94sNZPimnLT6nJR08ggGovZr52L1z2S0+7688vmtYrL36miM+XxMqPrndQzwZzaW9uQYivgxSKj5bSTq+QEisvZXj3D1DmCY+UWHAvkzdp71D574+SuYePvnREb+Yl5i9H3wjP+Q7Ez6wP7++g958vUmYpT6AlQs+xtk1vr9fYr3Vf0M8bfIHPk7BlTyCZWG9TmeYvkVSCD5lPTS+9cd5vbQGu7xxtwQ+DTujPL6me73hEKu+6R8FPvtzMr7Ngou9hZN7vTuOAT7FmLy+1gaOvXAIVj7/BfQ9pvoPvyR3hb0TRfU+nfzcPQGeu74HsGO90owqPjf6zT1lqS6+KQtWvRmSEr6t/cY9udS6vu7EYb0OzAc+YQu4PT4YLb7O51a9tS41vuQesT3JC7q+bGZlvaU6yj2rPKI9UYMrvphPXb1tIli+YGCbPTA+ub4Imm69AEODPZqOjD3WWA6/6Vlpvbb5rD4ikGs9t2a4vtasTb0WWeM8Dw9OPWNMKL7TZku93oiPvk6YQD0wxwA9AV5ivW+YFr+mK0M9a8YmvmRHib2cgaC+GdQ1PRnFtr6PHpa9q8wuvdeVGD1vDA2/DN6XvQ9vZz5J5tY8pLU+vyqcjr1sI/0+3rA5PIiGDL/It3S9n/Q4PvwuujnUPj6/5OtlvSJi6D4msm28XRoMv3m9QL3bfhM+jIPQvLbfPb/D8DS9itDXPj4EJb0RxQu//2gSvZ8B7D0vvlG92pU9v0v4CL0v+co+gjSHvVGEC78B/dC81Ea/PR6Hnb1FXz2/qK/BvCF/wT7N07u9SFYLv2vEg7yVf589scz6PHd4tzyX/9U8v/UfPQ94/jxnWF4+lGXcPGMpe76pBRE9FUOxPP81tDxBcGQ9dMsSPfehXT41Wb08fWhrvn2GJD28i9I+466XPMPjA79yNkY9+ScbPyqMBjznTk2/19x3PUpa0j47PwC8FLkBv3HCjD2sA10+bSWTvJG4Xb6hmZU9BNGrPB6ftrw5QpA9jnWWPRGaXT6zFKu8qrFqvsJSnz3d3rA8xKHQvCHGaD0nNaA9KUdePilSx7wXoXm+RxmpPR6ztjz5Qu+8enkoPSIDqj34Dl8+y4XovC5whb5B77I9tXC9PIecCb24Nrw8vOGzPYqdL76zuge9JRicPm/brD27QMU8J4LdvGpS/Drr1609n70uvmkx3bzDaJI+k9qmPb2UyzyRV668E7h3vCjfpz3ICS6+utGwvM+iij4B6aA97JnQPKt0hLwsm+q8BPShPQ99Lb7aJYm8eY6EPn4Dmz2qdNQ8o3U9vOrPH71wE5w9gBMtvpc+SrzT/n8+IyeVPXhB1zzlpvC7vrA+vao6lj2j0WI+yZQHvCffrr5KTZ89JV/VPsN/d7wbByO/JF+wPVAoYz5ZFuS8Mqiyvjx1uT1MJ94896AOvRF1hb2Xkbo9LxFkPpH3E729vLy+/7DDPYdz5jw/KjK9pk2zvfnXxD2hOGU+UFY5vU6Kyb4zA849ZdLwPGeVWb38puy9dDfPPRfiKL64DGO9Lr8jPhZ2yD1NgP08MPNVvbFTGb6Susk9kTdoPlI3Yr3otuq+eQTTPXgEBT2d4oO98w48vv9Y1D168Gk+VWiLvffq/b6JtN09/LkMPZG4n73Y3Wa+zBzfPbUIbD6k9Ki90KsKv8mN6D3nCRY9nSS/vQ5cjb7iDeo9R9Agvqhzyr1fYWm8J5/jPXiEs74FCcu9CpV7PqJC1T0L7x2+0vjAvXnFnL1l8c49Gh6yvn0bxL0MbD0+i7HAPZ8rG77Oh7y9xZMLvpp8uj1KwbC+FB3CvWv/AD6oWKw9YvMJvyT0vL0RTsY+MkaWPayJO7/bFq29U0gmP0yJcD0VUAm/7HuSvZe4qT6hmEQ9wT2uvgnohL06Wow8tbcoPfzJE75jNIS9aDGXvv3kHD37S62+0UyQvTDEwbzBKgE941IIv9dEkb2+5Xs+OBarPI//Ob9pMYe97fYDPwsx0Dtk0we/SChkvayeTz64hYu7Yl2rvjqMU724Edu9Q28zvE1vB79/T1y9MAotPkRlsLxLmKq+pHdOvb2ID75m/Oa8Dw4HvzjzWb1rdgs+5rUevRXNOL8Hy069ssDSPs7YWb0jrQa/mBItvenk0z2+eIK9WXc4v8yYJL3y2cM+e/yfvf9fBr+6QgW9XJuePXp8tb36Mzi/MNX9vAkpuD5v9dK9hCQGv8bmwrwrGms962vovScBOL9Uf7m8nFmvPl/uAr4E+QW/p2KBvFwULz0ipg2+yu2nvq3DdLygf4K+uV0UvjncBb9HJKS8UWMHPS4TH76FrKe+6LmevNUhiL4pyCW+TrcFv9pJyrwY+Kg8qnowvgyTN7+76Ma8mV+cPkUqP75Xb2m/n96UvM8gFz8E11G+eGs3vyGX0LsXgJU+dINgvv5Yab/X1gm6TCkVP2ouc74AZTe/wE82PFpilD4r7YC+AHwFv4KjijzLBy06DESGviZ/N78xv4o8leaYPg6bjb65mwW/0qy7PO/NOTw085K+RKQ3v3qIvTy8UJ8+sUuavtnGBb+Yg/A8auPTPJGln74h1Te/dsD0PLXDpz4DAKe+UP4Fv903FT3CezY9lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 1], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAD2J0L3pUTy+PVk6PXcJGz6jEdi9XmHCvmXARj0Xiuo+i57nvf6wPb4hR2w9yW05Pvw0771OjBM8txx7Pe4svr2O1u69EBNQPlGBcz0+Lru+34Pmvbyq7jtmjlU9klNgvX835r2kdU4+2RFRPYc+qb5a9d29eWe+O5r9NT2bDLa8bLjdvRwTTT6OKzQ9XOOZvnWE1b2yBZU7TIwbPUGHuTvFVNW9GK9DvgkDHD3hwp4+kyjdvQDbYjvmaTU9PuLyPEcE3b3gq0o+rtc3PQ21fr7s6NS9JmUPO0Z3Iz2Zk2w9+9HUvdhvST6MMig9omZjvkTDzL1eN4c6YAEWPRGNqj1zuMy9/E1IPtLTHD2eX0q+VLXEvdHRH7g2oww9M+TaPbu1xL2AVEi+qGQVPSAF0j4cucy9/WuJuhT/Nj0DVQQ+GsTMvTcMRj49lUE9KpsYvhjYxL1qphq73V81PTX6IT7X8MS9KrZEPilVQj25MPa9gxK9vUNXcLssfDg9gIs/Pvc4vb1tXEM+A89HPReLur14aLW9Ovaju89YQD0qx10+8Jy1vWr2QT7TFlI9Z495vcParb0cfcQ+ExlNPeCKrL6sIp69GXtAPrx9MT1zMO28q2+WvbnPwz6IHi89FoWdvnPFhr3BOD8+hOoVPUg07LquPn69/5sRvLnEFT0Cupo+D/l+vdImPj5Thi49vj2uPMHCb70Op8I+YkQwPXntg77JnVC9KOY8PqAoGz10tUU9IoFBvUwQwj7lHB89psxtvkl0Ir3q1RI/xBYMPRstA78G7+a8yqhEP585xDxhS0u/tyRSvEegEj+4OwQ8SO38vkSys7q7SsE+kx7tupWNS75Ufco7Nc46PseNvbt+I789xQUhPO9YwT6zY4C7SQBOvudhjjyQ8jo+dh0CvDfguD1GS6w8G2/BPuURybsB01G+YzHqPE8nOz6/rSe8/smvPZkRBD22jcE+bo0LvOUZV76NCSM9kW07PoRiULxQr6M9FAgyPWa1wT78MTa8M/JdvmIGUT2E2BI/zzd9vFdOA7/+AYA9+ubBPhWl0rx0i2a+HIWPPaxiPD4yiPe8hfpyPS4Olz16Zy+8GNDtvHxErz7rnZY9oUA9Pi66tby+ZiY93S+ePRmwwj47Eq+8J5iEvhPDrT3P4z0+XIDZvLZO3DyLW7U9fQnDPmMY1bzrT4y+5/XEPbaoPj5h/wC9YxEpPEGWzD2udMM++SYAvUqSlb4wOdw91ZM/PmsVGL2WShu88eLjPRb0wz4x3Bi9jZWgvhGQ8z2NqkA+uI0yvff9Br34RPs9korEPuFANb1umq2+EX8FPjzzQT6rB1G92Xh4vRdgCT5+O8U+2P9VvYDtvL7AQxE+WXVDPlM6dL0y8769gCwVPgL+Wrun3Xu9/l84PvsaFT75OUU+px1tvWuHBr7HDBk+K+rGPs7gd72BLOK+qgEhPtzzRj5yCI69Usksvk38JD6wTGw5x/GUvW5vzz18/SQ+iAVJPrXLkL1KjFq+twIpPnTkyD6kiZm9Sw0Hv9oLMT4yI0s+XiWvvbHPhL7rGzU+CtiVO1fFub3ONqg75DM1PqitQb6Dj7m96iaJPkJUMT7MVeo7pJauvSzOVL3BeTE+iSE/vmy3sL3iw1k+KactPqclHTyCAai9mvnYvXPZLT7vrzy+a1isvfqaIz5fEyo+ud1DPBnNpb25BiK+DFIqPltJOr5ASKy9lePcPUOYJj5RYcC+TN2nvUPnvj5K5h4++dERv5iXmL0ffCM/5DsTPrA/v76D3ny9SZilPoCVCz7G2TW+v19ivdV/Qzxt8gc+TsGVPIJlYb1OZ5i+RVIIPmU9NL71x3m9tAa7vHG3BD4NO6M8vqZ7veEQq77pHwU++3Myvs2Ci72Fk3u9O44BPsWYvL7WBo69cAhWPv8F9D2m+g+/JHeFvRNF9T6d/Nw9AZ67vgewY73SjCo+N/rNPWWpLr4pC1a9GZISvq39xj251Lq+7sRhvQ7MBz5hC7g9Phgtvs7nVr21LjW+5B6xPckLur5sZmW9pTrKPas8oj1Rgyu+mE9dvW0iWL5gYJs9MD65vgiabr0AQ4M9mo6MPdZYDr/pWWm9tvmsPiKQaz23Zri+1qxNvRZZ4zwPD049Y0wovtNmS73eiI++TphAPTDHAD0BXmK9b5gWv6YrQz1rxia+ZEeJvZyBoL4Z1DU9GcW2vo8elr2rzC6915UYPW8MDb8M3pe9D29nPknm1jyktT6/KpyOvWwj/T7esDk8iIYMv8i3dL2f9Dg+/C66OdQ+Pr/k62W9ImLoPiaybbxdGgy/eb1Avdt+Ez6Mg9C8tt89v8PwNL2K0Nc+PgQlvRHFC7//aBK9nwHsPS++Ub3alT2/S/gIvS/5yj6CNIe9UYQLvwH90LzURr89HoedvUVfPb+or8G8IX/BPs3Tu71IVgu/a8SDvJV/nz0MH9K9cjo9v8gDbryeG7s+D3j+PGdYXj6UZdw8Yyl7vqkFET0VQ7E8/zW0PEFwZD10yxI996FdPjVZvTx9aGu+fYYkPbyL0j7jrpc8w+MDv3I2Rj35Jxs/KowGPOdOTb/X3Hc9SlrSPjs/ALwUuQG/ccKMPawDXT5tJZO8kbhdvqGZlT0E0as8Hp+2vDlCkD2OdZY9EZpdPrMUq7yqsWq+wlKfPd3esDzEodC8IcZoPSc1oD0pR14+KVLHvBeheb5HGak9HrO2PPlC77x6eSg9IgOqPfgOXz7Lhei8LnCFvkHvsj21cL08h5wJvbg2vDy84bM9ip0vvrO6B70lGJw+b9usPbtAxTwngt28alL8OuvXrT2fvS6+aTHdvMNokj6T2qY9vZTLPJFXrrwTuHe8KN+nPcgJLr660bC8z6KKPgHpoD3smdA8q3SEvCyb6rwE9KE9D30tvtolibx5joQ+fgObPap01DyjdT286s8fvXATnD2AEy2+lz5KvNP+fz4jJ5U9eEHXPOWm8Lu+sD69qjqWPaPRYj7JlAe8J9+uvkpNnz0lX9U+w393vBsHI78kX7A9UChjPlkW5LwyqLK+PHW5PUwn3jz3oA69EXWFvZeRuj0vEWQ+kfcTvb28vL7/sMM9h3PmPD8qMr2mTbO9+dfEPaE4ZT5QVjm9TorJvjMDzj1l0vA8Z5VZvfym7L10N889F+IovrgMY70uvyM+FnbIPU2A/Tww81W9sVMZvpK6yT2RN2g+Ujdivei26r55BNM9eAQFPZ3ig73zDjy+/1jUPXrwaT5VaIu99+r9vom03T38uQw9kbifvdjdZr7MHN89tQhsPqT0qL3Qqwq/yY3oPecJFj2dJL+9DlyNvuIN6j1H0CC+qHPKvV9habwnn+M9eISzvgUJy70KlXs+okLVPQvvHb7S+MC9ecWcvWXxzj0aHrK+fRvEvQxsPT6LscA9nysbvs6HvL3Fkwu+mny6PUrBsL4UHcK9a/8APqhYrD1i8wm/JPS8vRFOxj4yRpY9rIk7v9sWrb1TSCY/TIlwPRVQCb/se5K9l7ipPqGYRD3BPa6+CeiEvTpajDy1tyg9/MkTvmM0hL1oMZe+/eQcPftLrb7RTJC9MMTBvMEqAT3jUgi/10SRvb7lez44Fqs8j/85v2kxh73t9gM/CzHQO2TTB79IKGS9rJ5PPriFi7tiXau+OoxTvbgR271DbzO8TW8Hv39PXL0wCi0+RGWwvEuYqr6kd069vYgPvmb85rwPDge/OPNZvWt2Cz7mtR69Fc04vwfLTr2ywNI+zthZvSOtBr+YEi296eTTPb54gr1Zdzi/zJgkvfLZwz57/J+9/18Gv7pCBb1cm549eny1vfozOL8w1f28CSm4Pm/10r2EJAa/xubCvCsaaz3ra+i9JwE4v1R/ubycWa8+X+4CvgT5Bb+nYoG8XBQvPSKmDb7K7ae+rcN0vKB/gr65XRS+OdwFv0ckpLxRYwc9LhMfvoWsp77ouZ681SGIvinIJb5OtwW/2knKvBj4qDyqejC+DJM3v7voxryZX5w+RSo/vldvab+f3pS8zyAXPwTXUb54aze/IZfQuxeAlT50g2C+/lhpv9fWCbpMKRU/ai5zvgBlN7/ATzY8WmKUPivtgL4AfAW/gqOKPMsHLToMRIa+Jn83vzG/ijyV5pg+DpuNvrmbBb/SrLs87805PDTzkr5EpDe/eoi9PLxQnz6xS5q+2cYFv5iD8Dxq49M8kaWfviHVN792wPQ8tcOnPgMAp75Q/gW/3TcVPcJ7Nj0bXKy+fFyovi7eGD2kF3K+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.9294587969779968, 0.18570725619792938, 0.9389818906784058, 0.8537620306015015, 0.5756825804710388, 0.7748913764953613, 0.6283503770828247, 0.7425888776779175, 0.6703974604606628, 0.7113828659057617, 0.296244740486145, 0.9111892580986023, 0.7489678859710693, 0.6216521859169006, 0.7777647376060486, 0.5808979868888855, 0.8015543222427368, 0.5408425331115723, 0.17833690345287323, 0.9369442462921143, 0.8489972949028015, 0.4280758202075958, 0.8676512837409973, 0.376761257648468, 0.8837069869041443, 0.327106237411499, 0.8976956605911255, 0.7202097177505493, 0.6347616910934448, 0.748841404914856, 0.6020268201828003, 0.2292359620332718, 0.921622097492218, 0.8006035685539246, 0.5160488486289978, 0.8193137049674988, 0.5181660056114197, 0.20087255537509918, 0.9219332337379456, 0.818394660949707, 0.5186336636543274, 0.8136024475097656, 0.5318852663040161, 0.8075593113899231, 0.5480309724807739, 0.7995930314064026, 0.5672764182090759, 0.7892532348632812, 0.41018688678741455, 0.8700136542320251, 0.6419435739517212, 0.2646118104457855, 0.9107948541641235, 0.7202782034873962, 0.6971978545188904, 0.6888979077339172, 0.7320569157600403, 0.6482886075973511, 0.7683077454566956, 0.5963469743728638, 0.8044509887695312, 0.5315272212028503, 0.8387905359268188, 0.45418357849121094, 0.8697337508201599, 0.631721019744873, 0.7413284182548523, 0.30365633964538574, 0.9114919304847717, 0.7701616287231445, 0.5772626996040344, 0.17918668687343597, 0.9383260607719421, 0.8686422109603882, 0.650863766670227, 0.696381151676178, 0.7289552092552185, 0.610662043094635, 0.7889453172683716, 0.5166381597518921, 0.8340479135513306, 0.5786466598510742, 0.254859060049057, 0.9007974863052368, 0.714186429977417, 0.3216126561164856, 0.894332230091095, 0.26670607924461365, 0.9086517095565796, 0.78337162733078, 0.4880681037902832, 0.8002294301986694, 0.4696184992790222, 0.8347337245941162, 0.4158012866973877, 0.8555151224136353, 0.36345183849334717, 0.8735378980636597, 0.6867836713790894, 0.641447126865387, 0.28130972385406494, 0.10146287083625793, 0.9494880437850952, 0.9127708673477173, 0.8083004951477051, 0.5610784888267517, 0.7359136939048767, 0.5986239910125732, 0.7132056951522827, 0.6249274015426636, 0.6977112293243408, 0.6417592167854309, 0.6897695064544678, 0.6503246426582336, 0.689414381980896, 0.6512210965156555, 0.7156968116760254, 0.6863346099853516, 0.7332778573036194, 0.33137035369873047, 0.10728571563959122, 0.9510492086410522, 0.9048927426338196, 0.7296422719955444, 0.6683666706085205, 0.7536039352416992, 0.6355584859848022, 0.7785553932189941, 0.5956146121025085, 0.8039855360984802, 0.4525139629840851, 0.8529237508773804, 0.4803846776485443, 0.8434273600578308, 0.5018799304962158, 0.8358321785926819, 0.5174911618232727, 0.8303287029266357, 0.5276968479156494, 0.8270243406295776, 0.4671238362789154, 0.14471691846847534, 0.9429692029953003, 0.8756961822509766, 0.3665030598640442, 0.8914253115653992, 0.3101690411567688, 0.9065641164779663, 0.7475004196166992, 0.6181640028953552, 0.21252992749214172, 0.9287680387496948, 0.16737596690654755, 0.9386832118034363, 0.12912197411060333, 0.9469860792160034, 0.9007706046104431, 0.7613030076026917, 0.5397393107414246, 0.8153049945831299, 0.4414641261100769, 0.8543521761894226, 0.6489368677139282, 0.3345296382904053, 0.8640318512916565, 0.6234999299049377, 0.2589537501335144, 0.9074349999427795, 0.7864785194396973, 0.5088066458702087, 0.7795531749725342, 0.4558618664741516, 0.8324680924415588, 0.41077080368995667, 0.8511573672294617, 0.6324989795684814, 0.6925164461135864, 0.655647337436676, 0.6772750616073608, 0.6708199977874756, 0.6688694357872009, 0.6791495680809021, 0.6672120094299316, 0.3187441825866699, 0.8853155970573425, 0.30418798327445984, 0.8911730051040649, 0.7137296199798584, 0.3683795928955078, 0.8657789826393127, 0.34493616223335266, 0.8797029256820679, 0.6938333511352539, 0.6563680171966553, 0.7246184945106506, 0.6215493679046631, 0.7580873370170593, 0.5766127109527588, 0.7927441000938416, 0.47962790727615356], "advantages": [0.22605036199092865, -0.42867884039878845, -0.39026838541030884, -1.1715447902679443, -0.6946521401405334, 0.8302022814750671, -1.5535041093826294, -0.13991041481494904, -2.4016759395599365, -1.101873755455017, -3.2444748878479004, -4.072220325469971, -4.134042263031006, -3.1309421062469482, -5.014195919036865, -4.158489227294922, -5.894867897033691, -5.186922550201416, -6.77764892578125, -7.112428188323975, -7.688177585601807, -7.322371482849121, -8.586888313293457, -8.391387939453125, -9.478861808776855, -9.448920249938965, -10.360526084899902, -10.490240097045898, -9.832526206970215, -11.496274948120117, -11.01651668548584, -12.474371910095215, -12.940675735473633, -13.439154624938965, -13.287638664245605, -14.382938385009766, -14.363130569458008, -13.511707305908203, -11.212190628051758, -14.58983325958252, -16.341259002685547, -17.146753311157227, -17.31726837158203, -18.075721740722656, -18.288162231445312, -19.007139205932617, -19.25596046447754, -19.941680908203125, -20.222368240356445, -19.703187942504883, -21.154199600219727, -21.80600929260254, -21.722909927368164, -22.79515266418457, -23.13188934326172, -23.75691032409668, -24.111248016357422, -24.729291915893555, -25.097023010253906, -25.714881896972656, -26.090248107910156, -26.716588973999023, -27.091075897216797, -27.737407684326172, -28.097753524780273, -28.77964210510254, -28.561779022216797, -29.934280395507812, -30.217309951782227, -31.02098846435547, -30.877267837524414, -32.234561920166016, -32.31761932373047, -33.347625732421875, -33.36613082885742, -32.347923278808594, -34.885929107666016, -34.01177978515625, -36.41779708862305, -35.70397186279297, -37.95089340209961, -37.414669036865234, -35.95866012573242, -34.20427703857422, -38.15470504760742, -41.30111312866211, -42.83456802368164, -42.96514129638672, -44.246761322021484, -44.590572357177734, -43.81086349487305, -42.177825927734375, -45.9426383972168, -48.067527770996094, -47.82322311401367, -49.5897216796875, -49.615081787109375, -50.965999603271484, -51.28434753417969, -50.43016815185547, -52.96976089477539, -53.268959045410156, -51.18247985839844, -53.687198638916016, -55.06084442138672, -55.323970794677734, -54.65584182739258, -56.81843948364258, -56.601871490478516, -58.014732360839844, -58.30796432495117, -58.95310592651367, -59.76710891723633, -59.71620178222656, -61.001075744628906, -60.4132080078125, 8.784515380859375, 8.630024909973145, 8.396060943603516, 8.196632385253906, 8.733899116516113, 10.577536582946777, 8.395715713500977, 7.446453094482422, 7.281351566314697, 7.0236921310424805, 6.848047256469727, 6.596985340118408, 6.404021739959717, 6.167863845825195, 5.949280738830566, 6.610771179199219, 5.4270195960998535, 6.049040794372559, 4.896714210510254, 5.482761383056641, 4.3576979637146, 4.911358833312988, 3.809469223022461, 4.334542274475098, 3.2517106533050537, 3.1204702854156494, 4.0510640144348145, 2.69480299949646, 2.2688372135162354, 2.237015724182129, 1.7450361251831055, 1.8114523887634277, 1.229501485824585, 1.5346274375915527, 0.6508389115333557, 1.0282247066497803, 0.1845501959323883, 0.8386350870132446, -0.1987219899892807, 0.8794696927070618, -0.4212284982204437, -0.6618558764457703, 0.18327677249908447, -1.2055025100708008, -0.6214536428451538, -1.562678575515747, -1.2707750797271729, 0.09874255955219269, 2.097134828567505, -1.05623459815979, -2.5054452419281006, -1.6038429737091064, -2.4810168743133545, -2.222179412841797, -0.9870624542236328, -2.698495626449585, -2.021012783050537, -2.715372323989868, -1.2873162031173706, -2.4002737998962402, -2.547067165374756, -1.9779797792434692, -2.6178901195526123, -1.4094657897949219, -2.4450089931488037, -0.8139810562133789, -2.099644899368286, -0.2879801094532013, 1.6387158632278442, 0.3563217520713806, 2.0343728065490723, 0.9628010392189026, -0.28174227476119995, -1.439690113067627, -0.020464925095438957, -1.1354455947875977, 0.03881257772445679, 0.8591410517692566, 0.10383941978216171, 0.6821367740631104, 0.06252273917198181, 0.4220903217792511, -0.07779884338378906, 0.08335113525390625]} +{"type": "SampleBatch", "eps_id": [219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 219733619, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAABtcrL58XKi+Lt4YPaQXcr4duq++XEMGvyGABT1i2oo9+Bi1vuHeqL78DQs9LJJbvpZ5uL6ogQa/V/ryPILNtT3v2r2+MJA4v9TCAD3eBMg+3DzFvtK6Br+cwyA9aEvdPX6gyr610Ti/qp0pPfph0z4KBdK+Cupqv+ZvSz2j4jc/kGrbvsolOb/kI4M9zyDiPnnS4r5QcQe//zqVPenmLT5oPei+ro2rvsEvnD07xcq9wqvrvj/8B7+RIZg9lO1dPkAc8b7qLzq/HQKhPawpCD/Ojvi+TooIv1fLtj2yqoc++gT+vnHirb7LpcE9nPThOqK/AL9OeRW+37fBPWn2hL72fgG/0EKvvswUtz3A7no9oT8Dv/jgCb8wl7k9tfLCPpEBBr/ekbC+uS/JPbWY8T2WxQe/AOgavrIEzj1elxG+3osIv5cDsr7YMcg9yMw4PpVTCr/ZyB2+MZbPPSKHo72MHQu/kpMhPe5QzD2Hha2+2OkKv3K1IL41b749kzuHvI23C78OdBY9HMK9PQSTjr5nhwu/UWkjvjBasj11zCs9klgMv00wtr7+EbQ9b7i4PvkqDr8k9SW+EtnCPV7Lxj1m/w6/W1ABPeXSxj1aFii+BdYOv5nAKL6vGcA9e0shPgauD78Sd+w8V43GPe5U1b0wiA+/M9BjPhVJwj1dwbq+lmQOv9AO1T5VWLM9BkcgvylDDL+YLWE+X7OZPZtDnb7uIgu/R+vCPJoejT0zqgY8vgMLv5QXXz7JdI09jx6Gvi/mCb9uzrI8BrqCPS27Uz2UyQm/ayldPg7YhD39eGG+fa4Iv7n60T5upnc9aMP7vvGUBr8aVls+NF5PPZ7zOL4xfAW/gi7RPmWSQD3z7+m+sGQDv0fpWT5TJBs9P2EZvsNNAr+JlIw8Gt8OPRlQHj5ENwK/Q9lYPlqJGz2s6wG+syEBv7sJ0D6UJBE9+o7Qvj8a/r6Fxlc+C4zfPMZg1L3d8fu+wEx5PIqOzjxyNEo++cn7vl+VOL7a6O48LPn+PoKi/b77ZcC+IkAgPbi6TD/LvQC/KZA5vpDDYT2PBAU/UasBv015VjwxKoY9eKh6PiiaAb8YMVQ+7jCQPT9m2byNigC/dXLNPqkajz3hp5e+N/n8viNnGD+/+II9BekQv5zg9r42HEo/c5JXPUaCVr8Ay+6+SfwXP+LtEj10dwe/q7bovt/Ryz7qKM88Kc9mvhyj5L5jeU8++DqqPBfpmT36j+K+4HvLPg+LtjwN91e+JH7evhLYTj4h/ZM8O7W1PZ5s3L6FMMs+goaiPCD2Sr5KXNi+0HgXPzUNgjy2GPi+N03Svrbtyj6PpMo75Gc/vjk+zr6IZBc/QUkgO/GN9L72L8i+x1VJP/ri6LunAUW/SiLAvrloFz9ITri8/E71vtwTur6ScEk/+2YDvdpfR78eBbK+55oXP7szQ71sJv6+rvSrvkK4ST+43Wu9R79NvxLjo752+xc/RNqWvaOYB7/Gzp2+S6HMPkmMrL0krIW+EbeZvvGLGD/mPbe9AUAUv/2ck74o480+OPbOvZW8ob7Zfo++z5RVPpXm271MSWi9FFyNvs8deDw9Od69pxBNPmE0jb5Yt1g+XwXWvfUm/72WCYu+TLqUPMAf270slgg+/tmKvg7JWz4aqdW9CaFDvliniL5KPa08WHzdvf4yiT3ob4i+3mcwvuK92r3fuqU+gTOKvvc3xjy8e829Ml0wuhP0ib4gay2+yoLNvSN6hD4HsIu+EJHdPKjpwr3Zt4K9IGmLvrjBZD7vhsW95RjFvoIfib6O1vM8fUvVvT9O/r170Yi+z6Ynvohh2r3bIgk+q36Kvrvetr5C5dS9IUXIPvYmjr4KmCS+ut/EvaJrij1SzI++hukRPQQbwr1Y84G+8G6PvsXRIb5ngMy9fwT3OzINkb4ELx09WzHMvTw7ob6ZqJC+rOsevmAX2b0vU2K9bz+Svl2Csr7EWtu9hgtPPmfRlb55xgq/oBLTvTUk6z52Xpu+Rfywvu5CwL0uNws+oOievintGL5dsbq9hD49vh5woL4zpK++OUPCvV5Mnz1n86O+s2QJv54Tv737na0+T3KpvrlFrr7wL7G9AaSXPJburL7elRO+1m2wvWqvmb5naK6+DgStvk+5vL3p1BK9P96xvpAXCL8zMb691NRnPtRPt76brDm/P+u0vfQP+j4ivb6+K28Hv/jpoL2xTC0++ifEvhpwqr5i+5m9WLEWvp+Qx74F4Aa/egKgvZST9z2+9cy+WVKpvuIOm71YE0i+q1jQvqlQBr+pD6O9BHOUPQ641b44Mai+mRegvUYUer4zFdm+1YcHvmkYqr3HfA+/KHDavhIGp76nDcG9ihaXvlHH3b7FFgW/8CPNveR7Cb0mGuO+fqM2v+WDzr3vVGc+XmjqvrhbBL8PQ8W9nz/Gvbiz77687DW/FzrJvX3xJz6g+va+sqYDv1qCwr2uxiG++gM7PDzhV7xA8mo8JkeivKyyNjxNGTo+hnRkPHepnb7kP3I8kIFevGAa/zvmWjK8qMxtPHXIOT4F+Pc7GSqavgGglDxnGmK8qY3KOm0AxrsuXZI8gPVVvpa2ujpjApM+1kJgPG/kzr6p2eo74osUPzOztztWFVa+QsiZPElmlD4iv7o6HBTPvidFyTyDoBY/uF/au5W2Vr7x1RQ9CWqbPirlMbyog8++tbMtPZGCGz8nWpu8NdtXvhZ3Xz0jKKg+oOO9vIQbh7zKXno9ueNaPWCXwLymiFm+gb9+PXHCuj6PZeO8wYCVvJZQjj3eE709BGPmvMEHND6qGJI9ZOo1vvqUybwYXL0+2dGKPeHd5L6g/Iy8zAAyPlEFcT2a9Qi+PANhvMJyvD5jEGY9CYrQvr3P0Lvf9Q8/n7JEPfyHLr8sup87xLG7PgfZDD2Mq7++7/xHPE6rDz9xXNw8r+8nvznxvzziP7s+h8NhPLm+tb613Ps8E38uPtvk2jv1dm29CuQLPe0Zuz5N5rQ7LXSyvrXTKT1VUS4+yRW+uhmxXb28xTc9E7rLvOL7JbvD0HM+Mbw1PUBgLj6SGRI70tVivWqvQz3zErs+dAaTOvfYsb73nWE9rVMuPnbjvru+gF69LpBvPb8buz4ufeK7ypqyvhLAhj1kgy4+LY1jvNP3br0Wu409663IvD2rdrz7cWs+N7qMPdnvLj6RUyu89yuKvZK5kz1lcrs+FW9BvIwTur57uKI9PkMvPupCnLwHkJi9LLupPcyruz5md6i8Swu/vq2+uD2w1y8+u5nlvHsvsr1Nx789Bwa8Pvja87zy28a+BtLOPQKwMD7Cvhm9MIvXvU7j1T1GX7S87F0ivTSSMz5t/NQ9ib9dvk8AFL2hGuk+uh3MPZeEq7y8aN281hAbPi9Cyz0kwDI+PpnEvOw+Gb6XaNI9CnO9PjAe3byQTea+g5DhPcl+Mz5OaBO9Gbwpvou+6D10570+efwgvb9r8L7I7/c9A5g0Phx0R72QFEK+Dyn/Pfi2k7zg+la9b/+yPfxr/j3OdVm+8NFPvdjkuT4xufU9paWHvLkTMr3OmWA9kQv1PYAYWL7ElS29H7+qPr9m7D0fNHu8/UMSvS9S4zz5xes9a/ZWvgz+D71YL54+wizjPfCcaryVXe28s8+wO5uW4j3aB1a+RHvsvEzdkz7uBto9MRFdvDUqvby7gFK8c3nZPcJROj4ZRb+8qB2gvlzt4D17MVK8yYHyvLhB4bzWZuA9pn5UvhoD97w064I+5ObXPYYzRLw8Hs28S809vVJp1z2P5js+z7XUvLmRsb5s7d49bgnCPiTEBr08xyW/THPuPe3APD6y0Du9Yhm7viIA9j1MqyW8R8BZvcBMs70blvU92VZRvk7sYL2daEA+eTbtPREmzL7Hh1G9ru/sPoLh3D25yk++3p4rvScdHj64kdQ9knrLvrL4Hr2y/t0+eUrEPdicTr6S5/a8bAQEPsIGvD0x+8q+JMjhvLnq0j60yas90cRNvtVJnrzMuOI9oY6jPXSkyr6RJoy8k2HLPoNYkz1XPE2+KyMWvCEqyz3mIos9sIWpuxhD67sxEkm+p+yKPR3/TL5J+TW80J3APX25gj2R+aC7uCcXvFH3Tr76hYI98bJMvmFiWbw6frM9uKt0PT4nyr5Yqjy8kZHAPoVTVD1L/Ba/zdeCu6dGKj/JAiQ9TwrKviKIGDxUDr4+N68DPZBpTL50FYk8y9mmPbSp5jzObZu7jm6WPPPTUr7B4uU8iaxCPhdmaTzmRfy+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 1, 1, 1, 0], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 1, 1, 1], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAB26r75cQwa/IYAFPWLaij34GLW+4d6ovvwNCz0sklu+lnm4vqiBBr9X+vI8gs21Pe/avb4wkDi/1MIAPd4EyD7cPMW+0roGv5zDID1oS909fqDKvrXROL+qnSk9+mHTPgoF0r4K6mq/5m9LPaPiNz+Qatu+yiU5v+Qjgz3PIOI+edLivlBxB7//OpU96eYtPmg96L6ujau+wS+cPTvFyr3Cq+u+P/wHv5EhmD2U7V0+QBzxvuovOr8dAqE9rCkIP86O+L5Oigi/V8u2PbKqhz76BP6+ceKtvsulwT2c9OE6or8Av055Fb7ft8E9afaEvvZ+Ab/QQq++zBS3PcDuej2hPwO/+OAJvzCXuT218sI+kQEGv96RsL65L8k9tZjxPZbFB78A6Bq+sgTOPV6XEb7eiwi/lwOyvtgxyD3IzDg+lVMKv9nIHb4xls89IoejvYwdC7+SkyE97lDMPYeFrb7Y6Qq/crUgvjVvvj2TO4e8jbcLvw50Fj0cwr09BJOOvmeHC79RaSO+MFqyPXXMKz2SWAy/TTC2vv4RtD1vuLg++SoOvyT1Jb4S2cI9XsvGPWb/Dr9bUAE95dLGPVoWKL4F1g6/mcAovq8ZwD17SyE+Bq4PvxJ37DxXjcY97lTVvTCID78z0GM+FUnCPV3Bur6WZA6/0A7VPlVYsz0GRyC/KUMMv5gtYT5fs5k9m0Odvu4iC79H68I8mh6NPTOqBjy+Awu/lBdfPsl0jT2PHoa+L+YJv27OsjwGuoI9LbtTPZTJCb9rKV0+DtiEPf14Yb59rgi/ufrRPm6mdz1ow/u+8ZQGvxpWWz40Xk89nvM4vjF8Bb+CLtE+ZZJAPfPv6b6wZAO/R+lZPlMkGz0/YRm+w00Cv4mUjDwa3w49GVAePkQ3Ar9D2Vg+WokbPazrAb6zIQG/uwnQPpQkET36jtC+Pxr+voXGVz4LjN88xmDUvd3x+77ATHk8io7OPHI0Sj75yfu+X5U4vtro7jws+f4+gqL9vvtlwL4iQCA9uLpMP8u9AL8pkDm+kMNhPY8EBT9RqwG/TXlWPDEqhj14qHo+KJoBvxgxVD7uMJA9P2bZvI2KAL91cs0+qRqPPeGnl743+fy+I2cYP7/4gj0F6RC/nOD2vjYcSj9zklc9RoJWvwDL7r5J/Bc/4u0SPXR3B7+rtui+39HLPuoozzwpz2a+HKPkvmN5Tz74Oqo8F+mZPfqP4r7ge8s+D4u2PA33V74kft6+EthOPiH9kzw7tbU9nmzcvoUwyz6ChqI8IPZKvkpc2L7QeBc/NQ2CPLYY+L43TdK+tu3KPo+kyjvkZz++OT7OvohkFz9BSSA78Y30vvYvyL7HVUk/+uLou6cBRb9KIsC+uWgXP0hOuLz8TvW+3BO6vpJwST/7ZgO92l9Hvx4Fsr7nmhc/uzNDvWwm/r6u9Ku+QrhJP7jda71Hv02/EuOjvnb7Fz9E2pa9o5gHv8bOnb5Locw+SYysvSSshb4Rt5m+8YsYP+Y9t70BQBS//ZyTvijjzT449s69lbyhvtl+j77PlFU+lebbvUxJaL0UXI2+zx14PD053r2nEE0+YTSNvli3WD5fBda99Sb/vZYJi75MupQ8wB/bvSyWCD7+2Yq+DslbPhqp1b0JoUO+WKeIvko9rTxYfN29/jKJPehviL7eZzC+4r3avd+6pT6BM4q+9zfGPLx7zb0yXTC6E/SJviBrLb7Kgs29I3qEPgewi74Qkd08qOnCvdm3gr0gaYu+uMFkPu+Gxb3lGMW+gh+Jvo7W8zx9S9W9P07+vXvRiL7Ppie+iGHavdsiCT6rfoq+u962vkLl1L0hRcg+9iaOvgqYJL6638S9omuKPVLMj76G6RE9BBvCvVjzgb7wbo++xdEhvmeAzL1/BPc7Mg2RvgQvHT1bMcy9PDuhvpmokL6s6x6+YBfZvS9TYr1vP5K+XYKyvsRa272GC08+Z9GVvnnGCr+gEtO9NSTrPnZem75F/LC+7kLAvS43Cz6g6J6+Ke0Yvl2xur2EPj2+HnCgvjOkr745Q8K9XkyfPWfzo76zZAm/nhO/vfudrT5Pcqm+uUWuvvAvsb0BpJc8lu6svt6VE77WbbC9aq+Zvmdorr4OBK2+T7m8venUEr0/3rG+kBcIvzMxvr3U1Gc+1E+3vpusOb8/67S99A/6PiK9vr4rbwe/+OmgvbFMLT76J8S+GnCqvmL7mb1YsRa+n5DHvgXgBr96AqC9lJP3Pb71zL5ZUqm+4g6bvVgTSL6rWNC+qVAGv6kPo70Ec5Q9DrjVvjgxqL6ZF6C9RhR6vjMV2b7Vhwe+aRiqvcd8D78ocNq+EganvqcNwb2KFpe+UcfdvsUWBb/wI8295HsJvSYa475+oza/5YPOve9UZz5eaOq+uFsEvw9Dxb2fP8a9uLPvvrzsNb8XOsm9ffEnPqD69r6ypgO/WoLCva7GIb68Pvy+Rzk1v/D6yL3HidM9rLI2PE0ZOj6GdGQ8d6mdvuQ/cjyQgV68YBr/O+ZaMryozG08dcg5PgX49zsZKpq+AaCUPGcaYrypjco6bQDGuy5dkjyA9VW+lra6OmMCkz7WQmA8b+TOvqnZ6jviixQ/M7O3O1YVVr5CyJk8SWaUPiK/ujocFM++J0XJPIOgFj+4X9q7lbZWvvHVFD0Japs+KuUxvKiDz761sy09kYIbPydam7w121e+FndfPSMoqD6g4728hBuHvMpeej2541o9YJfAvKaIWb6Bv349ccK6Po9l47zBgJW8llCOPd4TvT0EY+a8wQc0PqoYkj1k6jW++pTJvBhcvT7Z0Yo94d3kvqD8jLzMADI+UQVxPZr1CL48A2G8wnK8PmMQZj0JitC+vc/Qu9/1Dz+fskQ9/Icuvyy6nzvEsbs+B9kMPYyrv77v/Ec8TqsPP3Fc3Dyv7ye/OfG/POI/uz6Hw2E8ub61vrXc+zwTfy4+2+TaO/V2bb0K5As97Rm7Pk3mtDstdLK+tdMpPVVRLj7JFb66GbFdvbzFNz0Tusu84vslu8PQcz4xvDU9QGAuPpIZEjvS1WK9aq9DPfMSuz50BpM699ixvvedYT2tUy4+duO+u76AXr0ukG89vxu7Pi594rvKmrK+EsCGPWSDLj4tjWO80/duvRa7jT3rrci8Pat2vPtxaz43uow92e8uPpFTK7z3K4q9krmTPWVyuz4Vb0G8jBO6vnu4oj0+Qy8+6kKcvAeQmL0su6k9zKu7PmZ3qLxLC7++rb64PbDXLz67meW8ey+yvU3Hvz0HBrw++NrzvPLbxr4G0s49ArAwPsK+Gb0wi9e9TuPVPUZftLzsXSK9NJIzPm381D2Jv12+TwAUvaEa6T66Hcw9l4SrvLxo3bzWEBs+L0LLPSTAMj4+mcS87D4Zvpdo0j0Kc70+MB7dvJBN5r6DkOE9yX4zPk5oE70ZvCm+i77oPXTnvT55/CC9v2vwvsjv9z0DmDQ+HHRHvZAUQr4PKf89+LaTvOD6Vr1v/7I9/Gv+Pc51Wb7w0U+92OS5PjG59T2lpYe8uRMyvc6ZYD2RC/U9gBhYvsSVLb0fv6o+v2bsPR80e7z9QxK9L1LjPPnF6z1r9la+DP4PvVgvnj7CLOM98JxqvJVd7byzz7A7m5biPdoHVr5Ee+y8TN2TPu4G2j0xEV28NSq9vLuAUrxzedk9wlE6PhlFv7yoHaC+XO3gPXsxUrzJgfK8uEHhvNZm4D2mflS+GgP3vDTrgj7k5tc9hjNEvDwezbxLzT29UmnXPY/mOz7PtdS8uZGxvmzt3j1uCcI+JMQGvTzHJb9Mc+497cA8PrLQO71iGbu+IgD2PUyrJbxHwFm9wEyzvRuW9T3ZVlG+TuxgvZ1oQD55Nu09ESbMvseHUb2u7+w+guHcPbnKT77eniu9Jx0ePriR1D2Sesu+svgevbL+3T55SsQ92JxOvpLn9rxsBAQ+wga8PTH7yr4kyOG8uerSPrTJqz3RxE2+1UmevMy44j2hjqM9dKTKvpEmjLyTYcs+g1iTPVc8Tb4rIxa8ISrLPeYiiz2wham7GEPruzESSb6n7Io9Hf9Mvkn5NbzQncA9fbmCPZH5oLu4Jxe8UfdOvvqFgj3xsky+YWJZvDp+sz24q3Q9PifKvliqPLyRkcA+hVNUPUv8Fr/N14K7p0YqP8kCJD1PCsq+IogYPFQOvj43rwM9kGlMvnQViTzL2aY9tKnmPM5tm7uObpY889NSvsHi5TyJrEI+F2ZpPOZF/L5NhAI9Z62qu1njjzvJRki+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.8218080401420593, 0.5214031338691711, 0.8020656704902649, 0.4393726587295532, 0.8605348467826843, 0.37843596935272217, 0.11659073084592819, 0.9509966373443604, 0.9093955159187317, 0.7750025987625122, 0.5404054522514343, 0.17530225217342377, 0.9381769895553589, 0.8705894351005554, 0.6726978421211243, 0.6467463374137878, 0.2547138035297394, 0.9173524975776672, 0.8113523125648499, 0.43899378180503845, 0.8545355200767517, 0.6564990282058716, 0.6330134272575378, 0.7248623371124268, 0.554751455783844, 0.22242003679275513, 0.916826605796814, 0.8254109025001526, 0.3785548210144043, 0.8578136563301086, 0.694190263748169, 0.44027140736579895, 0.7742911577224731, 0.5131427645683289, 0.7728937268257141, 0.46035873889923096, 0.8010826110839844, 0.5884822010993958, 0.6715691685676575, 0.6175739765167236, 0.6515219211578369, 0.3627377152442932, 0.8444532155990601, 0.6638259887695312, 0.6092725396156311, 0.3237648606300354, 0.13984978199005127, 0.06951000541448593, 0.9559786915779114, 0.9348646998405457, 0.8877235054969788, 0.7829780578613281, 0.5870891809463501, 0.3448795974254608, 0.8308929800987244, 0.6584927439689636, 0.3929350972175598, 0.8162409663200378, 0.38085275888442993, 0.8228502869606018, 0.6297661066055298, 0.6454786062240601, 0.6234458684921265, 0.3373028337955475, 0.85651695728302, 0.29062625765800476, 0.8845455646514893, 0.22906357049942017, 0.9131240844726562, 0.8374058604240417, 0.34419429302215576, 0.8866040110588074, 0.7635060548782349, 0.49919307231903076, 0.779141902923584, 0.6160060167312622, 0.7008429765701294, 0.7241991758346558, 0.40860262513160706, 0.8400617837905884, 0.5190966129302979, 0.7886542081832886, 0.3732517957687378, 0.8963271975517273, 0.7419912219047546, 0.4069167971611023, 0.8473685383796692, 0.4897053837776184, 0.8626546263694763, 0.36847802996635437, 0.90073162317276, 0.7415105700492859, 0.40113651752471924, 0.8482184410095215, 0.5064860582351685, 0.8527261018753052, 0.6009272336959839, 0.740074872970581, 0.3205355107784271, 0.9082337617874146, 0.7596692442893982, 0.4305516481399536, 0.8323642611503601, 0.50350421667099, 0.8401395082473755, 0.42523887753486633, 0.8696585297584534, 0.34987014532089233, 0.10659728199243546, 0.9558961391448975, 0.917158305644989, 0.7981064319610596, 0.4780065417289734, 0.8402979373931885, 0.39332085847854614, 0.8721705675125122, 0.5617526769638062, 0.8077872395515442, 0.5643998980522156, 0.8086965084075928, 0.4406306743621826, 0.13279876112937927, 0.9484320282936096, 0.11702224612236023, 0.9529115557670593, 0.09847976267337799, 0.9577186703681946, 0.9198662638664246, 0.22036124765872955, 0.9326434135437012, 0.8307737708091736, 0.5276150107383728, 0.7980259656906128, 0.5837188363075256, 0.23010019958019257, 0.9141104817390442, 0.23539629578590393, 0.9159011840820312, 0.7745755910873413, 0.6059170961380005, 0.7809572219848633, 0.40703797340393066, 0.8745629191398621, 0.5971546173095703, 0.7890209555625916, 0.5786330103874207, 0.8034321665763855, 0.450589120388031, 0.8572885394096375, 0.5372458696365356, 0.8267911076545715, 0.4997854232788086, 0.8459077477455139, 0.45025214552879333, 0.8666173219680786, 0.6103419661521912, 0.23132109642028809, 0.9207680821418762, 0.7563594579696655, 0.33477190136909485, 0.8999382853507996, 0.28365153074264526, 0.9139088988304138, 0.7710989713668823, 0.4086507260799408, 0.8630256056785583, 0.4528726041316986, 0.84818035364151, 0.48957473039627075, 0.8347147107124329, 0.5189793109893799, 0.8230935335159302, 0.4582933187484741, 0.8597542643547058, 0.5858425498008728, 0.7847754955291748, 0.3888317942619324, 0.1180666983127594, 0.949183464050293, 0.9009569883346558, 0.7317249774932861, 0.3664522171020508, 0.8732879161834717, 0.39784175157546997, 0.865296483039856, 0.4176570177078247, 0.8613036870956421, 0.4257396161556244, 0.8616520166397095, 0.5778342485427856, 0.7956264019012451, 0.5651050209999084, 0.8032649159431458, 0.45069071650505066, 0.14573746919631958, 0.9439524412155151, 0.8720064759254456, 0.6130011081695557, 0.23096176981925964, 0.9187437295913696], "advantages": [18.04481315612793, 18.122724533081055, 18.112546920776367, 18.27110481262207, 18.169818878173828, 18.1669979095459, 18.140079498291016, 17.90390968322754, 17.8856201171875, 17.503557205200195, 16.959810256958008, 17.313705444335938, 17.5122013092041, 16.872713088989258, 16.0660457611084, 15.264167785644531, 15.776183128356934, 16.2425594329834, 15.203315734863281, 14.150713920593262, 14.728204727172852, 13.546517372131348, 12.5874662399292, 13.054508209228516, 11.968838691711426, 12.472203254699707, 13.175515174865723, 11.600423812866211, 10.209488868713379, 10.836267471313477, 9.275181770324707, 8.282868385314941, 8.299612045288086, 7.5688982009887695, 7.768186092376709, 6.547183990478516, 6.773626327514648, 5.4381022453308105, 5.1327643394470215, 4.48811674118042, 4.123455047607422, 3.48036789894104, 3.7074623107910156, 2.173457384109497, 1.7211731672286987, 1.04789137840271, 1.2491357326507568, 2.0297019481658936, 2.648629665374756, 0.4999796152114868, -2.145616292953491, -4.141756534576416, -4.900667667388916, -4.34340763092041, -2.423415184020996, -5.454281806945801, -7.364708423614502, -8.143362998962402, -8.946764945983887, -9.780247688293457, -10.558634757995605, -9.931825637817383, -11.977564811706543, -11.276748657226562, -9.09858512878418, -12.306340217590332, -9.893783569335938, -13.164565086364746, -10.436964988708496, -13.77761459350586, -16.078516006469727, -14.402709007263184, -16.776628494262695, -18.236873626708984, -19.182212829589844, -19.335824966430664, -20.306180953979492, -20.292177200317383, -21.272266387939453, -22.08904266357422, -22.37391471862793, -23.158414840698242, -23.394296646118164, -23.037290573120117, -24.040708541870117, -24.73094367980957, -25.5306453704834, -25.742321014404297, -25.868749618530273, -26.452571868896484, -26.53635025024414, -27.080211639404297, -27.55298614501953, -28.36653709411621, -28.488910675048828, -28.939193725585938, -29.194520950317383, -29.731197357177734, -30.112783432006836, -30.659534454345703, -30.804677963256836, -31.074373245239258, -31.809494018554688, -32.06523132324219, -32.81196212768555, -32.91990280151367, -33.72016525268555, -33.81428909301758, -34.620662689208984, -34.7878532409668, -35.28911209106445, -35.541778564453125, -35.91837692260742, -36.66606521606445, -37.08808898925781, -37.8045539855957, 16.902423858642578, 17.206523895263672, 16.650371551513672, 16.95181655883789, 16.39695930480957, 16.690797805786133, 18.187639236450195, 16.366186141967773, 17.901554107666016, 16.073989868164062, 17.70115089416504, 15.82987117767334, 15.12055778503418, 15.648594856262207, 14.837496757507324, 14.747261047363281, 15.27219009399414, 14.402836799621582, 14.8204345703125, 16.16831398010254, 14.439570426940918, 15.732061386108398, 14.11117172241211, 13.509349822998047, 13.782867431640625, 13.218619346618652, 13.371761322021484, 12.904123306274414, 13.116780281066895, 12.616473197937012, 12.812353134155273, 12.32984733581543, 12.581377983093262, 12.007198333740234, 12.177565574645996, 11.719172477722168, 11.891339302062988, 11.42978572845459, 11.618071556091309, 11.13872241973877, 11.449567794799805, 12.829580307006836, 11.018871307373047, 10.369399070739746, 10.614158630371094, 10.061699867248535, 10.359978675842285, 9.756139755249023, 10.00848388671875, 11.305398941040039, 9.487624168395996, 10.715190887451172, 8.963594436645508, 10.123252868652344, 8.43553638458252, 9.529356956481934, 7.903004169464111, 7.468772888183594, 7.441303253173828, 8.418562889099121, 6.888324737548828, 6.555356979370117, 7.442780494689941, 6.237778663635254, 6.016102313995361, 6.776116371154785, 8.6322603225708, 5.965451717376709, 7.700962066650391, 5.185009479522705, 6.796955108642578, 4.431941986083984, 5.923365592956543, 3.7026560306549072, 2.9259538650512695, 3.0685935020446777, 2.378781318664551, 2.437511444091797, 3.626401662826538, 5.901596546173096, 2.734513998031616, 0.9856656193733215, 0.5314849019050598, 1.1544876098632812]} +{"type": "SampleBatch", "eps_id": [1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1693691551, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAE2EAj1nraq7WeOPO8lGSL4RFwI93DNNvrRg+zkfsck9CVnjPKfzrLtJgSA72bRGvqh74jzTPk2+r627uvuUyz3OpME8voPKvhrQETqGhMg+ztaAPII7Tb7ScQk87QPLPfgAQDyep7G7VO0pPDZ6Q74sOj48x4FNvqW/1juUH9c9F+74Oz+QubtGyw08YgU+vgE49Tu8uk2+gfmhOybw4D2XGmM7Duu/u2706Tscozm+WW1bO/znTb5IS2Y7TLzoPfmHMLrM5cS7Vp+9O8IzNr7nCFC6CZFBPqkGEjtnBPC+qMFDO+6nyLuCNeq72ZszvvG6OztL+E2+TpQuvM+P6z2L0pe6887KvrLjCLxgA88+csYUvAyxTb711Iy5ekTfPbOYVryuP7y7DZP6OgwpPL6eeli8MrdNvrUd57rrU+A9ayeNvE6/yr4dGOA5iKbNPnoIzrwmsk2+iZ4KPKZ23z3I8e68zoTAu5lfLjzgOjm+NejvvI6hQT4eM+Y7bb3wvhTt0Ly0qsQ+GeUbu1S6Qr8i/pG8+I9BPtkckLxa/O++ogtmvDV5v7t76Ny8q/45vs71Z7wEYk2+1Kr6vJLB0T1j15S8CpOkuwbj6bxEjUy+CqqVvJ2ETL67TgW96JarPRljtrxA6sm+UOP8vHBnuz76//a8GJdLvizrwLw8mYI9hMkLvWmpXLuDeLa8jutxviEQDL1y60q+kC3dvGcISj3tSxy9Wawuu78Y1bzN5YC+0oMcveIkSr4LWP68w5QFPbuvLL1Zvci+LAD5vER0oT4Hzky9rj5JvtlVxbyavFg8hedcvQRYyL4AK8O8oK+YPpr1fL0fi0i+9k6SvG8j97pegIa95EX2uQuekrwzrZm+S4WGvSEFSL46y8O8v69XvH+Fjr2du8e+Y/PFvDc0iz4HgJ69A1FHvsxnmby3BOi8B3mmvQ0bQjq7C568eVmnvkNxpr2Rwka+AZnTvAEkJb2RZK69kwS1Ogs02rxgnK6+FVauvRD+Rb4VCgm9hPFovYdBtr2oXBI7wLINvVRHuL4cKra9VP9Evs4uK73kcaC9XAu+vf8Uxr7DmTG9O7BNPhXkzb2jwEO+QyUhvVFk172WuNW9onrFvuDCKb1ECjM+9YTlvWaIFL8icBu94ojoPtxI/b025sS+BnfsvF5dGT6uhAa+M5VBvjTt07zKhRu+02MKvth/xL5pz+y8vbYHPvo/Er6IFxS/kxjXvGj41D7mGB6+lBjEvgvykrzswOs97PAlvqnuE7/PFYC8/N3NPpPGMb6x2cO+stT4u/4K1j0VnDm+Bsk/vkxWtLsTIUO+BnI9vinBw74inBi8HpfNPY1GRb5DyxO/Tm7vuzu9xz5fGVG+AqLDvtzhATqn18I9p+xYvvvEE79xKx0776TGPvi+ZL5qp8O+vWwmPIe2xD13kmy+qtATvxrmRTzyqcg+uGV4vgrRw75vKaM8CBfTPXEdgL5RIEC+jgy0PAGpO75JCYK+fB/Evv8FljxbIO49b/WFviIUFL/TEqk8flDUPsLhi75eaMS+nwPtPCOoAz5ez4++Z3JBviQKAT3Pjx6+l76RvsQ4vDuYteg8Pvbfvoivkb6ZUkK+owqhPOgvC77/oJO+fzTFvojFijzSzSY+sJKXvpPbQr7SdaU8nsP+vYaFmb73ecW+QRSRPJjHMj6aeJ2+e2lDvhGvrTx2Tua93GyfvrNogjtkQps8oPfLvm1in76DmUs+yPozPJBdL782WZ2+B/xoO+3zMbv2Hce+5E+dvvsfRL5R7Cu8M8/GvfhFn74Z48W+jLtLvMrkRD4nO6O+GcpDvgS6DLyAm9W9YDClvs27xb5g5y68th0+PsUkqb7ygUO+IyLku0QK4r1FGau+UMKMO6w7Frzve8++Ag6rvpNFQ77vgo28nHjsvecBrb49B5k73W2gvGm8076p9ay+pLtCvkYv5LzvJgK+Luiuvh8Pxb5OAvm8yWwgPh/Zsr6k4UG+TFffvOrwFL51ybS+T6PEvuwr97wq1Q0+Pri4vsMKQb51euC8qHYnvm6mur6dN8S+w0X7vFaH9j0Qk76+qfETv9uM57yWc84+An7EvkbJw75ZfKW8PGvQPW9oyL7NxBO/7s+UvBepxj6WUc6+aIHDvlJ7KryknLc9kzrSvleqE7+UGg28CBHCPqsi2L7DXcO+wDWHutZQqz3yCty+m+k+vv8lKDrtYVa+rvPdvgddw75eX2i7VxCrPfHb4b673j6+isj1uptRV76RxOO+U2UPPPY/x7tm6AC/oK3jvh3BPr4/UIS8eORZvvWV5b7VKMO+Hy2nvDcbmT0sfem+X3QTv4HtmrxazLg+G2Pvvinfwr6plT+852B/PdpI877XxD2+gicrvPGdb76pLvW+rbXCvu7Ud7w4ymI9khP5vu0/E79GsGW8Rr2vPmn3/r4Of8K+Rm7quy0dPT2dbQG/XRU9vioszLs+u36+D6b3uTeAmTyO0ag8lkO4upurzLji6TS+l5aoPBRtmD4i9227nK2UPFZd2Tx1zz08Ry1Wu2r9WT5AQ9s8dpOLvk+zgTrcb448L5muPCSa6Dy+R686PEpZPhtAszyc1YO+uuK2O1Fi0D47EIk8jg4LvwHPYDwAFRo/oETAO0ifVL9OBNM8iDjQProFMLx0OQm/4NIKPZIZGj+31a+8TQNVv70iPD1pc9A+3xQcvXLVC7/ifF09ZqpZPg7USL1pFYi+reZuPUoT0T4Lml69dtISvzMtiD3fKls+28qGvcDNmL568ZA9QBGjPEYEk71hBN+8NMKRPc9LMr68IZS9IMh3PnOgij1DfLy+dDiKvYbQAj+KGHc9HUMwvpOUar3fk0o+sP5oPViXu77JX1q9Rp7xPvT6Sj1Jpi6+GrczvV/SJj4hAj09XOS6vpheJr3+A+I+CBsfPVhqLb7+NAK9WIALPnw7ET2PudY8AxjuvK1MKr4uYRM9s4Ysvr2rBL1SwO892JMFPenn3TxgKfa87xw+vu3LBz0NnSu+NEoKvQ94xz2PIvQ8ylvlPKRPAr1/rlK+4Lj4POnsZD5mKhO9UB8Cv9WsDj0aQu08Bc48vSWYaL42DBE9bXEpvo1pT72Nd089BX4DPSGI+DxSQ0u9C+CDvkL6BT1x+ie+7VxgvdsfnDwhFPE853e3vkDNXr34gpY+Zl62PPdhJr5LuEa9cLF7vF2/mzyRuLa+dvpHvQr0hT66jUI8+h8Nv7yLMr123Ak/tJ5eOkUMtr4mbga9cwFuPqExzbvu2Qy/lcfmvIG7Az97cY28Wpq1vop4krwbRVo+ZY7HvNokI75nGF+8n1CuvcWo4byGX7W+V/x6vHgeUD5u2Q29dbgivjtjOLz+/cC98t0aveAttb4xRFe8wI1HPgzbN70WXSK+wGgXvGi80L1B2ES9b00WPZPOOLzohcu+tNZBvUUQIr7mh528IAHevcTNTr1cyrS+jEqvvPxtNj7xumu9NEQMvzsakrzJku0+yU6MvfOAtL5wKAy89rwpPoC/mr0FFCG++66ru1a4BL7xMKG9ZjYbPetPALzBDtm+maOfvSPjIL5SnYW8VPIIvhUTpr17ObS+p4abvDhyHT4VfrS9G/4Lv6hVgrzAeuE+LeTKvSf4s76Duei76SgSPvNJ2b3g6Au/sC6LuzTI3T6mrO+94OKzvtuykDshfQ4+uBD+vRTpC7835Os7Rc/dPro5Cr6W+LO+s/OBPKM+Ej6hbBG+nF8gvuJZmTyISxS+vqEUvnybHD24n4E8l/LcvknZE7734CC+wa7rO5gfCb78EBe+6mS0vnTskzsI5iQ+OEgevssMIb5vdf070VgFvsyAIb7xjxo93x2oO8hF1771uiC+BT0hvk3dVruoLwG+f/Qjvlx/tL6DHL67zHUpPsosK74GGyG+UlAju5gdBL6nZS6+2G+0vvs1prsdySY+U501vrT+IL5O3+26To4Gvp7VOL72YrS+dpWRu4iQJD7HDEC+R+cgvs4MoboYkwi+mkRDvlpYtL4qV3+7S7wiPlZ7Sr4Y1CC+9ic8ulM6Cr7Hsk2+uU+0vn34X7tdPyE+K+lUvp3EIL40l4y5048Lvk0gWL7USLS+hDZEuxIPID5qVl++Z7ggvi+ZCjk2nQy+TY1ivnlDtL7bUiu7yyIfPjPDab4cryC+MXsBOi1qDb7n+Wy+VHQcPfyjFLsye9y+pDFsvoeoIL6kRDK8+fwNvjZob76kJrS+ULRfvBIvGj71nHa+sEwgvpJdLrzw5hW+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAABEXAj3cM02+tGD7OR+xyT0JWeM8p/Osu0mBIDvZtEa+qHviPNM+Tb6vrbu6+5TLPc6kwTy+g8q+GtAROoaEyD7O1oA8gjtNvtJxCTztA8s9+ABAPJ6nsbtU7Sk8NnpDviw6PjzHgU2+pb/WO5Qf1z0X7vg7P5C5u0bLDTxiBT6+ATj1O7y6Tb6B+aE7JvDgPZcaYzsO67+7bvTpOxyjOb5ZbVs7/OdNvkhLZjtMvOg9+YcwuszlxLtWn707wjM2vucIULoJkUE+qQYSO2cE8L6owUM77qfIu4I16rvZmzO+8bo7O0v4Tb5OlC68z4/rPYvSl7rzzsq+suMIvGADzz5yxhS8DLFNvvXUjLl6RN89s5hWvK4/vLsNk/o6DCk8vp56WLwyt02+tR3nuutT4D1rJ428Tr/Kvh0Y4DmIps0+egjOvCayTb6Jngo8pnbfPcjx7rzOhMC7mV8uPOA6Ob416O+8jqFBPh4z5jttvfC+FO3QvLSqxD4Z5Ru7VLpCvyL+kbz4j0E+2RyQvFr8776iC2a8NXm/u3vo3Lyr/jm+zvVnvARiTb7Uqvq8ksHRPWPXlLwKk6S7BuPpvESNTL4KqpW8nYRMvrtOBb3olqs9GWO2vEDqyb5Q4/y8cGe7Pvr/9rwYl0u+LOvAvDyZgj2EyQu9aalcu4N4tryO63G+IRAMvXLrSr6QLd28ZwhKPe1LHL1ZrC67vxjVvM3lgL7Sgxy94iRKvgtY/rzDlAU9u68svVm9yL4sAPm8RHShPgfOTL2uPkm+2VXFvJq8WDyF51y9BFjIvgArw7ygr5g+mvV8vR+LSL72TpK8byP3ul6Ahr3kRfa5C56SvDOtmb5LhYa9IQVIvjrLw7y/r1e8f4WOvZ27x75j88W8NzSLPgeAnr0DUUe+zGeZvLcE6LwHeaa9DRtCOrsLnrx5Wae+Q3GmvZHCRr4BmdO8ASQlvZFkrr2TBLU6CzTavGCcrr4VVq69EP5FvhUKCb2E8Wi9h0G2vahcEjvAsg29VEe4vhwqtr1U/0S+zi4rveRxoL1cC769/xTGvsOZMb07sE0+FeTNvaPAQ75DJSG9UWTXvZa41b2iesW+4MIpvUQKMz71hOW9ZogUvyJwG73iiOg+3Ej9vTbmxL4Gd+y8Xl0ZPq6EBr4zlUG+NO3TvMqFG77TYwq+2H/EvmnP7Ly9tgc++j8SvogXFL+TGNe8aPjUPuYYHr6UGMS+C/KSvOzA6z3s8CW+qe4Tv88VgLz83c0+k8YxvrHZw76y1Pi7/grWPRWcOb4GyT++TFa0uxMhQ74Gcj2+KcHDviKcGLwel809jUZFvkPLE79Obu+7O73HPl8ZUb4CosO+3OEBOqfXwj2n7Fi++8QTv3ErHTvvpMY++L5kvmqnw769bCY8h7bEPXeSbL6q0BO/GuZFPPKpyD64ZXi+CtHDvm8pozwIF9M9cR2AvlEgQL6ODLQ8Aak7vkkJgr58H8S+/wWWPFsg7j1v9YW+IhQUv9MSqTx+UNQ+wuGLvl5oxL6fA+08I6gDPl7Pj75nckG+JAoBPc+PHr6XvpG+xDi8O5i16Dw+9t++iK+RvplSQr6jCqE86C8Lvv+gk75/NMW+iMWKPNLNJj6wkpe+k9tCvtJ1pTyew/69hoWZvvd5xb5BFJE8mMcyPpp4nb57aUO+Ea+tPHZO5r3cbJ++s2iCO2RCmzyg98u+bWKfvoOZSz7I+jM8kF0vvzZZnb4H/Gg77fMxu/Ydx77kT52++x9EvlHsK7wzz8a9+EWfvhnjxb6Mu0u8yuREPic7o74ZykO+BLoMvICb1b1gMKW+zbvFvmDnLry2HT4+xSSpvvKBQ74jIuS7RArivUUZq75Qwow7rDsWvO97z74CDqu+k0VDvu+CjbyceOy95wGtvj0HmTvdbaC8abzTvqn1rL6ku0K+Ri/kvO8mAr4u6K6+Hw/Fvk4C+bzJbCA+H9myvqThQb5MV9+86vAUvnXJtL5Po8S+7Cv3vCrVDT4+uLi+wwpBvnV64Lyodie+bqa6vp03xL7DRfu8Vof2PRCTvr6p8RO/24znvJZzzj4CfsS+RsnDvll8pbw8a9A9b2jIvs3EE7/uz5S8F6nGPpZRzr5ogcO+UnsqvKSctz2TOtK+V6oTv5QaDbwIEcI+qyLYvsNdw77ANYe61lCrPfIK3L6b6T6+/yUoOu1hVr6u892+B13Dvl5faLtXEKs98dvhvrvePr6KyPW6m1FXvpHE475TZQ889j/Hu2boAL+greO+HcE+vj9QhLx45Fm+9ZXlvtUow74fLae8NxuZPSx96b5fdBO/ge2avFrMuD4bY+++Kd/CvqmVP7znYH892kjzvtfEPb6CJyu88Z1vvqku9b6ttcK+7tR3vDjKYj2SE/m+7T8Tv0awZbxGva8+aff+vg5/wr5Gbuq7LR09PZ1tAb9dFT2+KizMuz67fr6kXwK/+GXCvq6ZN7y00ys9m6vMuOLpNL6Xlqg8FG2YPiL3bbucrZQ8Vl3ZPHXPPTxHLVa7av1ZPkBD2zx2k4u+T7OBOtxvjjwvma48JJroPL5Hrzo8Slk+G0CzPJzVg7664rY7UWLQPjsQiTyODgu/Ac9gPAAVGj+gRMA7SJ9Uv04E0zyIONA+ugUwvHQ5Cb/g0go9khkaP7fVr7xNA1W/vSI8PWlz0D7fFBy9ctULv+J8XT1mqlk+DtRIvWkViL6t5m49ShPRPguaXr120hK/My2IPd8qWz7byoa9wM2YvnrxkD1AEaM8RgSTvWEE37w0wpE9z0syvrwhlL0gyHc+c6CKPUN8vL50OIq9htACP4oYdz0dQzC+k5Rqvd+TSj6w/mg9WJe7vslfWr1GnvE+9PpKPUmmLr4atzO9X9ImPiECPT1c5Lq+mF4mvf4D4j4IGx89WGotvv40Ar1YgAs+fDsRPY+51jwDGO68rUwqvi5hEz2zhiy+vasEvVLA7z3YkwU96efdPGAp9rzvHD6+7csHPQ2dK740Sgq9D3jHPY8i9DzKW+U8pE8CvX+uUr7guPg86exkPmYqE71QHwK/1awOPRpC7TwFzjy9JZhovjYMET1tcSm+jWlPvY13Tz0FfgM9IYj4PFJDS70L4IO+QvoFPXH6J77tXGC92x+cPCEU8Tznd7e+QM1evfiClj5mXrY892Emvku4Rr1wsXu8Xb+bPJG4tr52+ke9CvSFPrqNQjz6Hw2/vIsyvXbcCT+0nl46RQy2viZuBr1zAW4+oTHNu+7ZDL+Vx+a8gbsDP3txjbxamrW+iniSvBtFWj5ljse82iQjvmcYX7yfUK69xajhvIZftb5X/Hq8eB5QPm7ZDb11uCK+O2M4vP79wL3y3Rq94C21vjFEV7zAjUc+DNs3vRZdIr7AaBe8aLzQvUHYRL1vTRY9k844vOiFy7601kG9RRAivuaHnbwgAd69xM1OvVzKtL6MSq+8/G02PvG6a700RAy/OxqSvMmS7T7JToy984C0vnAoDLz2vCk+gL+avQUUIb77rqu7VrgEvvEwob1mNhs9608AvMEO2b6Zo5+9I+MgvlKdhbxU8gi+FROmvXs5tL6nhpu8OHIdPhV+tL0b/gu/qFWCvMB64T4t5Mq9J/izvoO56LvpKBI+80nZveDoC7+wLou7NMjdPqas773g4rO+27KQOyF9Dj64EP69FOkLvzfk6ztFz90+ujkKvpb4s76z84E8oz4SPqFsEb6cXyC+4lmZPIhLFL6+oRS+fJscPbifgTyX8ty+SdkTvvfgIL7Brus7mB8JvvwQF77qZLS+dOyTOwjmJD44SB6+ywwhvm91/TvRWAW+zIAhvvGPGj3fHag7yEXXvvW6IL4FPSG+Td1Wu6gvAb5/9CO+XH+0voMcvrvMdSk+yiwrvgYbIb5SUCO7mB0EvqdlLr7Yb7S++zWmux3JJj5TnTW+tP4gvk7f7bpOjga+ntU4vvZitL52lZG7iJAkPscMQL5H5yC+zgyhuhiTCL6aREO+Wli0vipXf7tLvCI+VntKvhjUIL72Jzy6UzoKvseyTb65T7S+ffhfu10/IT4r6VS+ncQgvjSXjLnTjwu+TSBYvtRItL6ENkS7Eg8gPmpWX75nuCC+L5kKOTadDL5NjWK+eUO0vttSK7vLIh8+M8NpvhyvIL4xewE6LWoNvuf5bL5UdBw9/KMUuzJ73L6kMWy+h6ggvqREMrz5/A2+NmhvvqQmtL5QtF+8Ei8aPvWcdr6wTCC+kl0uvPDmFb6x0Xm+L/mzvo1VXrwJWRI+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.7756350636482239, 0.612125813961029, 0.7756436467170715, 0.3885174095630646, 0.8815892934799194, 0.6356493234634399, 0.7581368088722229, 0.6445144414901733, 0.7527977228164673, 0.6516579389572144, 0.7484222650527954, 0.6572870016098022, 0.2550281286239624, 0.9149109125137329, 0.7610495686531067, 0.36838531494140625, 0.8866824507713318, 0.6457452774047852, 0.7557428479194641, 0.354793906211853, 0.8928549289703369, 0.668721079826355, 0.26495277881622314, 0.08777742087841034, 0.9563407301902771, 0.9214736819267273, 0.7911956906318665, 0.5715227723121643, 0.8113599419593811, 0.46872013807296753, 0.8444793224334717, 0.5148441195487976, 0.8360649943351746, 0.4795854985713959, 0.8509207963943481, 0.5608783960342407, 0.7989695072174072, 0.5798503160476685, 0.7905892133712769, 0.4081323742866516, 0.8741276264190674, 0.623587965965271, 0.7611796855926514, 0.362978458404541, 0.8880791068077087, 0.33047470450401306, 0.8981094360351562, 0.29254063963890076, 0.9087594747543335, 0.7488438487052917, 0.6212868094444275, 0.7758610248565674, 0.4188234508037567, 0.8572765588760376, 0.567042350769043, 0.8044984340667725, 0.4615945816040039, 0.8433027267456055, 0.46471813321113586, 0.8454596996307373, 0.5439815521240234, 0.8119010329246521, 0.4629794657230377, 0.8489465117454529, 0.44517841935157776, 0.8587840795516968, 0.41622641682624817, 0.8716719150543213, 0.623212456703186, 0.7551260590553284, 0.3523470163345337, 0.8933494091033936, 0.6929872632026672, 0.30915963649749756, 0.8949166536331177, 0.680034875869751, 0.729566752910614, 0.6561221480369568, 0.7493543028831482, 0.3704221546649933, 0.12344934046268463, 0.9468966126441956, 0.8860577344894409, 0.6618601679801941, 0.7321107387542725, 0.6663256287574768, 0.7279475331306458, 0.3306964635848999, 0.8942685127258301, 0.30627503991127014, 0.9030520915985107, 0.726478636264801, 0.6553636193275452, 0.7468454241752625, 0.6281380653381348, 0.7662988901138306, 0.4013173282146454, 0.8649280667304993, 0.4087170958518982, 0.8637696504592896, 0.40591245889663696, 0.8660423159599304, 0.6065399050712585, 0.7637236714363098, 0.6049522161483765, 0.23537707328796387, 0.9175612330436707, 0.7841527462005615, 0.43730372190475464, 0.8495862483978271, 0.562748908996582, 0.7911516427993774, 0.4525795578956604, 0.8426398634910583, 0.551680862903595, 0.7927470803260803, 0.35888081789016724, 0.8957231640815735, 0.6768782138824463, 0.7230063676834106, 0.6946155428886414, 0.29071521759033203, 0.09575942158699036, 0.9542697072029114, 0.08424267917871475, 0.9584065675735474, 0.9297118782997131, 0.17087379097938538, 0.9403315782546997, 0.8702080249786377, 0.6345273852348328, 0.26655271649360657, 0.9053779244422913, 0.30668720602989197, 0.8960314393043518, 0.3381303548812866, 0.8890500068664551, 0.6409712433815002, 0.7704555988311768, 0.604201078414917, 0.7936940789222717, 0.5633335709571838, 0.18421390652656555, 0.934615969657898, 0.8475006222724915, 0.4323859214782715, 0.8700407147407532, 0.6335121393203735, 0.7338801622390747, 0.6772472858428955, 0.3027171492576599, 0.8957626819610596, 0.31503212451934814, 0.8953123092651367, 0.6868011951446533, 0.7152880430221558, 0.678637683391571, 0.7223745584487915, 0.6722382307052612, 0.2721746265888214, 0.9104246497154236, 0.7503911852836609, 0.3685230016708374, 0.8811478018760681, 0.6412357687950134, 0.24763242900371552, 0.9166780710220337, 0.7716503143310547, 0.3969353437423706, 0.8723196387290955, 0.38646960258483887, 0.8790027499198914, 0.3632466793060303, 0.8891210556030273, 0.6716672778129578, 0.278555303812027, 0.9067665338516235, 0.7231478095054626, 0.6862280368804932, 0.2860000431537628, 0.9065383076667786, 0.7248930335044861, 0.6793040037155151, 0.7249322533607483, 0.679607093334198, 0.7244158983230591, 0.680435061454773, 0.7234086394309998, 0.6817116141319275, 0.7219614386558533, 0.6833726763725281, 0.720114529132843, 0.6853653192520142, 0.717898428440094, 0.6876468658447266, 0.2846648395061493, 0.9079360961914062, 0.7332726120948792, 0.6636269092559814, 0.740626871585846], "advantages": [-0.9090890884399414, -1.127325415611267, -1.4979877471923828, -1.7770287990570068, -0.9555240273475647, -2.489248037338257, -2.7517693042755127, -3.15179705619812, -3.3717381954193115, -3.820936679840088, -3.996267795562744, -4.496857166290283, -4.625176906585693, -3.5213992595672607, -5.18390417098999, -5.796466827392578, -5.304215431213379, -6.538675785064697, -6.4541015625, -7.231523036956787, -6.871771335601807, -7.994811058044434, -7.805431842803955, -6.283897399902344, -3.234347105026245, -6.558821678161621, -8.778098106384277, -9.84870719909668, -9.216177940368652, -10.44337272644043, -10.484763145446777, -11.097777366638184, -10.046831130981445, -11.628363609313965, -10.332232475280762, -12.08273696899414, -12.596323013305664, -12.58414077758789, -13.297853469848633, -13.064221382141113, -11.137035369873047, -13.369363784790039, -14.462318420410156, -13.730950355529785, -11.508196830749512, -13.889410018920898, -11.53868293762207, -13.903693199157715, -11.468234062194824, -13.756550788879395, -15.538286209106445, -13.697656631469727, -15.57264518737793, -16.732980728149414, -15.741827964782715, -13.910992622375488, -15.761367797851562, -17.291433334350586, -15.994025230407715, -17.592037200927734, -16.347126007080078, -15.096712112426758, -16.604503631591797, -18.157657623291016, -17.127408981323242, -18.591249465942383, -17.779804229736328, -19.132686614990234, -18.557809829711914, -18.199935913085938, -19.238386154174805, -20.373611450195312, -20.18935203552246, -20.186275482177734, -19.96440887451172, -20.91289710998535, -21.710935592651367, -21.88909339904785, -22.617202758789062, -22.915283203125, -22.948945999145508, -22.311323165893555, -23.52683448791504, -24.237096786499023, -24.829599380493164, -25.225664138793945, -25.7595272064209, -26.241615295410156, -26.46504020690918, -27.02013397216797, -27.275028228759766, -27.77414321899414, -28.14527130126953, -28.757505416870117, -29.06433868408203, -29.748523712158203, -29.99290657043457, -30.538898468017578, -31.161794662475586, -31.647903442382812, -32.37772750854492, -32.81642150878906, -33.64552307128906, -34.683815002441406, -34.75780487060547, -35.85243606567383, -36.564735412597656, -36.82870101928711, -36.87998962402344, -37.2624626159668, -38.26713943481445, -39.47517395019531, -39.5224723815918, -39.87834930419922, -40.99521255493164, -42.26863098144531, 10.817754745483398, 11.002147674560547, 10.40869426727295, 10.572393417358398, 10.015243530273438, 10.132020950317383, 11.14441204071045, 13.743602752685547, 10.855884552001953, 13.747347831726074, 10.741920471191406, 9.208016395568848, 10.72360897064209, 9.030326843261719, 8.344115257263184, 8.487646102905273, 9.67080307006836, 7.91990852355957, 8.974520683288574, 7.3722710609436035, 8.30816650390625, 6.837972164154053, 6.667980670928955, 6.378802299499512, 6.317411422729492, 5.931929111480713, 5.999307632446289, 7.51806116104126, 5.8334760665893555, 5.213784694671631, 5.706218242645264, 4.912362575531006, 5.191457271575928, 4.591506004333496, 4.683869361877441, 5.95835018157959, 4.115170955657959, 5.216538429260254, 3.5854387283325195, 3.6977410316467285, 3.1883649826049805, 3.5112197399139404, 2.8282289505004883, 3.3661630153656006, 5.586302280426025, 3.419462203979492, 2.3931772708892822, 2.6711783409118652, 2.0857045650482178, 3.3379178047180176, 5.920799732208252, 3.55290150642395, 2.003239393234253, 1.6574945449829102, 1.8609412908554077, 1.3016539812088013, 1.6781538724899292, 0.9470158219337463, 1.4215353727340698, 3.134845018386841, 5.222073078155518, 3.1220991611480713, 1.2644894123077393, 2.969377040863037, 4.856164932250977, 3.004380464553833, 1.257417917251587, 2.876025676727295, 1.206418752670288, 2.7077701091766357, 1.131942868232727, 2.4993934631347656, 1.0279994010925293, 2.251680374145508, 0.8901981711387634, 1.9660530090332031, 0.715822160243988, 1.6443406343460083, 0.5035478472709656, 1.2885972261428833, 2.169264316558838, 1.168209195137024, 0.23358669877052307, 0.8329963684082031]} +{"type": "SampleBatch", "eps_id": [1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1085404201, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAALHReb4v+bO+jVVevAlZEj5PgoC+EfIfvr2AL7wsth2+xRuCvr7Ls755+GG8AYQKPlO0hb5CzAu/RqU1vA/e2D7bS4u+JJ2zvtxmK7tCeAI+euOOvp+/C78A0Yy4XarWPoF6lL4YmLO+KUkIPBuaAT4GEpi+TccLvyjCMTwJ/tc+W6mdvvK7s74r/508xcwHPphBob6k8x++iLmzPAadHb4S26K+Ggm0vrCBmjwMGRU+23SmvkaMIL68XLI8tnQQvtsPqL4iVbS+0j+bPPYzIj4pq6u+qyQhvqgztTwuUgO+sEetvvahtL6/MKA8gnIvPobksL6nwCG+EUO8PIzB672dgrK+itQWPclmqTwbCc2+FSKyvi9kIr6KlE88GIbPvc3Bs76v5hQ9XmAuPKCsx76BYrO+jbsivuhWOjsHc8C9GgO1viBKtb4JsHw6XWZMPk6juL7SySK+4WaiO+r9vb0LRLq+M1W1vn01SzvvTk4+d+S9vgPoIr5NpOk7nMq4vYGFv75UaLW+MIKuOz2bUT5QJsO+rhYjvg1UGjxmwLC90sfEvpNOEj2dGPw7K4PAvi9qxL7UViO+C8U1OfSvpb1VDMa+xJK1vspbvbpv61g+/K3JvhJSI75X+jY7B4KmvRZQy75LlbW+XtOYOqNaWT7L8c6+6WAjviFQsTvs86O9C5TQvpehtb47snk7XXlbPv411L54gyO+4KcEPAUAnr2X2NW+2be1vmLA1jvTT18+/HrZvl66I77l1TI83YqUvSEe277fqA89mREbPEI3ub4wwtq+wgYkvlIfEjseXoe9GGbcvpD7Dj2/8m06wFW3vpYK3L6KEiS+3OzMu99Whb2crt2+eOS1vgSY97uuAmc+5lHhvsveI76Gfke7WkGOvWn14r6k6g890ESRu47oub5NmeK+zMIjvqSdP7ywFpO9iDzkvsHaED1kJle81YC8vtPf474ANWw+YuWnvCM7Kr8ig+G+78wSPQRsCr248sG+LiXhvteYIr4hdCm9KJbGvW7F4r4uKBc9qGUxvYwIzr6wZOK+L1whvshcUr2mUP29xQHkvu8vtL67fly9/wMcPlSc574F0x++iANQvYySIL57Nem+l2yzvg7cXL25l/Q9IszsvutLHr5tE1O9ylFCvl9h7r6Up7K+Fp9ivY2XsD0V9PG+lRALv8mOW73azrg+G4T3vtzbsb4M/T29q2xUPb4S+76kQxu+b705vdqagr44oPy+RDCxvgOjTr3PCrw8thUAv7PeGb6gwUy97waSvqraAL9udLC+5R5kvUWTDrxkngK/gFUYvmTVZL11CKO+YWEDv+Sjr7456369u7AzvQQjBb+Fhwm/nkGBvfwHaj4r4we/l7muvkfKb73Kt6q9dqIJvz8WCb9unna9qdxCPllgDL+l262+qAdnvbdP971sHQ6/h6kIvyDscL0hTB0+ItkQv2MErb6tVmS9Os0gvg+UEr+dPgi/5TNxvd3A8D2hTRW/vPc5v5SSZ711Pcg+yQUZv8yza7+/iEe9k1kqP5W8Hb8tljm/rwURvXkztz7JciG/kGdrv4hr57zInSM/Digmv/9aOb9EaX28wOSsPhPdKb9JWQe/Y8IOvLaBJD0Qkiy/UbuqvkqZAbz7sIK+Ikcuv7ZJB7/CPVW8TAkPPc/7ML/VMDm/YMxJvMKdpT78rzS/+TEHv5+bv7tql9w8L2Q3v/UfOb/o9a27b7GiPgYYO7+EJwe/eCWJOizBvzwEzD2/+WmqvgGCxjors4m+RoA/v+opB7/9QX273F7GPFA0Qr/IGTm/u4Rdu1OgoT4H6EW/IAprv6o+QDtydxs/bptKv08ZOb/1Dnc8po2hPiNPTr8WNAe/7TmvPDt24jxhA1G/RDs5v2nBszxYbKc+w7dUv6tcB7+4VOk8OzopPdFsV79VCau+nBnwPIUPeL6sIlm/zJIHvwxpyDwR23M9zthbv6Rtq74hKtI85bxmvqqPXb+8wQe/Gz+tPJ5Imj29RmC/AsWrvtaWuTwrqFe+eP5hv8USEL6HFZc8wfr9vuK2Yr9l1109Dp8LPAVqSL/kb2K/oXYQvk3R6buxm/m+zihjvwwVrL46VIq8EtJJvlbhZL/QHRC+zJ6qvHF3/b7OmWW/V+5fPci6+7xZUku/JlJlv3FgD76A7T692egCv6sJZr9DPau+ndFovYRWb74LwGe/B1oHv0H3e72I4CU9C3Vqv1sQOb/3pXi91lWgPpIobr/E5wa/of5eve6WADtJ23C/c6U4v3vVXr290I0+rIx0v/OBBr+5JEi9b4EEvVo9d79kRTi/J8tKvaVkej7S7Hq/WQhqvxnDNr2jbAU/EZt/v5ntN7/6EAy96/BbPmSkgb8r2wW/KvH0vFg3tb0Q+4K/2Jinvj24Ab3s6Me+ltGDv5ChBb+LtCG9lgPdva4nhb+Vbje/uosqveEuMD5E/Ya/eDlpv39zHL3sEuc+NDZJvZfbArzW1f08AcM3vbPdSb2TbFC+Hnz2PCjOgz46ilq9IScRvM5UED03zdG8BURbvZwyPj62Ow49SCudvsYMTL1fiCG8IizqPDogaLuJ20y9EUc9PpKX6Tzo+pK+Ibc9vaTnLrz2jro8UwltPAKXPr0MiDw+xu28PC25ir7igS+9+KM5vI+JkDx+7+w8gW8wvXKpU76rRpU8QXynPldeQb0VCkK8Et/KPEnRJD22VkK9E047Ps120TzVZXq+tFozvf5cwT6CZqk8EG8Hv4tqFL1MoTo+DnIlPApua75cfAW9kiTBPko3tDvq9gS/dirNvOlmOj5ULKC7mWFmvm9Xr7zLS1W8/84ZvHZ8hz15ebG8Q5w6PnshBLxk+2q+6Z2TvFc1Ubw1U0+85G94PXy1lbys6To+N3M7vJ6ocb4Sm2+8CnHBPvJjhLx8Qgi/N5vnu5RQOz69mNu8L5d6viBzX7tzZ0G8d9gBvaNYIT0L7G67rWlTvr48/bzh0KQ+5cP+uzC+MrwFf8i81t2gPBz1Ary5m1K+YkfFvDznmz4kWka8tF4nvMpjk7xCh407FLNJvCL/Ub6irpK8iiGVPgBzhrzt9R68pOtFvDSF5bvwCYi8iws+Pjc3SLz2YJu+YkNTvMVEGbw81JW82lpwvB5UVrxaHFG+izuYvApciz48n4y8o6EQvGxGV7xAcde8fhGOvLH0Pj6N5F+8K22lvuIHX7yDWgq8A+KkvIpcDr1BzGG8jSZQvsuTqrzJxYA++jOSvD7DALy+XoG83D9DvZt9k7xE/D8+GC6JvBPKsL7Fi2m8tknyu67AwbxKTm29B/hrvIOSQD6xPsu8+0m3voFYLrwFmtu73fICvfD4lb2vijC8eGVBPpTyCL0ObMC+Tk/lu9zVxD4wvCe9tcktv7sdNTrjeEI+5Vhfvfd2zL4LGpM7D/ePu+IHgL3/pf698TiQOxBVS76vH4W99IgXPrNv4TkKEMm+7x9+vQnF1j46RfO7nnpJvvfCW70n99w9xRs6vDc1brpI7FK9RQ1Ovv9nOryp8Ue+OWhjvUMnmT1pY3q8AnLHvu9HXb1h6rI+RAS9vOVXRr6Np0C9WcUkPWnA3LycmQc77Vs9vVCpiL6gady8+PpEvpM5U719gDE88e37vDSoYjtfVlK9U2eYvuFc+7wLekO+07hqvdoCsbzOUQ29dDLFvvl9bL05QoE+/94svbpTFL+Jz1e9M8EGP/NVXL2lZMS+b7AsvWrMXj42wnu9G2RAvoXdGr2vbrS9MZOFvXLPw74mFSK96AhFPmM9lb15Qj++4FESvd5c5r3i45y9wM0QPMqIG713GtW+Noecvd8uPr6BoT29UAALvq8ipL2so8K+QMBIvaljET7mtLO9LxUTv60ePb0SbtY+aD3LvbXxwb6gzxq99D3lPWLB2r3jjDu+MKQRvRsMRb7mQeK9f2jBvrdnIb2/67U95rrxvVWBEr/YIBq9Zta8PuCVBL5J2MC+H9T3vGwhhD2cTAy+ZUISvxdC7bz867E+AAAYvl9pwL68UrS8l7g7PUuyH74RExK/edCsvMS4qT7mYSu+ffJDv7sBbbzUEB4/5g47vu7xEb8Duwq7lPajPtq7Rr7l+L++9IGMOzxK3DymaU6+7TM4vn4hnjsrZIe+xBhSvvkJwL5TuPK5x9XzPD/HWb4Q+BG/GscKOb8EpT6xdGW+WQnAvoCP1zvq+vI8JSNtvopdOL67/+o73JmFvhnTcL4eI8C+afX/Oj1ACz2Vgni+5wYSvySKLDs9k6c+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 1], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAE+CgL4R8h++vYAvvCy2Hb7FG4K+vsuzvnn4YbwBhAo+U7SFvkLMC79GpTW8D97YPttLi74knbO+3GYru0J4Aj56446+n78LvwDRjLhdqtY+gXqUvhiYs74pSQg8G5oBPgYSmL5Nxwu/KMIxPAn+1z5bqZ2+8ruzviv/nTzFzAc+mEGhvqTzH76IubM8Bp0dvhLbor4aCbS+sIGaPAwZFT7bdKa+Rowgvrxcsjy2dBC+2w+oviJVtL7SP5s89jMiPimrq76rJCG+qDO1PC5SA76wR62+9qG0vr8woDyCci8+huSwvqfAIb4RQ7w8jMHrvZ2Csr6K1BY9yWapPBsJzb4VIrK+L2QivoqUTzwYhs+9zcGzvq/mFD1eYC48oKzHvoFis76NuyK+6FY6OwdzwL0aA7W+IEq1vgmwfDpdZkw+TqO4vtLJIr7hZqI76v29vQtEur4zVbW+fTVLO+9OTj535L2+A+givk2k6Tucyri9gYW/vlRotb4wgq47PZtRPlAmw76uFiO+DVQaPGbAsL3Sx8S+k04SPZ0Y/Dsrg8C+L2rEvtRWI74LxTU59K+lvVUMxr7EkrW+ylu9um/rWD78rcm+ElIjvlf6NjsHgqa9FlDLvkuVtb5e05g6o1pZPsvxzr7pYCO+IVCxO+zzo70LlNC+l6G1vjuyeTtdeVs+/jXUvniDI77gpwQ8BQCevZfY1b7Zt7W+YsDWO9NPXz78etm+XrojvuXVMjzdipS9IR7bvt+oDz2ZERs8Qje5vjDC2r7CBiS+Uh8SOx5eh70YZty+kPsOPb/ybTrAVbe+lgrcvooSJL7c7My731aFvZyu3b545LW+BJj3u64CZz7mUeG+y94jvoZ+R7taQY69afXivqTqDz3QRJG7jui5vk2Z4r7MwiO+pJ0/vLAWk72IPOS+wdoQPWQmV7zVgLy+09/jvgA1bD5i5ae8IzsqvyKD4b7vzBI9BGwKvbjywb4uJeG+15giviF0Kb0olsa9bsXivi4oFz2oZTG9jAjOvrBk4r4vXCG+yFxSvaZQ/b3FAeS+7y+0vrt+XL3/Axw+VJznvgXTH76IA1C9jJIgvns16b6XbLO+DtxcvbmX9D0izOy+60sevm0TU73KUUK+X2HuvpSnsr4Wn2K9jZewPRX08b6VEAu/yY5bvdrOuD4bhPe+3Nuxvgz9Pb2rbFQ9vhL7vqRDG75vvTm92pqCvjig/L5EMLG+A6NOvc8KvDy2FQC/s94ZvqDBTL3vBpK+qtoAv250sL7lHmS9RZMOvGSeAr+AVRi+ZNVkvXUIo75hYQO/5KOvvjnrfr27sDO9BCMFv4WHCb+eQYG9/AdqPivjB7+Xua6+R8pvvcq3qr12ogm/PxYJv26edr2p3EI+WWAMv6Xbrb6oB2e9t0/3vWwdDr+HqQi/IOxwvSFMHT4i2RC/YwStvq1WZL06zSC+D5QSv50+CL/lM3G93cDwPaFNFb+89zm/lJJnvXU9yD7JBRm/zLNrv7+IR72TWSo/lbwdvy2WOb+vBRG9eTO3PslyIb+QZ2u/iGvnvMidIz8OKCa//1o5v0RpfbzA5Kw+E90pv0lZB79jwg68toEkPRCSLL9Ru6q+SpkBvPuwgr4iRy6/tkkHv8I9VbxMCQ89z/swv9UwOb9gzEm8wp2lPvyvNL/5MQe/n5u/u2qX3DwvZDe/9R85v+j1rbtvsaI+Bhg7v4QnB794JYk6LMG/PATMPb/5aaq+AYLGOiuzib5GgD+/6ikHv/1BfbvcXsY8UDRCv8gZOb+7hF27U6ChPgfoRb8gCmu/qj5AO3J3Gz9um0q/Txk5v/UOdzymjaE+I09OvxY0B7/tOa88O3biPGEDUb9EOzm/acGzPFhspz7Dt1S/q1wHv7hU6Tw7Oik90WxXv1UJq76cGfA8hQ94vqwiWb/Mkge/DGnIPBHbcz3O2Fu/pG2rviEq0jzlvGa+qo9dv7zBB78bP608nkiaPb1GYL8Cxau+1pa5PCuoV754/mG/xRIQvocVlzzB+v2+4rZiv2XXXT0Onws8BWpIv+RvYr+hdhC+TdHpu7Gb+b7OKGO/DBWsvjpUirwS0km+VuFkv9AdEL7Mnqq8cXf9vs6ZZb9X7l89yLr7vFlSS78mUmW/cWAPvoDtPr3Z6AK/qwlmv0M9q76d0Wi9hFZvvgvAZ78HWge/Qfd7vYjgJT0LdWq/WxA5v/eleL3WVaA+kihuv8TnBr+h/l697pYAO0nbcL9zpTi/e9Vevb3QjT6sjHS/84EGv7kkSL1vgQS9Wj13v2RFOL8ny0q9pWR6PtLser9ZCGq/GcM2vaNsBT8Rm3+/me03v/oQDL3r8Fs+ZKSBvyvbBb8q8fS8WDe1vRD7gr/YmKe+PbgBvezox76W0YO/kKEFv4u0Ib2WA929rieFv5VuN7+6iyq94S4wPkT9hr94OWm/f3McvewS5z5SUom/9SM3v2v17rzRXRY+s91JvZNsUL4efPY8KM6DPjqKWr0hJxG8zlQQPTfN0bwFRFu9nDI+PrY7Dj1IK52+xgxMvV+IIbwiLOo8OiBou4nbTL0RRz0+kpfpPOj6kr4htz29pOcuvPaOujxTCW08Apc+vQyIPD7G7bw8LbmKvuKBL734ozm8j4mQPH7v7DyBbzC9cqlTvqtGlTxBfKc+V15BvRUKQrwS38o8SdEkPbZWQr0TTjs+zXbRPNVler60WjO9/lzBPoJmqTwQbwe/i2oUvUyhOj4OciU8Cm5rvlx8Bb2SJME+Sje0O+r2BL92Ks286WY6PlQsoLuZYWa+b1evvMtLVbz/zhm8dnyHPXl5sbxDnDo+eyEEvGT7ar7pnZO8VzVRvDVTT7zkb3g9fLWVvKzpOj43czu8nqhxvhKbb7wKccE+8mOEvHxCCL83m+e7lFA7Pr2Y27wvl3q+IHNfu3NnQbx32AG9o1ghPQvsbrutaVO+vjz9vOHQpD7lw/67ML4yvAV/yLzW3aA8HPUCvLmbUr5iR8W8POebPiRaRry0Xie8ymOTvEKHjTsUs0m8Iv9RvqKukryKIZU+AHOGvO31Hryk60W8NIXlu/AJiLyLCz4+NzdIvPZgm75iQ1O8xUQZvDzUlbzaWnC8HlRWvFocUb6LO5i8ClyLPjyfjLyjoRC8bEZXvEBx17x+EY68sfQ+Po3kX7wrbaW+4gdfvINaCrwD4qS8ilwOvUHMYbyNJlC+y5OqvMnFgD76M5K8PsMAvL5egbzcP0O9m32TvET8Pz4YLom8E8qwvsWLaby2SfK7rsDBvEpObb0H+Gu8g5JAPrE+y7z7Sbe+gVguvAWa27vd8gK98PiVva+KMLx4ZUE+lPIIvQ5swL5OT+W73NXEPjC8J721yS2/ux01OuN4Qj7lWF+993bMvgsakzsP94+74geAvf+l/r3xOJA7EFVLvq8fhb30iBc+s2/hOQoQyb7vH369CcXWPjpF87ueekm+98JbvSf33D3FGzq8NzVuukjsUr1FDU6+/2c6vKnxR745aGO9QyeZPWljerwCcse+70ddvWHqsj5EBL285VdGvo2nQL1ZxSQ9acDcvJyZBzvtWz29UKmIvqBp3Lz4+kS+kzlTvX2AMTzx7fu8NKhiO19WUr1TZ5i+4Vz7vAt6Q77TuGq92gKxvM5RDb10MsW++X1svTlCgT7/3iy9ulMUv4nPV70zwQY/81VcvaVkxL5vsCy9asxePjbCe70bZEC+hd0ava9utL0xk4W9cs/DviYVIr3oCEU+Yz2VvXlCP77gURK93lzmveLjnL3AzRA8yogbvXca1b42h5y93y4+voGhPb1QAAu+ryKkvayjwr5AwEi9qWMRPua0s70vFRO/rR49vRJu1j5oPcu9tfHBvqDPGr30PeU9YsHaveOMO74wpBG9GwxFvuZB4r1/aMG+t2chvb/rtT3muvG9VYESv9ggGr1m1rw+4JUEvknYwL4f1Pe8bCGEPZxMDL5lQhK/F0LtvPzrsT4AABi+X2nAvrxStLyXuDs9S7IfvhETEr950Ky8xLipPuZhK7598kO/uwFtvNQQHj/mDju+7vERvwO7CruU9qM+2rtGvuX4v770gYw7PErcPKZpTr7tMzi+fiGeOytkh77EGFK++QnAvlO48rnH1fM8P8dZvhD4Eb8axwo5vwSlPrF0Zb5ZCcC+gI/XO+r68jwlI22+il04vrv/6jvcmYW+GdNwvh4jwL5p9f86PUALPZWCeL7nBhK/JIosOz2Tpz6bGIK+tSvAvgxiFjxtLRE9lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.6537982225418091, 0.7480552792549133, 0.35675248503685, 0.8867892622947693, 0.34405601024627686, 0.8922792077064514, 0.32081732153892517, 0.900262176990509, 0.7113386988639832, 0.6747792959213257, 0.731661856174469, 0.6502748727798462, 0.7509464621543884, 0.6238872408866882, 0.7694880962371826, 0.40483903884887695, 0.8611644506454468, 0.41056203842163086, 0.8615332841873169, 0.594387412071228, 0.7814565896987915, 0.5843990445137024, 0.7871851325035095, 0.5718697309494019, 0.7941381931304932, 0.443391352891922, 0.8467154502868652, 0.5640233159065247, 0.7952790260314941, 0.5560895204544067, 0.7993307113647461, 0.5452324748039246, 0.8047934770584106, 0.5313411355018616, 0.8115816712379456, 0.4857438802719116, 0.822441816329956, 0.4813523292541504, 0.8271824717521667, 0.5324507355690002, 0.8056276440620422, 0.46916747093200684, 0.8332657814025879, 0.44921472668647766, 0.1553802192211151, 0.938803493976593, 0.8697534799575806, 0.35048192739486694, 0.8889293074607849, 0.7061596512794495, 0.6530068516731262, 0.7467408180236816, 0.6007909178733826, 0.7833701372146606, 0.45725783705711365, 0.831234872341156, 0.5036577582359314, 0.831467866897583, 0.44791221618652344, 0.8557592034339905, 0.38761278986930847, 0.8783908486366272, 0.6753438711166382, 0.6768608689308167, 0.7193108797073364, 0.6272778511047363, 0.7571271657943726, 0.5746353268623352, 0.7901135683059692, 0.4809771776199341, 0.18898361921310425, 0.9214778542518616, 0.1946869194507599, 0.9212345480918884, 0.8093040585517883, 0.5057886242866516, 0.8030053973197937, 0.5049028992652893, 0.8032615780830383, 0.4986136257648468, 0.8062254190444946, 0.5137951374053955, 0.787377655506134, 0.4885161519050598, 0.1918935924768448, 0.9216127991676331, 0.8230971097946167, 0.43188774585723877, 0.8354758620262146, 0.6013743281364441, 0.6997705101966858, 0.6221972107887268, 0.6802588701248169, 0.6396864652633667, 0.33783024549484253, 0.1504456251859665, 0.9276463985443115, 0.8611058592796326, 0.3066280484199524, 0.12906235456466675, 0.9389564990997314, 0.8911041021347046, 0.770799994468689, 0.5167937874794006, 0.7597653269767761, 0.5536934733390808, 0.7326628565788269, 0.5852333307266235, 0.29404735565185547, 0.8741856813430786, 0.6928068995475769, 0.3840416669845581, 0.8318026065826416, 0.6462596654891968, 0.3603768050670624, 0.8428529500961304, 0.4279141128063202, 0.8768115043640137, 0.6203874945640564, 0.7618516087532043, 0.6473038196563721, 0.745112955570221, 0.6674724221229553, 0.731564462184906, 0.3179558217525482, 0.9055505990982056, 0.7141088843345642, 0.314890056848526, 0.8963093757629395, 0.3072061538696289, 0.9014476537704468, 0.714926540851593, 0.6900429725646973, 0.7282330393791199, 0.6736318469047546, 0.2557612657546997, 0.918137788772583, 0.7811395525932312, 0.40977048873901367, 0.8699129819869995, 0.4324297606945038, 0.863659679889679, 0.44798779487609863, 0.8595702648162842, 0.5433138608932495, 0.8255105018615723, 0.4877859652042389, 0.843747615814209, 0.5010747313499451, 0.8442225456237793, 0.5334628820419312, 0.8223601579666138, 0.4510672390460968, 0.8635732531547546, 0.411889910697937, 0.8779820799827576, 0.3639357388019562, 0.10682330280542374, 0.9534547328948975, 0.9134367108345032, 0.7707125544548035, 0.41555944085121155, 0.8564978837966919, 0.5332484245300293, 0.8350723385810852, 0.5360636115074158, 0.8027918934822083, 0.41831517219543457, 0.8751989006996155, 0.35682371258735657, 0.8933011889457703, 0.7044357657432556, 0.3348449766635895, 0.8834495544433594, 0.6413818597793579, 0.759468674659729, 0.6051672101020813, 0.21704421937465668, 0.9261196851730347, 0.8165813088417053, 0.504275918006897, 0.813959002494812, 0.46903300285339355, 0.8487949371337891, 0.5728330016136169, 0.777737021446228, 0.5921180248260498, 0.769203782081604, 0.6028798818588257, 0.23415561020374298, 0.9180946350097656, 0.7853320837020874, 0.43054547905921936, 0.8574968576431274, 0.5692688226699829, 0.7983015179634094, 0.451924592256546, 0.849402666091919, 0.5448324084281921, 0.8133159875869751], "advantages": [13.10532283782959, 13.717348098754883, 13.11357307434082, 12.386098861694336, 12.870925903320312, 12.268976211547852, 12.5468111038208, 12.066407203674316, 12.141822814941406, 12.150308609008789, 11.878232955932617, 11.793551445007324, 11.579999923706055, 11.404545783996582, 11.245341300964355, 10.980606079101562, 11.080321311950684, 10.78364372253418, 10.843639373779297, 10.602641105651855, 10.547697067260742, 10.17845630645752, 10.16540813446045, 9.72918701171875, 9.754876136779785, 9.252480506896973, 9.068318367004395, 9.003835678100586, 9.083035469055176, 8.512777328491211, 8.624432563781738, 7.99703311920166, 8.138105392456055, 7.453843116760254, 7.621213436126709, 6.880115509033203, 6.397294998168945, 6.515971660614014, 6.003903865814209, 6.151065349578857, 6.356845855712891, 5.554975509643555, 4.978507995605469, 5.169253349304199, 4.590817451477051, 5.088864803314209, 4.489740371704102, 4.663486957550049, 4.215814113616943, 4.354055881500244, 4.629079818725586, 3.8671088218688965, 4.1565375328063965, 3.4034528732299805, 3.6891298294067383, 3.7238264083862305, 3.053161382675171, 2.3295602798461914, 2.5724568367004395, 1.9142333269119263, 2.086911201477051, 1.526548147201538, 1.599770188331604, 1.6244124174118042, 0.9612365365028381, 0.9450251460075378, 0.3171690106391907, 0.23584073781967163, -0.3332659602165222, -0.5028881430625916, -0.7209151983261108, -1.1251213550567627, -1.6873725652694702, -2.0129201412200928, -2.6844675540924072, -3.4554150104522705, -3.9211435317993164, -4.36565637588501, -4.60673189163208, -5.413742542266846, -5.6569695472717285, -6.502972602844238, -6.857282638549805, -7.545830726623535, -7.813848495483398, -7.862382888793945, -8.984331130981445, -9.98784351348877, -10.111977577209473, -11.2002592086792, -11.644142150878906, -12.368237495422363, -12.832771301269531, -13.56917667388916, -14.060062408447266, -13.622071266174316, -12.163074493408203, -14.433777809143066, -16.19291114807129, -15.418161392211914, -13.598258972167969, -16.11433982849121, -18.21383285522461, -19.677331924438477, -20.355953216552734, -21.185104370117188, -21.927322387695312, -22.685121536254883, -23.491466522216797, -23.77059555053711, -25.13155174255371, -25.83732795715332, -25.3378849029541, -27.16590690612793, -28.12954330444336, -28.48270606994629, 10.988126754760742, 10.701079368591309, 10.53136920928955, 11.316716194152832, 10.090356826782227, 10.812117576599121, 9.648663520812988, 10.314984321594238, 9.204715728759766, 9.106141090393066, 8.722440719604492, 9.25240421295166, 10.888703346252441, 8.776987075805664, 10.434324264526367, 8.351195335388184, 7.403763771057129, 7.910619258880615, 6.9539289474487305, 7.478797912597656, 9.292895317077637, 7.13918924331665, 6.120083332061768, 6.030979156494141, 5.648583889007568, 5.506397724151611, 5.166876792907715, 4.978718280792236, 4.672268867492676, 5.469540596008301, 4.21985387802124, 3.95556902885437, 3.720889091491699, 4.641626358032227, 3.2715249061584473, 2.915140151977539, 2.7774789333343506, 3.8700737953186035, 2.345771551132202, 3.581430196762085, 1.9568488597869873, 3.3888776302337646, 6.570765495300293, 3.4980804920196533, 1.574702501296997, 0.7716585397720337, 1.068275809288025, 0.34088781476020813, 1.334826946258545, 0.08975276350975037, 0.013926140032708645, -0.17896458506584167, 1.4879997968673706, -0.21621160209178925, 1.8389824628829956, -0.05637194961309433, -0.8427844047546387, -0.4693267345428467, -1.1085247993469238, 0.2136649638414383, -1.1392728090286255, 0.5177525281906128, 3.271544933319092, 1.1374188661575317, -0.5443050861358643, -1.4161131381988525, -0.27429434657096863, 1.9518009424209595, 0.21473242342472076, -1.1469491720199585, 0.5348289012908936, -0.9374476671218872, 0.7671730518341064, -0.7198417782783508, -1.671157956123352, -0.7234440445899963, 0.6133512854576111, 1.7369062900543213, 0.6506044864654541, -0.5637465119361877, 0.42721620202064514, 1.1776400804519653, 0.3433816134929657, -0.6211967468261719]} +{"type": "SampleBatch", "eps_id": [1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 1411378167, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAJsYgr61K8C+DGIWPG0tET2F8IW+6RESv0j/ITzveak+R8iLvm5PwL4kO4c8z9EpPeigj77fDzm+GAaOPGXfe76qepG+i8JnPMhySzx1ewi/lVWRvhd/Ob4zAuY64T1yvnQwk77ZqcC+qhBDu6YgaD3kCpe+okUSv4+R8bpWY7I+t+Scvs+gwL7M8ac7lOVhPfm+oL61SBK/iBbMO5vqsj7smKa+E7bAvu+MWDy/knA9mnOqvs5aEr/fy2s8mgu2PkZOsL6IW0S/FyewPDAaJz/6KLi+EnwSv5GMDT1o27s++gS+vqhTwb4xmys91LuuPdDiwb4BwxK/d5gyPbomyD6nwce+ivPBvqmeUj00/OU9r6LLvrDjPL610Vs9sqgnvj2Gzb73t8K+DWhOPe/gFD4za9G+M2g+vhZRWj2MHwa+pFLTvv03CTw+lk89VxHQvq88077pm08+yksuPTnDLr81KdG+Fz7nO3m+7DyC+MC+tRbRvoT8QL5Q/q487BiavcEE076OjMS+ZqqiPBx7ZT4V89a+6ZZBvu9hxzxQ+369rOLYvokvuzvsLr08kb6xvrPT2L53T00+Gk6EPIDpIb8axta+wuqoO2BzZTuTY6u+l7jWvruTQr4aTlG7xMQnvbWq2L7eKsW+2n6Du6RdgD40nNy+LYUUvzFAgzr2yAo/DY3iviwlxb4YDUI8hcF/Pm9+5r4oqUK+T/KJPGBoIL3EcOi+cZGcO7yHgzwFKae+PmTovrd1TD7vExw8IjMdv9NY5r6go5A7G400u3gEo75BTea+a11MPg54FbxcqBy/FELkvlUjljvR/q68n/CkvhE25L6Fy0K+ssbjvA6YFL2+KOa+GxfFvky46bzKcX0+2Bnqvp5jFL82K8G8HvMHP1kJ8L7LtcS+b1JUvFuUbD6B+PO+HkIUv9edCLyhBAU/q+b5vvSNxL4YlgY7s69lPgfV/b5pPBS/20rWO5qDBD984QG/8Z3EvtFhijxtd2g+0tgDvy2uQb6mk688ePF2vbzQBL/6s7k787KlPGM4sb5OyQS/S0pCviH6WTxMF0G9/8EFv9gkxb6gh0o8zLN/Pq+6B796qkK+YS2OPOrxH73bswi/UFzFvovHhzxbo4Q+Gq0Kv58pQ75CObI8Uz/ovOmmC7+wHIo7JpStPHzQoL5ioQu/kspDvng8dDzKgXK8/5sMv+nnxb7WYm88CqqQPqOWDr8yOUS+T/ylPOP0s7vNkQ+/l8pRO/cVpTx+W5W+m40Pv2XISj43lWo8EPYTvwuKDr/4oDE7LsU0O+7Gj75+hg6/0hFFvpBMO7vwrlA8voIPv4xqxr66miq7oumbPrB+Eb9Y/US+LYhkO2R0NDzVehK/O2zGvuP3cjsdDpw+y3YUv08YRb4MniA8QqRZPBRzFb+qISQ7X/gkPON1jb7LbxW/ijhKPvfelDv61hC/82wUvxR3Fjs26927RBiLvvFpFL+vWkW+5fpHvKOUmjyOZhW/Jn3Gvv3LQbzNh50+sGIXv5cARb5l9Lm7GO04PNpeGL8Lqzg7wI6yuxr9kL4oWxi/5NZEvlISNrw/4f47HFcZv1o9xr7UhTO8ygeYPptSG78phES+U3Kku9XpVTolThy/oB/Gvmzpo7tndZU+V0kev6D+FL9iKVs6TkAVPzJEIb/EF8a+N71MPJvIlD5QPyO/xZBEvvb6lTyZD/Y66jokv3XNPTuzSZY8CeiRvh83JL8nGkW+GTJPPO4tXDxpMyW/r/AgO2qZUzyg6oy+MTAlv+MgSj5Q0/I7zVUQv3gtJL+96cg+XVh9u3epWr8hKyK/SxNKPqCcq7wGDBC/eSghvwX7JDuk5gO9IaiNviwlIb9v1ko+5pAavWVPFL+LISC/XwxmO3cGSr0n+Zi+8Rwgv8p0Q75AgGK9apy0vCAXIb//M8W+nU5kvWdggT73DyO/ZNVBvlqbT71VpGm9Eggkv9htxL6ZR1S9jn1gPu7+Jb/dVUC+CFJCvYf4tr0e9Sa/xLTDvqejSb1EhEA+IOoovz7sPr7qPDq90VX1vYLeKb9cSRg8JQ1EvdNZ2r5T0im/e489vsz8Zr2Y1Bi+9sQqvzo/wr7ENnO9dyAAPjy2LL8p3zu+uvZovVIkPr62pi2/J2bBvtUseL2WTbU90JUvv2VqEr9K7HC95BW5PnaDMr/vhsC+ME9TvUlTUD1UcDS/GwQSv5AkT71iVqc+7ls3vyLGQ79sXjS90WcaP0tGO78hrBG/hfUCvY8GmD4jMD6/cze/vhpF1bzEtrm7pxlAv49BNr7RMta88uacvvACQb+21b6+HjQEvbVBtbx660K/B2g1viMEBr0vTqa+rdNDv1jgljwBoCC9oawgv4m7Q7/xXDS+cgpUvbPysb5nokS/KrC9vjODcL1WtZK9AYhGvw7DMr5+YXa9WrbDvtFsR78l0ry+79iKvSV8370zUEm/mukwvi1Rj70RRdi+fRsKPUwbFb23BOS84esDvRAgBz3n7CI+l0vpvNrFqr7GKBQ9DpG1Pqf4D724kyK/vjUxPRXZIz72/kO9ghK1vllRPj1i6Qy9r/dgvaiLnL3hfzs9LC9qvrY6Z71ZL0s+ycMoPXtnBr17+Va90GTkvaMTJj3gmGi+PBxgvcYXKD4MeBM9lyQAvbGpUr1SwxS+9ecQPZsMZ75ckF69u+IFPiLY/DwCAte+YdpTvbHFzz6xCrg8FYBlvgecMr0AOMc9XFKTPFms6LwHpCq9TtpVvhKrjjwlQWS+vL87vcc1kD2JS1Q8q63VvgX7Nb02U7I+8BSXO0LvYr7Uchm98sQrPR3/OjmwFdW+XgMWvY8npT7QcwW89Vocv0ct97yefRo/KcumvFGZ1L6GTZS8TFeaPkPT6rz1Lhy/0NNFvCyfFj8tZCe9QRdOv3ICobmUkWA/JFdpvXgjHL9LNY08/J0VPwGnjb0dgNS+j/bsPKEvmD4Dp569VaFhvtLUDj3FX2I8eK2nvej21L6U9g89dXSiPvu2uL1/p2K+vPQpPdELEz3rx8G9Wd3bvJzlLD38P3m+WeHCvebgY7749Bg9Uh5/PdL+y703IuW8LQ8ePaagX74dJM29M/1kvkkrDD2LkbA99kzWvayj1r5ZOxM972vHPsZ4572/A2a+qCMzPTjr3T0grPC9WjfXvhsEPD0IO9Q+4PEAvtFSZ74S+V09Pe0LPkGSBb4kawG9xSppPQzIDb7pNwa+8fFovhbTXT3Tvi8+luAKvnLmB71a4ms9sfPTvYqOC74Vk2q+92djPZvAUz6QPxC+zYAOvadYdD2rAou99/UQvu9AIz4vyW49JqyuvhuyDb4xZRW9mtZSPdlK+7xVcQ6+uKkhPktTUD2kBZ2+njULvjxYtD6sMzc9GzkVv+P+A77jRSA+UXMHPcSNjb5KygC+mf4fvY2a4Twut9g8FZcBvhdkHz4i8OU8vMuDvgHO/L0AQCO9a8O7PG8tND3tb/699aUePnD4wjwjK3e+Xhf4vWcMsz5qbJs858UGv3TE6b0jBh4+ZVYKPExTab5KcuO9JdqyPoWxfjvEkgS/ZiPVvfTXHT57CtS7nlVlvhTTzr0L37I+UWgzvPzHBL/Mg8C95hgePgevrryp+Gq+4TC6vWGOJb10R9S8eRVnPbTYu73IzXC+JwnLvAV/rj6JesW9kR7cvm0yk7zlJSA/lBbXvY0ucL4Zz7K7rpKnPguy4L3/MCG917yOOva5Bj2xTuK9s30fPmr25Do71oS+ge3bvX5bIb2+lGG7qWMKPZSK3b0ShB8+5Es1uyochb4jKde9ZP4gvbiDAryzXgI9SMXYvSmsHz58K/C7u9aGvjxi0r0PGSC9u2FOvPM53TwV/NO9X6Rvvl+IRbzDoaE+BZLdvWinHr00Lby7Q4GdPCwo372qOSA+gZOvuzjvjL53v9i99QAevXT8Mbzk0oA89VPavTV2ID5M1Sy8AYyPvtTo073vvxy9AlqEvFYAEzwbetW90dkgPq/hgrzi2JO+/grPvULeGr1UMbK8imCYunWX0L0YMm6+FmKyvE+0kT6VHtq9vVEYvf3Bg7wO/HO8haTbvQ+lbb6XMoa84JuLPgAm5b0Daxa9qQszvAnfzbwSp+a97lsiPshHO7zEeaS+gyjgvRQqtT68RZK88xQev0Sq0b2syCI+3XH3vG03qb5cJ8u9TEUSvQrMFr2EiUK90J3MvefBIz4SsBq97f+zvvAQxr2x6A293Hw3vWd0kb05fMe9e/QkPlBOPb0pR8G+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 0, 0, 0, 0, 1, 1, 1, 1, 0, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 0, 0, 0, 0, 1, 1, 1, 1, 0, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAIXwhb7pERK/SP8hPO95qT5HyIu+bk/AviQ7hzzP0Sk96KCPvt8POb4YBo48Zd97vqp6kb6Lwmc8yHJLPHV7CL+VVZG+F385vjMC5jrhPXK+dDCTvtmpwL6qEEO7piBoPeQKl76iRRK/j5HxulZjsj635Jy+z6DAvszxpzuU5WE9+b6gvrVIEr+IFsw7m+qyPuyYpr4TtsC+74xYPL+ScD2ac6q+zloSv9/LazyaC7Y+Rk6wvohbRL8XJ7A8MBonP/oouL4SfBK/kYwNPWjbuz76BL6+qFPBvjGbKz3Uu6490OLBvgHDEr93mDI9uibIPqfBx76K88G+qZ5SPTT85T2vosu+sOM8vrXRWz2yqCe+PYbNvve3wr4NaE497+AUPjNr0b4zaD6+FlFaPYwfBr6kUtO+/TcJPD6WTz1XEdC+rzzTvumbTz7KSy49OcMuvzUp0b4XPuc7eb7sPIL4wL61FtG+hPxAvlD+rjzsGJq9wQTTvo6MxL5mqqI8HHtlPhXz1r7plkG+72HHPFD7fr2s4ti+iS+7O+wuvTyRvrG+s9PYvndPTT4aToQ8gOkhvxrG1r7C6qg7YHNlO5Njq76XuNa+u5NCvhpOUbvExCe9tarYvt4qxb7afoO7pF2APjSc3L4thRS/MUCDOvbICj8NjeK+LCXFvhgNQjyFwX8+b37mviipQr5P8ok8YGggvcRw6L5xkZw7vIeDPAUpp74+ZOi+t3VMPu8THDwiMx2/01jmvqCjkDsbjTS7eASjvkFN5r5rXUw+DngVvFyoHL8UQuS+VSOWO9H+rryf8KS+ETbkvoXLQr6yxuO8DpgUvb4o5r4bF8W+TLjpvMpxfT7YGeq+nmMUvzYrwbwe8wc/WQnwvsu1xL5vUlS8W5RsPoH4874eQhS/150IvKEEBT+r5vm+9I3EvhiWBjuzr2U+B9X9vmk8FL/bStY7moMEP3zhAb/xncS+0WGKPG13aD7S2AO/La5BvqaTrzx48Xa9vNAEv/qzuTvzsqU8Yzixvk7JBL9LSkK+IfpZPEwXQb3/wQW/2CTFvqCHSjzMs38+r7oHv3qqQr5hLY486vEfvduzCL9QXMW+i8eHPFujhD4arQq/nylDvkI5sjxTP+i86aYLv7AcijsmlK08fNCgvmKhC7+SykO+eDx0PMqBcrz/mwy/6efFvtZibzwKqpA+o5YOvzI5RL5P/KU84/Szu82RD7+XylE79xWlPH5blb6bjQ+/ZchKPjeVajwQ9hO/C4oOv/igMTsuxTQ77saPvn6GDr/SEUW+kEw7u/CuUDy+gg+/jGrGvrqaKrui6Zs+sH4Rv1j9RL4tiGQ7ZHQ0PNV6Er87bMa+4/dyOx0OnD7LdhS/TxhFvgyeIDxCpFk8FHMVv6ohJDtf+CQ843WNvstvFb+KOEo+996UO/rWEL/zbBS/FHcWOzbr3btEGIu+8WkUv69aRb7l+ke8o5SaPI5mFb8mfca+/ctBvM2HnT6wYhe/lwBFvmX0ubsY7Tg82l4YvwurODvAjrK7Gv2QvihbGL/k1kS+UhI2vD/h/jscVxm/Wj3GvtSFM7zKB5g+m1IbvymERL5TcqS71elVOiVOHL+gH8a+bOmju2d1lT5XSR6/oP4Uv2IpWzpOQBU/MkQhv8QXxr43vUw8m8iUPlA/I7/FkES+9vqVPJkP9jrqOiS/dc09O7NJljwJ6JG+HzckvycaRb4ZMk887i1cPGkzJb+v8CA7aplTPKDqjL4xMCW/4yBKPlDT8jvNVRC/eC0kv73pyD5dWH27d6lavyErIr9LE0o+oJyrvAYMEL95KCG/BfskO6TmA70hqI2+LCUhv2/WSj7mkBq9ZU8Uv4shIL9fDGY7dwZKvSf5mL7xHCC/ynRDvkCAYr1qnLS8IBchv/8zxb6dTmS9Z2CBPvcPI79k1UG+WptPvVWkab0SCCS/2G3EvplHVL2OfWA+7v4lv91VQL4IUkK9h/i2vR71Jr/EtMO+p6NJvUSEQD4g6ii/Puw+vuo8Or3RVfW9gt4pv1xJGDwlDUS901navlPSKb97jz2+zPxmvZjUGL72xCq/Oj/CvsQ2c713IAA+PLYsvynfO7669mi9UiQ+vramLb8nZsG+1Sx4vZZNtT3QlS+/ZWoSv0rscL3kFbk+doMyv++GwL4wT1O9SVNQPVRwNL8bBBK/kCRPvWJWpz7uWze/IsZDv2xeNL3RZxo/S0Y7vyGsEb+F9QK9jwaYPiMwPr9zN7++GkXVvMS2ubunGUC/j0E2vtEy1rzy5py+8AJBv7bVvr4eNAS9tUG1vHrrQr8HaDW+IwQGvS9Opr6t00O/WOCWPAGgIL2hrCC/ibtDv/FcNL5yClS9s/KxvmeiRL8qsL2+M4NwvVa1kr0BiEa/DsMyvn5hdr1atsO+0WxHvyXSvL7v2Iq9JXzfvTNQSb+a6TC+LVGPvRFF2L6mMkq/NNG7vmKeoL2ZPBy+ECAHPefsIj6XS+m82sWqvsYoFD0OkbU+p/gPvbiTIr++NTE9FdkjPvb+Q72CErW+WVE+PWLpDL2v92C9qIucveF/Oz0sL2q+tjpnvVkvSz7Jwyg9e2cGvXv5Vr3QZOS9oxMmPeCYaL48HGC9xhcoPgx4Ez2XJAC9salSvVLDFL715xA9mwxnvlyQXr274gU+Itj8PAIC175h2lO9scXPPrEKuDwVgGW+B5wyvQA4xz1cUpM8WazovAekKr1O2lW+EquOPCVBZL68vzu9xzWQPYlLVDyrrdW+Bfs1vTZTsj7wFJc7Qu9ivtRyGb3yxCs9Hf86ObAV1b5eAxa9jyelPtBzBbz1Why/Ry33vJ59Gj8py6a8UZnUvoZNlLxMV5o+Q9PqvPUuHL/Q00W8LJ8WPy1kJ71BF06/cgKhuZSRYD8kV2m9eCMcv0s1jTz8nRU/AaeNvR2A1L6P9uw8oS+YPgOnnr1VoWG+0tQOPcVfYjx4rae96PbUvpT2Dz11dKI++7a4vX+nYr689Ck90QsTPevHwb1Z3du8nOUsPfw/eb5Z4cK95uBjvvj0GD1SHn890v7LvTci5bwtDx49pqBfvh0kzb0z/WS+SSsMPYuRsD32TNa9rKPWvlk7Ez3va8c+xnjnvb8DZr6oIzM9OOvdPSCs8L1aN9e+GwQ8PQg71D7g8QC+0VJnvhL5XT097Qs+QZIFviRrAb3FKmk9DMgNvuk3Br7x8Wi+FtNdPdO+Lz6W4Aq+cuYHvVriaz2x89O9io4LvhWTar73Z2M9m8BTPpA/EL7NgA69p1h0PasCi7339RC+70AjPi/Jbj0mrK6+G7INvjFlFb2a1lI92Ur7vFVxDr64qSE+S1NQPaQFnb6eNQu+PFi0PqwzNz0bORW/4/4DvuNFID5Rcwc9xI2NvkrKAL6Z/h+9jZrhPC632DwVlwG+F2QfPiLw5Ty8y4O+Ac78vQBAI71rw7s8by00Pe1v/r31pR4+cPjCPCMrd75eF/i9ZwyzPmpsmzznxQa/dMTpvSMGHj5lVgo8TFNpvkpy470l2rI+hbF+O8SSBL9mI9W99NcdPnsK1LueVWW+FNPOvQvfsj5RaDO8/McEv8yDwL3mGB4+B6+uvKn4ar7hMLq9YY4lvXRH1Lx5FWc9tNi7vcjNcL4nCcu8BX+uPol6xb2RHty+bTKTvOUlID+UFte9jS5wvhnPsruukqc+C7Lgvf8wIb3XvI469rkGPbFO4r2zfR8+avbkOjvWhL6B7du9flshvb6UYbupYwo9lIrdvRKEHz7kSzW7KhyFviMp171k/iC9uIMCvLNeAj1Ixdi9KawfPnwr8Lu71oa+PGLSvQ8ZIL27YU688zndPBX8071fpG++X4hFvMOhoT4Fkt29aKcevTQtvLtDgZ08LCjfvao5ID6Bk6+7OO+Mvne/2L31AB69dPwxvOTSgDz1U9q9NXYgPkzVLLwBjI++1OjTve+/HL0CWoS8VgATPBt61b3R2SA+r+GCvOLYk77+Cs+9Qt4avVQxsryKYJi6dZfQvRgybr4WYrK8T7SRPpUe2r29URi9/cGDvA78c7yFpNu9D6Vtvpcyhrzgm4s+ACblvQNrFr2pCzO8Cd/NvBKn5r3uWyI+yEc7vMR5pL6DKOC9FCq1PrxFkrzzFB6/RKrRvazIIj7dcfe8bTepvlwny71MRRK9CswWvYSJQr3Qncy958EjPhKwGr3t/7O+8BDGvbHoDb3cfDe9Z3SRvTl8x7179CQ+UE49vSlHwb4W48C9JJcIvfs6XL2MQsy9lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.5199524164199829, 0.828216552734375, 0.5140385627746582, 0.18035228550434113, 0.9324020743370056, 0.8260346055030823, 0.4908517897129059, 0.8378863334655762, 0.47002214193344116, 0.8485099077224731, 0.43999046087265015, 0.1383698731660843, 0.9451899528503418, 0.8839677572250366, 0.3263181447982788, 0.9009080529212952, 0.7320062518119812, 0.6222200989723206, 0.7733772397041321, 0.43819767236709595, 0.16584333777427673, 0.9304763078689575, 0.8318285942077637, 0.5231486558914185, 0.8209273815155029, 0.5099197626113892, 0.1928052008152008, 0.9249619841575623, 0.8177680969238281, 0.5048879384994507, 0.17730273306369781, 0.932253360748291, 0.8360740542411804, 0.5454598665237427, 0.2143557071685791, 0.9181556701660156, 0.19947047531604767, 0.9252221584320068, 0.8249421119689941, 0.5356269478797913, 0.20226994156837463, 0.9230884909629822, 0.20110706984996796, 0.9242172241210938, 0.19192714989185333, 0.9272116422653198, 0.823756217956543, 0.5338760018348694, 0.7814702391624451, 0.45769694447517395, 0.8371107578277588, 0.4346748888492584, 0.8461841940879822, 0.592106819152832, 0.7362754344940186, 0.3974513113498688, 0.8586617708206177, 0.6268390417098999, 0.29485592246055603, 0.883057177066803, 0.7175191044807434, 0.3823699951171875, 0.8604565262794495, 0.3734237849712372, 0.8636725544929504, 0.6407763361930847, 0.3074577748775482, 0.8782146573066711, 0.7121286392211914, 0.38472887873649597, 0.8560131788253784, 0.6158082485198975, 0.7195335030555725, 0.39733144640922546, 0.8498152494430542, 0.396481454372406, 0.14997293055057526, 0.9339860677719116, 0.8583023548126221, 0.6429347991943359, 0.673652708530426, 0.6506326794624329, 0.33214858174324036, 0.13844002783298492, 0.9359982013702393, 0.8820406794548035, 0.26333263516426086, 0.8994787335395813, 0.7817967534065247, 0.5110031366348267, 0.7817914485931396, 0.5555487275123596, 0.7534081339836121, 0.5966616272926331, 0.7227045893669128, 0.36503785848617554, 0.8674027919769287, 0.6905041933059692, 0.6285871863365173, 0.7324374914169312, 0.4257999360561371, 0.8285356163978577, 0.4654594361782074, 0.19057074189186096, 0.9185949563980103, 0.8013376593589783, 0.4992280900478363, 0.7963412404060364, 0.46718254685401917, 0.18560022115707397, 0.9243147373199463, 0.8452135324478149, 0.35927814245224, 0.868760347366333, 0.30183133482933044, 0.8911712169647217, 0.415315181016922, 0.12365061789751053, 0.9488186240196228, 0.8980846405029297, 0.716732919216156, 0.659044623374939, 0.7603343725204468, 0.6008750796318054, 0.7970541715621948, 0.46184778213500977, 0.8372563719749451, 0.4994930326938629, 0.8425098061561584, 0.5560781955718994, 0.7909103631973267, 0.5886493921279907, 0.226023867726326, 0.9201210141181946, 0.2205154150724411, 0.07576221227645874, 0.9616395831108093, 0.935249388217926, 0.8421018123626709, 0.4603724479675293, 0.8659160733222961, 0.602634847164154, 0.768596887588501, 0.6461200714111328, 0.739647626876831, 0.31555724143981934, 0.9092029333114624, 0.2605026662349701, 0.9231515526771545, 0.7939956188201904, 0.5552677512168884, 0.8310684561729431, 0.4811696708202362, 0.8614320755004883, 0.5957092046737671, 0.7622209191322327, 0.6478646397590637, 0.27306675910949707, 0.9024930596351624, 0.714667022228241, 0.698287844657898, 0.6941055059432983, 0.7173634171485901, 0.3238767385482788, 0.892363429069519, 0.31465864181518555, 0.8983293175697327, 0.29090261459350586, 0.9077141880989075, 0.7451168894767761, 0.3576626479625702, 0.1131540983915329, 0.9507300853729248, 0.8918394446372986, 0.6574879884719849, 0.7465486526489258, 0.6497708559036255, 0.7548961043357849, 0.6366959810256958, 0.766804039478302, 0.3823505938053131, 0.8815728425979614, 0.6185122132301331, 0.778278648853302, 0.6005059480667114, 0.7919353246688843, 0.5752210021018982, 0.8085626363754272, 0.4582545757293701, 0.8534752130508423, 0.47200465202331543, 0.8493980169296265, 0.5202584862709045, 0.16641069948673248, 0.9388980269432068, 0.8601585626602173, 0.41317588090896606, 0.8800144791603088, 0.3510757386684418, 0.8989355564117432], "advantages": [13.86804485321045, 13.175607681274414, 13.719236373901367, 14.118341445922852, 14.801876068115234, 14.30284309387207, 14.008076667785645, 13.588269233703613, 13.82448959350586, 13.508691787719727, 13.582075119018555, 13.360777854919434, 13.023870468139648, 12.973896980285645, 12.651301383972168, 12.653926849365234, 12.162520408630371, 11.682069778442383, 11.809429168701172, 11.227328300476074, 10.974092483520508, 11.448655128479004, 10.89777660369873, 10.953437805175781, 11.128548622131348, 10.45494556427002, 10.011300086975098, 10.346689224243164, 9.921895027160645, 10.106973648071289, 10.3045015335083, 10.285378456115723, 9.615755081176758, 8.806126594543457, 8.1491060256958, 8.309440612792969, 7.983579635620117, 8.211711883544922, 7.857133388519287, 8.145045280456543, 8.392200469970703, 8.293962478637695, 7.687145233154297, 7.6393232345581055, 6.946270942687988, 6.941007137298584, 6.160737991333008, 5.158932209014893, 4.272641658782959, 4.648043155670166, 4.963296890258789, 3.9142699241638184, 4.2271246910095215, 3.1382246017456055, 2.133297920227051, 2.4980688095092773, 2.7819325923919678, 1.6436471939086914, 0.5795941948890686, 0.28523167967796326, 0.0589417926967144, 0.3272989094257355, 0.5336767435073853, -0.6180878281593323, -0.4089493155479431, -1.5944722890853882, -2.663055658340454, -2.961348533630371, -3.3038089275360107, -3.1845593452453613, -3.0634560585021973, -4.240775108337402, -5.215975284576416, -5.176181793212891, -5.092827320098877, -6.2793402671813965, -6.199551582336426, -6.284739017486572, -7.427353382110596, -8.703426361083984, -9.742770195007324, -9.774018287658691, -10.799638748168945, -11.069645881652832, -9.942083358764648, -11.582765579223633, -12.444509506225586, -12.274998664855957, -13.322125434875488, -13.908228874206543, -14.255329132080078, -15.217801094055176, -15.626358032226562, -16.52485466003418, -17.006019592285156, -17.83319091796875, -17.938188552856445, -18.991668701171875, -19.71113395690918, -20.28329849243164, -21.123502731323242, -21.606306076049805, -22.64682388305664, -23.177936553955078, -23.443754196166992, -24.81296157836914, -25.845666885375977, -26.1445369720459, -27.323701858520508, -27.49441146850586, -26.639259338378906, -28.63787269592285, -30.18505096435547, -29.91583251953125, -31.645811080932617, -31.13223648071289, 14.471434593200684, 15.03508472442627, 17.12688636779785, 15.078790664672852, 14.167634010314941, 14.203629493713379, 14.005565643310547, 13.904351234436035, 13.906237602233887, 13.652503967285156, 14.492643356323242, 13.354822158813477, 13.834983825683594, 13.211174964904785, 13.724588394165039, 13.037026405334473, 13.361952781677246, 14.903573989868164, 12.935162544250488, 14.34245777130127, 16.584243774414062, 13.7691068649292, 12.052897453308105, 11.985196113586426, 11.721636772155762, 11.650335311889648, 13.043719291687012, 11.345305442810059, 12.73244857788086, 11.027462005615234, 10.773356437683105, 10.599921226501465, 10.376823425292969, 10.099592208862305, 11.247505187988281, 9.597785949707031, 10.61972427368164, 9.061104774475098, 9.93366527557373, 11.822427749633789, 9.26461410522461, 11.083911895751953, 13.52344799041748, 10.450069427490234, 8.047859191894531, 9.800442695617676, 7.450640678405762, 9.151045799255371, 11.682793617248535, 8.617552757263184, 11.2544584274292, 8.196488380432129, 10.96518611907959, 7.911034107208252, 5.662945747375488, 4.671236515045166, 5.125830173492432, 4.045373439788818, 4.580812931060791, 6.474283695220947, 4.0751471519470215, 6.000354290008545, 3.5899806022644043, 5.561758518218994, 3.132011651992798, 1.9265995025634766, 2.5898499488830566, 4.613146781921387, 2.1183156967163086, 4.202277183532715, 1.6860337257385254, 3.8475253582000732, 1.3056964874267578, -0.12257706373929977, 0.8492254018783569, -0.6747030019760132, 0.364379346370697, 2.6335201263427734, 5.519772529602051, 2.5334067344665527, -0.07665462046861649, 2.468492269515991, -0.12134800106287003, 2.5339698791503906]} +{"type": "SampleBatch", "eps_id": [53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 53199550, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAABbjwL0klwi9+zpcvYxCzL3CQMK9dR9pvppmZL31uDM+75PLvZUI2L7jBVa9tnLmPk3c3L38i2e+uyYxvYPNED5XH+a9Nxb5vCyRJb2SjCi+LF7nvTZTZr4ODTO9SKbrPbOU8L3GQu+8AaApvZerQ770xvG94BRlvlVHOb3BvbQ9wPD6vcgY1r6LDDK9V4+7ProIBr7gyGO+FgoUvfHUdj38lgq+/b/bvE8aD72/cnm+nyMLvsi+Yr4CDyO9oRIbPY+sD763+NS+CfUfvV2soj5jMRi+cZdhvvDtBb2StlQ8arQcvoKUyryq3QS9L2WUvhE2Hb6Eo2C+75scvZaY97s4tCG+Qe7TvmU6Hb1ir4s+Yy4qvoyEX77l4Aa9uOgBvcymLr5sadO+B3oJvS80gD6oGze+eYtevpnt6bygxle9Fo87vpr10r4lj/K8OmZsPk//Q77IURu/OrzMvDmZAz9AbFC+mI/Svj8GcbwLu1o+ZdhYvlMtG7/VByu8jGkAP2xCZb4YF02/NTbVuQZ+Sj+rqnW+NyMbv/qGfDy/EP8+8QmBvlR80r5z4s88I3BXPqE/hb4XjF2+ylryPPzsl73Kdoe+Z4OxvFoz5jxg+7a+mK+HvjBkXr54pas8FUhlvevoib6pP9O+oXmiPI0UeT6DIo6+/fxevvVTyjyunTC9XF2QvseS075oQ8M8praDPp6YlL6p0hu/YGntPP6uDj8+1Jq+KvXTvlNdJD1yQow+dxGfvhESHL9bzjo97jkUP6FPpb6klNS+Dj1qPTwhmj4KkKm+nDxivhtzgT3uitw8NdOrvlN01b5mjYI94H6tPhgYsL7aFmS+l26QPXkXiT0AYLK+M23WvoAskz0YFMM+3qm2vh0qZr61x6I9twTlPRb3uL6J5vy8SFynPZZ5Hr4ESLm+JoNovoAFoT2ScSY+P5u7vuqv2L7irac90Tn1Pq/wv75L2Gq+G0y7PcJGWj7jScK+5O3ZvkIHxD3+gQg/r6XGvl6Obb6d3tk9C1uLPtQFyb5R1B29ngTlPa9ZczzXasm+4IkePl2g5T1+VXa++9THvpNEsj7nxds95OP9vkBExL63XBs+O3bHPd6CL76GtsK+Rc6wPv9wwD3GGd2+Ry2/vu2UGD7bwK49wHrjvaumvb4Hc0G9KTSqPcA8VD56Ir6+kyEWPnixsj3C7229JKK8vrdYS71aULA9F4eFPkgkvb7SmxM+AP+6Pe3vZ7tnqru+z/qsPuTZuj3zOoi+vzS4vlcUCD/l8689OlUHv0vDsr7gsas+qEyaPcA3V743VK++sJEOPtSwkT1go9c9Peetvkyhqj7lAJY9rxAovp19qr62+gY/6EePPVKc3b5tF6W++Kg4P6Uaez3/4TO/grSdvnKBBj+mikE93FzIvitTmL5KyKg+ynshPQdPrL0B85S+NTEGP1mXGj1jbrq+4JSPvkYAOD9Ahvk8JQklv7Q4iL7U8QU/zeaPPDVjr74c3YK+dtynPhuOLzyyyzW9UgN/vpHXBT/uAiE8X9aqvjxOdL71tac+urNOOyg+G73hmG2+DZIHPj0GHTv79II+wuJqvo6rpz4YI/Y7/RIUvdEtZL4JbQc++nHeO46NhD5veGG+wZCnPnYORDyWmgG9kMRavoOwBT8tsDk83RukvpoSUL5TmTc/KlGhO48PHL9/YkG+U6EFP9Uy7rvheaG+wLE2vkxppz6ocV68rM/MvHb/L76/tAU/7aJmvOjUpL4pTSW+NrU3P3kQqLx3eB6/05oWvhq7aT8lvga9hQhrvwToA7655Tc/EvRRvYXMIr+YY+q9igVqP0cGg70nvnG/EfLEvSdOOD8QtKm9+CEsv+x0p70lrAY/nj7FvS9W0L696JG9UzSqPlrp1b2SlBW+9EqEvcBkBz8O5du9M4nwvnhCXb0Kvas+OyPvvUXmWb4QyEG9AZARPoba970lgic98SI2vct0rT6zLfa9ZSSTvi1iGr3fDRU+lvkAvmf1Dr2MdQ698sRCvZOwAb5Ftlw+xFoSvTG2GD4Ojfq9pQDqvTsjBr3kZjS9Jjv/vZTCDD7jvgm9IkkcPsSZ+b31ZUS+UHz6vHsmJr1xugC+2yN2PdmQAL3nSW++xP7+vYYJnz56tRO9TYnavq9F8r09xA8/uqw2vf3Ja74BRdu9SbhvPrCJSb0cZAq9Rq7RvScFub0/Tky9SsJovpRh1b3qZyw+J+1evQly174kfM69SFfaPuaygL1iyGW+hAS9vdp/1D3g44m9Dw/WvoXEuL2vXrs+zAObvfCfHL8wx6m9kP4gPyYTtL3RztS+2wSQvRxhnz50GcW94xQcv8REg73PuxQ/jhLevRHV075H8Va9aqOJPuEE7713Il++nOtAvbiUI73H8fe9uhvTviUxRL3hOHM+pGoEvui+Xb70uzC9lRuPvfrZCL4HctK+YXU2vVvmVT7wRBG+snZcvrVYJb2rsse9h2y1PHszb7xidE87A0CWPCwIszzlzjg+oH5nO2jzi77qmdA8I8VwvBKP/bohjqc8izHOPBXTVr7z8Me6VgOgPlTSqzyL9W+81dSaO7ycnjwJbKk827w4PjaFpzvZLIu+5PrGPH9Dcrxj+6m5Xge4PLKOxDwMqzg+VSYDOXtnir60GuI8fDdyvBkPrbubg7c8oa7fPFW+OD67YJ67GzyLvrk9/TwwT8A+mEwovPjuEL/aYx09fvY4PjLosLzirY2+5C8sPThRaLyPPt68LrgUPIYGKz0orDk+1sHcvFaGlb4Z4Tk9GKhbvHRNBr24sQK88Mc4PbGKOj6+9Aa9FCSfvlC0Rz1rRUy8J2sgvSUw67zYrkY9lvlTvjzFIr3DsoA+mbk1PcDNObzBLQ69znpbvcXLND0L5FK+fZESvUhraT667CM9sBPNviLK/7yEVgI/xhwDPb7pUb6QX6y8Ar1TPoOj5Dz55xy8w36KvNVgvb3VEeM8+FtRvjqlmbzwgUc+epLBPMuOFLzMcnO8DWXUvSsWwDwcvD4+PreKvF9Vzb6rmt48pB4NvClszLxT8+i9ZzHdPPI/UL78Dt+8iBAvPn3fuzz+6AC8VwzDvFJPBb57lbo8HoFPvrZg2Lz3mx4+GmKZPOdS6rsUAL+8dYIVvis2mDwdx06+/uvWvL2SDj4ZQW48ERjLvjEcwLwV2Ng+WYzYOxcNTr6ncHW8DQP9Pd1ZKTsa9MC7SfVMvI4AMr4FoiE7KqZNvpz1grxORus9IzLLuiChs7tgRmC8vTA7vvOQ2bqnY0I+gRaOvMSk9b4ACQw7YDjFPqGx3LyUIke/zzohPHkGQz72ES69XdH8vkijXzw0fYS7YYVWvYXyW74cUF480yNLvuYdaL31h3k94k4dPLEIyb5NIGO9Q5+rPtwr5TpsgEm+qKpHvWRg0TwiVg+7U0fIvqeSRb2U3Zo+LgMkvAboFb9dyyy9FV4UPz/ysbz4nse+MaL6vJo7jD4y0/G8CAFHvlLCzbwOrue8M9UIvSw3x76FZNK8PUODPhG1KL12dhW/fGOovNV0Cj8OiVi91eDGvpaNH7wwj3c+HFt4vYHPRb4Nq6C7uxtdvSIXhL2Exca+pgvEu43Xcj76/ZO9t6ZFvquBorqfKmu97eWbvQy4xr6sgRy7z4NwPrDLq70gmUW+PloXO+/Zb70Ys7O9TQ4LOz8zlToodrO+2JyzvRCmRb4macC7zWRrvcSEu731rsa+4BLmu7T0bj7Oacu9wHVFvrBIGrsaC3y9y0/TvXedxr4V8Gq7WO9rPm4z472BX0W+KB6GOjDbgb2IGOu9spjGvntkgLkQHGs+yfv6vaI/Fb8Vco47Oa8FP4FuCb5LoMa+xFZyPLxvbD5wYBG+wapFvtL/njzZ0mm9flQVvirgxr53pZU88nR3PvpIHb7pN0a+TD29PGIsOb3bPyG+Yi3Hvh/VtTxxZII+bjcpvpzhRr7djt88h1v9vLQxLb7oHgM6rH3aPIc3ob4VLy2+z6tHvsnmpjzZymO8Zi0xvp6JL7mjn6Q8SKWZvkYuMb59Q0i++OlmPB0AlbqgLzW+8SgrupyKZjxZJJS+DDM1vg+tSL4Huwc8aN/9O4I2Ob7/TMi+8UQKPBwnmz6XOUG+vutIviyRbTytX1U8Tj5FvsF3yL6l1XE8hNiePhhDTb5lWUm+eL+rPBBOtjwASlG+4uPtut5krzwSCYe+hFNRvjP4Sb7ELoQ8beMRPZpdVb72AMm+qgSKPHusqj7hZ12+yHNKvkKiwDwohDw9lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAMJAwr11H2m+mmZkvfW4Mz7vk8u9lQjYvuMFVr22cuY+TdzcvfyLZ767JjG9g80QPlcf5r03Fvm8LJElvZKMKL4sXue9NlNmvg4NM71Ipus9s5TwvcZC77wBoCm9l6tDvvTG8b3gFGW+VUc5vcG9tD3A8Pq9yBjWvosMMr1Xj7s+uggGvuDIY74WChS98dR2PfyWCr79v9u8TxoPvb9yeb6fIwu+yL5ivgIPI72hEhs9j6wPvrf41L4J9R+9XayiPmMxGL5xl2G+8O0FvZK2VDxqtBy+gpTKvKrdBL0vZZS+ETYdvoSjYL7vmxy9lpj3uzi0Ib5B7tO+ZTodvWKviz5jLiq+jIRfvuXgBr246AG9zKYuvmxp074Hegm9LzSAPqgbN755i16+me3pvKDGV70Wjzu+mvXSviWP8rw6Zmw+T/9DvshRG786vMy8OZkDP0BsUL6Yj9K+PwZxvAu7Wj5l2Fi+Uy0bv9UHK7yMaQA/bEJlvhgXTb81NtW5Bn5KP6uqdb43Ixu/+oZ8PL8Q/z7xCYG+VHzSvnPizzwjcFc+oT+FvheMXb7KWvI8/OyXvcp2h75ng7G8WjPmPGD7tr6Yr4e+MGRevnilqzwVSGW96+iJvqk/076heaI8jRR5PoMijr79/F6+9VPKPK6dML1cXZC+x5LTvmhDwzymtoM+npiUvqnSG79gae08/q4OPz7Umr4q9dO+U10kPXJCjD53EZ++ERIcv1vOOj3uORQ/oU+lvqSU1L4OPWo9PCGaPgqQqb6cPGK+G3OBPe6K3Dw106u+U3TVvmaNgj3gfq0+GBiwvtoWZL6XbpA9eReJPQBgsr4zbda+gCyTPRgUwz7eqba+HSpmvrXHoj23BOU9Fve4vonm/LxIXKc9lnkevgRIub4mg2i+gAWhPZJxJj4/m7u+6q/YvuKtpz3ROfU+r/C/vkvYar4bTLs9wkZaPuNJwr7k7dm+QgfEPf6BCD+vpca+Xo5tvp3e2T0LW4s+1AXJvlHUHb2eBOU9r1lzPNdqyb7giR4+XaDlPX5Vdr771Me+k0SyPufF2z3k4/2+QETEvrdcGz47dsc93oIvvoa2wr5FzrA+/3DAPcYZ3b5HLb++7ZQYPtvArj3AeuO9q6a9vgdzQb0pNKo9wDxUPnoivr6TIRY+eLGyPcLvbb0kory+t1hLvVpQsD0Xh4U+SCS9vtKbEz4A/7o97e9nu2equ77P+qw+5Nm6PfM6iL6/NLi+VxQIP+Xzrz06VQe/S8OyvuCxqz6oTJo9wDdXvjdUr76wkQ4+1LCRPWCj1z09562+TKGqPuUAlj2vECi+nX2qvrb6Bj/oR489Upzdvm0Xpb74qDg/pRp7Pf/hM7+CtJ2+coEGP6aKQT3cXMi+K1OYvkrIqD7KeyE9B0+svQHzlL41MQY/WZcaPWNuur7glI++RgA4P0CG+TwlCSW/tDiIvtTxBT/N5o88NWOvvhzdgr523Kc+G44vPLLLNb1SA3++kdcFP+4CITxf1qq+PE50vvW1pz66s047KD4bveGYbb4Nkgc+PQYdO/v0gj7C4mq+jqunPhgj9jv9EhS90S1kvgltBz76cd47jo2EPm94Yb7BkKc+dg5EPJaaAb2QxFq+g7AFPy2wOTzdG6S+mhJQvlOZNz8qUaE7jw8cv39iQb5ToQU/1TLuu+F5ob7AsTa+TGmnPqhxXrysz8y8dv8vvr+0BT/toma86NSkvilNJb42tTc/eRCovHd4Hr/Tmha+GrtpPyW+Br2FCGu/BOgDvrnlNz8S9FG9hcwiv5hj6r2KBWo/RwaDvSe+cb8R8sS9J044PxC0qb34ISy/7HSnvSWsBj+ePsW9L1bQvr3okb1TNKo+WunVvZKUFb70SoS9wGQHPw7l270zifC+eEJdvQq9qz47I++9ReZZvhDIQb0BkBE+htr3vSWCJz3xIja9y3StPrMt9r1lJJO+LWIavd8NFT6W+QC+Z/UOvYx1Dr3yxEK9k7ABvkW2XD7EWhK9MbYYPg6N+r2lAOq9OyMGveRmNL0mO/+9lMIMPuO+Cb0iSRw+xJn5vfVlRL5QfPq8eyYmvXG6AL7bI3Y92ZAAvedJb77E/v69hgmfPnq1E71Nidq+r0XyvT3EDz+6rDa9/clrvgFF271JuG8+sIlJvRxkCr1GrtG9JwW5vT9OTL1Kwmi+lGHVvepnLD4n7V69CXLXviR8zr1IV9o+5rKAvWLIZb6EBL292n/UPeDjib0PD9a+hcS4va9euz7MA5u98J8cvzDHqb2Q/iA/JhO0vdHO1L7bBJC9HGGfPnQZxb3jFBy/xESDvc+7FD+OEt69EdXTvkfxVr1qo4k+4QTvvXciX76c60C9uJQjvcfx9726G9O+JTFEveE4cz6kagS+6L5dvvS7ML2VG4+9+tkIvgdy0r5hdTa9W+ZVPvBEEb6ydly+tVglvauyx722rRW+ltPRvp5VLb2/ijo+LAizPOXOOD6gfmc7aPOLvuqZ0DwjxXC8Eo/9uiGOpzyLMc48FdNWvvPwx7pWA6A+VNKrPIv1b7zV1Jo7vJyePAlsqTzbvDg+NoWnO9ksi77k+sY8f0NyvGP7qbleB7g8so7EPAyrOD5VJgM5e2eKvrQa4jx8N3K8GQ+tu5uDtzyhrt88Vb44PrtgnrsbPIu+uT39PDBPwD6YTCi8+O4Qv9pjHT1+9jg+MuiwvOKtjb7kLyw9OFFovI8+3rwuuBQ8hgYrPSisOT7Wwdy8VoaVvhnhOT0YqFu8dE0GvbixArzwxzg9sYo6Pr70Br0UJJ++ULRHPWtFTLwnayC9JTDrvNiuRj2W+VO+PMUivcOygD6ZuTU9wM05vMEtDr3Oelu9xcs0PQvkUr59kRK9SGtpPrrsIz2wE82+Isr/vIRWAj/GHAM9vulRvpBfrLwCvVM+g6PkPPnnHLzDfoq81WC9vdUR4zz4W1G+OqWZvPCBRz56ksE8y44UvMxyc7wNZdS9KxbAPBy8Pj4+t4q8X1XNvqua3jykHg28KWzMvFPz6L1nMd088j9QvvwO37yIEC8+fd+7PP7oALxXDMO8Uk8FvnuVujwegU++tmDYvPebHj4aYpk851LquxQAv7x1ghW+KzaYPB3HTr7+69a8vZIOPhlBbjwRGMu+MRzAvBXY2D5ZjNg7Fw1OvqdwdbwNA/093VkpOxr0wLtJ9Uy8jgAyvgWiITsqpk2+nPWCvE5G6z0jMsu6IKGzu2BGYLy9MDu+85DZuqdjQj6BFo68xKT1vgAJDDtgOMU+obHcvJQiR7/POiE8eQZDPvYRLr1d0fy+SKNfPDR9hLthhVa9hfJbvhxQXjzTI0u+5h1ovfWHeT3iTh08sQjJvk0gY71Dn6s+3CvlOmyASb6oqke9ZGDRPCJWD7tTR8i+p5JFvZTdmj4uAyS8BugVv13LLL0VXhQ/P/KxvPiex74xovq8mjuMPjLT8bwIAUe+UsLNvA6u57wz1Qi9LDfHvoVk0rw9Q4M+EbUovXZ2Fb98Y6i81XQKPw6JWL3V4Ma+lo0fvDCPdz4cW3i9gc9Fvg2roLu7G129IheEvYTFxr6mC8S7jddyPvr9k723pkW+q4Giup8qa73t5Zu9DLjGvqyBHLvPg3A+sMurvSCZRb4+Whc779lvvRizs71NDgs7PzOVOih2s77YnLO9EKZFviZpwLvNZGu9xIS7vfWuxr7gEua7tPRuPs5py73AdUW+sEgauxoLfL3LT9O9d53GvhXwartY72s+bjPjvYFfRb4oHoY6MNuBvYgY672ymMa+e2SAuRAcaz7J+/q9oj8VvxVyjjs5rwU/gW4Jvkugxr7EVnI8vG9sPnBgEb7BqkW+0v+ePNnSab1+VBW+KuDGvnellTzydHc++kgdvuk3Rr5MPb08Yiw5vds/Ib5iLce+H9W1PHFkgj5uNym+nOFGvt2O3zyHW/28tDEtvugeAzqsfdo8hzehvhUvLb7Pq0e+yeamPNnKY7xmLTG+nokvuaOfpDxIpZm+Ri4xvn1DSL746WY8HQCVuqAvNb7xKCu6nIpmPFkklL4MMzW+D61Ivge7Bzxo3/07gjY5vv9MyL7xRAo8HCebPpc5Qb6+60i+LJFtPK1fVTxOPkW+wXfIvqXVcTyE2J4+GENNvmVZSb54v6s8EE62PABKUb7i4+263mSvPBIJh76EU1G+M/hJvsQuhDxt4xE9ml1VvvYAyb6qBIo8e6yqPuFnXb7Ic0q+QqLAPCiEPD1vdGG+/wlAu6osyDwX43S+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.7145386338233948, 0.3301219940185547, 0.8899467587471008, 0.6370378732681274, 0.7793131470680237, 0.5874241590499878, 0.8089688420295715, 0.46758198738098145, 0.8422499299049377, 0.5003751516342163, 0.8467786908149719, 0.550575852394104, 0.8031160235404968, 0.4204248785972595, 0.8743100762367249, 0.6273601055145264, 0.7532482743263245, 0.65560382604599, 0.7326817512512207, 0.6786774396896362, 0.28631067276000977, 0.9049944281578064, 0.2818832993507385, 0.09157051146030426, 0.9566946029663086, 0.9197820425033569, 0.7836957573890686, 0.4231114983558655, 0.857382595539093, 0.5579886436462402, 0.8157505989074707, 0.5205936431884766, 0.1657259464263916, 0.9400999546051025, 0.1377422958612442, 0.9469895958900452, 0.8893464207649231, 0.30538681149482727, 0.9083792567253113, 0.23912914097309113, 0.9237581491470337, 0.817291259765625, 0.45908769965171814, 0.14540739357471466, 0.9435144662857056, 0.11238325387239456, 0.9506571292877197, 0.9121183156967163, 0.8074678778648376, 0.5779584646224976, 0.6944699287414551, 0.6532284021377563, 0.630101203918457, 0.2886457145214081, 0.8928754925727844, 0.23312027752399445, 0.9082863926887512, 0.8120977878570557, 0.6071277856826782, 0.6598420739173889, 0.34575900435447693, 0.8603193163871765, 0.7030950784683228, 0.43988409638404846, 0.7909899353981018, 0.5444848537445068, 0.7443684339523315, 0.4748729467391968, 0.7838408946990967, 0.5340209007263184, 0.7517281174659729, 0.5389045476913452, 0.2501910924911499, 0.8950940370559692, 0.24435243010520935, 0.897934079170227, 0.7639405131340027, 0.47292789816856384, 0.8035104274749756, 0.5599677562713623, 0.7422976493835449, 0.41125062108039856, 0.15176455676555634, 0.938137412071228, 0.11105752736330032, 0.9521937370300293, 0.9244604706764221, 0.8477113842964172, 0.36517229676246643, 0.8995999693870544, 0.7670295834541321, 0.5389034748077393, 0.8523119688034058, 0.6268474459648132, 0.7116072773933411, 0.7535645365715027, 0.5744153261184692, 0.8408138155937195, 0.5867356657981873, 0.25469496846199036, 0.9003883004188538, 0.6576654314994812, 0.7835851907730103, 0.4603838324546814, 0.8245253562927246, 0.5588733553886414, 0.22580257058143616, 0.9132584929466248, 0.2633396089076996, 0.904869019985199, 0.7092060446739197, 0.7081540822982788, 0.6708568334579468, 0.7411358952522278, 0.6319013237953186, 0.7692068219184875, 0.6107634902000427, 0.7839415669441223, 0.39703473448753357, 0.8809493780136108, 0.6164395213127136, 0.7797384262084961, 0.6107505559921265, 0.7852819561958313, 0.5989381670951843, 0.2057776153087616, 0.9290640950202942, 0.8207107186317444, 0.5182713866233826, 0.8403854370117188, 0.46897831559181213, 0.860327422618866, 0.5880143046379089, 0.7830023765563965, 0.6252779364585876, 0.24076998233795166, 0.918492317199707, 0.7558153867721558, 0.6562982201576233, 0.743545413017273, 0.329177588224411, 0.8988929390907288, 0.7061043381690979, 0.6914294958114624, 0.728039026260376, 0.6665208339691162, 0.7485456466674805, 0.3601571321487427, 0.8857332468032837, 0.6377137303352356, 0.7662349939346313, 0.6207844614982605, 0.22204458713531494, 0.07517490535974503, 0.9601967334747314, 0.9349498748779297, 0.8495979309082031, 0.5772209763526917, 0.7747247219085693, 0.6230380535125732, 0.2545163631439209, 0.9106013178825378, 0.7390216588973999, 0.6630235910415649, 0.27492284774780273, 0.9086012840270996, 0.7358675599098206, 0.6606889367103577, 0.7389239072799683, 0.656338632106781, 0.7449959516525269, 0.35186490416526794, 0.8877314925193787, 0.6601843237876892, 0.7403911352157593, 0.6570512056350708, 0.7450990080833435, 0.6503136157989502, 0.24748466908931732, 0.920200765132904, 0.7808467149734497, 0.5918764472007751, 0.8006954193115234, 0.5567577481269836, 0.821295440196991, 0.4851776659488678, 0.8356659412384033, 0.5061039924621582, 0.8285458087921143, 0.5198307633399963, 0.8241803646087646, 0.4732814133167267, 0.8542349934577942, 0.44508469104766846, 0.8658439517021179, 0.5906569361686707, 0.7823728322982788, 0.3923027813434601, 0.8830647468566895, 0.6449106931686401], "advantages": [5.34372615814209, 3.4772355556488037, 2.641378402709961, 3.393198251724243, 5.50675630569458, 3.498453378677368, 5.678830146789551, 3.6962521076202393, 2.296417474746704, 3.7592406272888184, 5.846028804779053, 4.012153148651123, 2.433779239654541, 4.092923641204834, 5.93383264541626, 4.364585876464844, 2.822707414627075, 4.435563564300537, 2.9776952266693115, 4.449057102203369, 3.104326009750366, 1.805916666984558, 2.97487211227417, 1.7299500703811646, 0.9767391681671143, 1.421822428703308, 2.063217878341675, 2.5110833644866943, 3.0324864387512207, 2.1789679527282715, 1.3903967142105103, 1.6036921739578247, 0.905038595199585, 0.15614436566829681, 0.16684097051620483, -0.425083190202713, -0.6813451051712036, -1.1224215030670166, -1.48048996925354, -2.1088414192199707, -2.3572349548339844, -3.1857850551605225, -3.7631912231445312, -4.1742448806762695, -4.231551170349121, -5.400660037994385, -5.330511093139648, -6.695231914520264, -7.817572116851807, -8.301892280578613, -8.01997184753418, -9.54063606262207, -9.350038528442383, -10.8381929397583, -11.669363975524902, -12.334385871887207, -13.227333068847656, -13.911215782165527, -13.977618217468262, -13.32863712310791, -15.46882152557373, -17.1589412689209, -17.140392303466797, -16.56845474243164, -15.080556869506836, -18.081636428833008, -20.54050064086914, -19.756214141845703, -18.168752670288086, -21.35610580444336, -24.229143142700195, -23.090972900390625, -26.15956687927246, -29.049827575683594, -28.207744598388672, -31.165536880493164, -30.333826065063477, -28.902372360229492, -26.87515640258789, -30.78998565673828, -34.42581558227539, -32.72133255004883, -30.247812271118164, -26.430543899536133, -31.429292678833008, -26.980817794799805, -32.148597717285156, -36.648887634277344, -40.435855865478516, -37.44397735595703, -41.3135871887207, -44.285980224609375, -42.04058837890625, -45.176239013671875, -47.27995681762695, -45.952308654785156, -48.31342315673828, -46.48448181152344, -49.06890106201172, -50.6634521484375, -51.27536392211914, -51.82233428955078, -50.408817291259766, -52.61117935180664, -53.92329406738281, -53.3603515625, -55.02437210083008, -55.7225456237793, -56.133888244628906, -57.18886947631836, -57.2010498046875, -55.85426712036133, -58.09061050415039, -56.71928405761719, -58.945430755615234, -57.6478271484375, 12.230320930480957, 12.367692947387695, 11.885704040527344, 12.243776321411133, 11.499811172485352, 11.624293327331543, 11.139490127563477, 11.263777732849121, 10.773929595947266, 10.906808853149414, 12.110933303833008, 10.617769241333008, 10.080604553222656, 10.311936378479004, 9.72043514251709, 10.033855438232422, 9.370911598205566, 9.55990982055664, 8.975624084472656, 9.086871147155762, 10.301041603088379, 8.553838729858398, 8.127638816833496, 8.087565422058105, 7.722953796386719, 8.489952087402344, 7.385662078857422, 7.220555305480957, 7.0152482986450195, 6.764087677001953, 6.660372734069824, 6.316971778869629, 7.056035041809082, 5.808440208435059, 5.913172721862793, 5.369409084320068, 5.582090377807617, 7.242582321166992, 10.706400871276855, 7.521082878112793, 5.484389305114746, 4.586575031280518, 4.8134355545043945, 4.327906131744385, 4.34528923034668, 5.550821781158447, 3.8153395652770996, 3.873429536819458, 3.444118022918701, 4.303328514099121, 2.992663860321045, 3.5000691413879395, 2.664701223373413, 3.3611855506896973, 2.3582327365875244, 3.224130392074585, 5.521113395690918, 3.248332977294922, 1.960426926612854, 3.1943047046661377, 1.7602624893188477, 3.1233749389648438, 1.569790244102478, 1.2972478866577148, 1.2165945768356323, 2.6268157958984375, 0.9137896299362183, 2.301605224609375, 0.5635809898376465, 1.8988845348358154, 3.6868131160736084, 1.597379207611084, 3.3193776607513428, 1.3135194778442383, 2.9633233547210693, 1.0475983619689941, -0.6342313885688782, 0.5902262926101685, -1.068387746810913, 0.04324698820710182, 1.3158276081085205, -0.3969970941543579, -1.9769048690795898, -1.0476150512695312]} +{"type": "SampleBatch", "eps_id": [1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1936018840, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAG90Yb7/CUC7qizIPBfjdL7Mg2G+cCdLvhf+oDzEb3o98pNlvnArabuPAqs84q5mvpqmZb5Iv0u+xxmGPAxilz3JuWm+x/eFuxo2kjx0rlq+OM9pvj1aQz7RcV48byQAvwTnZb40qZS7kbBpOy6DUL7N/mW+qIBMviTTBLposLg92xVqvo8Wlrta/ak6xIZPvt4tar56GEM+laM0u0Jo/b77Rma+XmuVuyRXT7x+/0++415mvl5JTL4r84i8KjCvPdV0ar7I68m+ot51vCftvj6BiHK+fNJLvoJa97udrZo9E5x2vvrByb5G28W7h0y7PhOufr6Mn0u+XounOvrkkT1QYIG+qpN1uwglMTvKWGK+I2qBvnWuS76bKOG69HaUPY9zg74Fu8m+036MuSyxuj5sfIe+/qZLvlUv5js+LpM9xYWJvrzhfLsvpAo8y9ZfvuOPib4E4Uu+lQaGO7UtnT3RmYu+MN/JvqZSuDu9zr0+ZqOPvmQFTL5lo1U8e3WjPbGtkb4sH427rMpvPHK8Vb78uJG+E21MvmllKzzWUrU9UMSTvqPXmLttaEg8qKVNvovQk74Vwky+0ZkGPJ/4wz253JW+2WSiu8v0JTxYDke+t+mVvswGTb5AhMw7qc/PPZX2l77oCqq7EoIHPFfHQb4vBJi+b15CPnL/kju7ZvW+mhKWvucNxT6mHae7wENFv68hkr5BYEI+QQeovN1/9b4UMJC+zAqku5KW9ryo+0W+ND2QvjJ6TL76IQu9zbG3PapIkr77HYa78sgDvQqiWr5lU5K+uxFEPoxGFb1fIwS/dV2Qvgc9TLtOjz+9aNRwvqBlkL7yR0q+gNNSvaidLT13a5K+fqXIvpZaT73OBKM+x26Wvt7JSL5WRTW9u+omPMtwmL5eA7q5r280vXmdl766cZi+LQ5IPtexTL3DNBq/lXGWvjMugDpnCn692NOmvgVvlr7dpkk+1F2MvZAtI7/KapS+lQE5O5l5pr2r8bu+ZGOUvg2kQ76xgrW9rwnPvTtYlr7hBMW+uqa5vQlDJj74SJq+whkUvzQAs72qqtk+hTWgvmO5w75klqG91ovZPaAfpL6jgBO/jjydven6vj4NBqq+7ZbCvkj1jb2NCmo9WuqtviNuPL4jnou9216DvrvMr75HlsG+myCWvd7JYjzmq7O+FWA6vnaPlb3HIZq+BYm1vjiFwL4V5KG9bhoEvbliub5X5hG/RDajvcEbcD69OL++qIlDv46bmb2KQQA/CwvHvusydb8xFoW9COZEP+HZ0L4HdJO/cypLvSA1hT+5pdy+vct0v3bT67wlnDs/bnDmvvO+Qr/ngme8rMzcPqE67r6DvhC/OWa0u2ddCT7PBPS+uYu9vrr4OLukHSS+Sc/3vri2EL8YhcW7q60GPieZ/b6KpEK/xqZeu1A32D4msQK/4q0Qv0xupTtMoQM+6JUFvxymQr+RrPk7ZHvYPoJ6Cb+6uRC/WbGDPO65Bz6AXwy/F6u9vrJomTxZvh6+DUUOv0zaEL+PAoA84fMSPrMqEb8+3EK/vYWXPDrV4T5iEBW/M/oQv/nJ3zy//B0+q/YXvwNCvr4iEfk8tMQEvrrdGb8uMBG/8NLjPKaZMD4XxRy/7K6+vj4KAD3s++O9Pa0ev2VnEb9e1+084KRDPrWVIb+/H7++eZIGPWAVvb38fiO/NqERv4UE/jyBl1c+nGgmv9ywQ7+UQRA9dkkDP4xSKr/93hG/m0Q6PdIDbT5oPS2/2y/Avqw6TT1+nz69aCkvvxg4Er+vakk9O+aFPgwWMr+G6cC+M9dePdTsebznAzS/h5gSv0yXXT1hkpY+efI2v2G0wb63rnU9ay+bPFviOL8MiTy+/Tt3PSgqgr6v0zm/ZpXCvnJoYj2Y7Gg90cU7v3c3Pr4EEWc9Pxtfvku5PL9DZsO+zDdVPYCGvD2ErT6/6dQTv0zCXD3cJ80+aqJBvzYsxL56lX09pH0APp6YQ7+LgEG+fO6DPYOTFr5MkES/IJ2pOyrRez33Y9a+hIlEv79WQ767g1k9mMnbvYyDRb/za2k7G7lQPXP8w77hfkW/9NtEvoRdMT1/iZi93HpGv8Cixr6JQys9sRltPl53SL9oGka+Wjs+PSg6Q73wdEm/fkZ9OspTOj3Bkqa+rHNJv5ByR77yrB89MuWYvPdySr/7W3K5iSUePeb7mL5Fc0q/xftGPlCrBT0YPRS/knRJv4cLo7o0d6w87nONvjN2Sb8xYUm+xWZ+PKUxvDz3d0q/ZTTnuvD2gjwKkIe+R3pKv0/XSb5KKy88E84GPaJ8S7/2dwy7GfQ5PCJEg75yf0u/OHFFPubiyztRowu/uIJKv67mHLs6lpm7/WuAvtuFSr8GMEq+v/sevE1iJT2oiEu/s3YPu6/AEbx1vYK+hotLv1DqSb4ibWW8hVwNPfqNTL+uLvS6Dh5avLlthr5rkEy/8YNJvmsTmLwLKdQ8zcFMO4gbFDzDc/g8KbMavQubWDvS+j6+o0PyPN9Vhz6qyd65AxYGPCjJDj2FsZq8BPmIufGnTz4kPQ09uMmZvpqteDtpeMs++UPpPI/8FL/7Y0A8SL1OPgTqiTzqmI++CUaBPD8r1jsM7Tc86TSPPCxYgjy7VE4+fKc9PDoUi759W6M8fvvKPo9JyTulhg+/zk/kPHkSTj6dI6a77DWIvkOkAj2z+co+jz4qvMBwD78iHiM9+ktOPo/ssLxqtoq+F58zPfcN2jvkT928BaRzPKUqND1EAU8+K+DavGWKkr4bukQ9FjfzO2PiBL2F7we7xFVFPYS2P77jDQW9N3aPPnz/NT1jzwg8YzPcvKG6uLyarjY9Y9k+vjPl37yR54U+AmonPV6BFTy9C7W8CWIivWApKD1bd1E+ioq7vOWtrb4/6zg9CwogPFge87wAiFy9GLg5Pdw6Uj6W8Pu8NSO2vpqJSj1VLi48pRwbvRZRlb2OaEs9n1c8vqYVIb1/klQ+Tlc8PUc+QDwrFBC98SLHvWBNPT1aUVQ+VAsYvUE2zb6lSU49fCVRPM/gOL3R3fW9WlVPPdEKOr57tkK9etkhPjVzQD0nj8C+zcM1vQHp3j77oyE9LbM4vmcZEr0eIgQ+VN0SPVP7v75Qhwe9Hg3SPoNL6DzTsTe+OtfLvObO2z1i58o8u7qDPI5BurwKvEW+1onNPIv/Nr7D5Nm8shS9PTtCsDzzfYk8Y8TKvBijVb4wArM8Rj82vvby7LwS8Zs9WNmVPMjOvr5HeeC8hRi4Pr+UMTxFbDW+JpClvHwRbz0qDe87i2GVPBcAnLwkZna+fQD7OzXZNL6YbMO8BmI8PUZChzvGYpo8jeO7vCscgr4ZnJM76yk0viqG5bxT6/88RziBOq5ToDzcZ+C8B1GKvkGGtDpdWjO+YFUGvRqYYTxCTwu7mFS9vp00Bb22g5c+t/8bvHZ+EL8p7dm8fwMUP8F5qrxh5ry+NWV2vMXxjT5v7Oa8UscxvgmNG7xzF6W8H68Bvay4vL6SJyK8ef6JPibhH72jfjG+Qq2Tu+gw17w+FC691p28vl/kpLuprIc++UFMveNaMb7tVow5p9XvvDVyWr34lLy+EqamuUDohj6Fnni9sz0QvwZEojvTWg4/X2OTvb81Qr9qrIM8Ea5ZPzZ2sr3WUBC/jH4HPdwOED9jjcm9uWFCv8mXNT3sml0/Q6fovYiXEL+tgXw9W1QWP7/J/737v72+Vk6WPbIQoT7qewe+/tY0vvQwoz0jEjw90BkLvrvbvr5qEqU9mau5PjS8Er6HLDe+8eyzPSd3xT0NZha+XchzPPfftz3pUSu+CxgWvpydVT6mBbE9N1ncvlTSEb5aCM4+6WSfPUvRMb+OlAm+kD9TPoHxgj1RxsG+91oFvrU3KTz/4WY9T9eHvdAkBb5MglE+/HJhPUhxrr4h9AC+3SsPPNKJRT2Xf/+8UMYAvpmHP76++0I9v3iNPvKaBL6th/E7bJ5ZPRYfgbpNdAS+aYhOPsOJWT3yiY2+2lIAvtrCyj5V5EI9NEoNvyht8L09EU0+3q0VPYh/er5EOei94u+XO6qjAT3aHXM9pgjova6BQ75sgAY9zUS5PqLa7708hXM7ByUkPRg7oz2ss++9tf9KPoOsKj3D30y+95TnvZ4VyT6xSBo9GWr1vsF+170/10k+CAnmPHU2M77na8+99W3ROntcyTy0bQE+JlvPvWESST7fEd48Ojwivi1Qx70gZWE6txzEPPwZEj4pR8e9WURHvgh92zy2uOI+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 0, 0, 0, 1, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 0, 0, 0, 1, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAMyDYb5wJ0u+F/6gPMRvej3yk2W+cCtpu48Cqzzirma+mqZlvki/S77HGYY8DGKXPcm5ab7H94W7GjaSPHSuWr44z2m+PVpDPtFxXjxvJAC/BOdlvjSplLuRsGk7LoNQvs3+Zb6ogEy+JNMEumiwuD3bFWq+jxaWu1r9qTrEhk++3i1qvnoYQz6VozS7Qmj9vvtGZr5ea5W7JFdPvH7/T77jXma+XklMvivziLwqMK891XRqvsjryb6i3nW8J+2+PoGIcr580ku+glr3u52tmj0TnHa++sHJvkbbxbuHTLs+E65+voyfS75ei6c6+uSRPVBggb6qk3W7CCUxO8pYYr4jaoG+da5Lvpso4br0dpQ9j3ODvgW7yb7Tfoy5LLG6Pmx8h77+pku+VS/mOz4ukz3FhYm+vOF8uy+kCjzL1l++44+JvgThS76VBoY7tS2dPdGZi74w38m+plK4O73OvT5mo4++ZAVMvmWjVTx7daM9sa2Rviwfjbusym88crxVvvy4kb4TbUy+aWUrPNZStT1QxJO+o9eYu21oSDyopU2+i9CTvhXCTL7RmQY8n/jDPbnclb7ZZKK7y/QlPFgOR7636ZW+zAZNvkCEzDupz889lfaXvugKqrsSggc8V8dBvi8EmL5vXkI+cv+SO7tm9b6aEpa+5w3FPqYdp7vAQ0W/ryGSvkFgQj5BB6i83X/1vhQwkL7MCqS7kpb2vKj7Rb40PZC+MnpMvvohC73Nsbc9qkiSvvsdhrvyyAO9CqJavmVTkr67EUQ+jEYVvV8jBL91XZC+Bz1Mu06PP71o1HC+oGWQvvJHSr6A01K9qJ0tPXdrkr5+pci+llpPvc4Eoz7Hbpa+3slIvlZFNb276iY8y3CYvl4DurmvbzS9eZ2XvrpxmL4tDkg+17FMvcM0Gr+VcZa+My6AOmcKfr3Y06a+BW+Wvt2mST7UXYy9kC0jv8pqlL6VATk7mXmmvavxu75kY5S+DaRDvrGCtb2vCc+9O1iWvuEExb66prm9CUMmPvhImr7CGRS/NACzvaqq2T6FNaC+Y7nDvmSWob3Wi9k9oB+kvqOAE7+OPJ296fq+Pg0Gqr7tlsK+SPWNvY0Kaj1a6q2+I248viOei73bXoO+u8yvvkeWwb6bIJa93sliPOars74VYDq+do+Vvcchmr4FibW+OIXAvhXkob1uGgS9uWK5vlfmEb9ENqO9wRtwPr04v76oiUO/jpuZvYpBAD8LC8e+6zJ1vzEWhb0I5kQ/4dnQvgd0k79zKku9IDWFP7ml3L69y3S/dtPrvCWcOz9ucOa+875Cv+eCZ7yszNw+oTruvoO+EL85ZrS7Z10JPs8E9L65i72+uvg4u6QdJL5Jz/e+uLYQvxiFxburrQY+J5n9voqkQr/Gpl67UDfYPiaxAr/irRC/TG6lO0yhAz7olQW/HKZCv5Gs+Ttke9g+gnoJv7q5EL9ZsYM87rkHPoBfDL8Xq72+smiZPFm+Hr4NRQ6/TNoQv48CgDzh8xI+syoRvz7cQr+9hZc8OtXhPmIQFb8z+hC/+cnfPL/8HT6r9he/A0K+viIR+Ty0xAS+ut0Zvy4wEb/w0uM8ppkwPhfFHL/srr6+PgoAPez74709rR6/ZWcRv17X7TzgpEM+tZUhv78fv755kgY9YBW9vfx+I782oRG/hQT+PIGXVz6caCa/3LBDv5RBED12SQM/jFIqv/3eEb+bRDo90gNtPmg9Lb/bL8C+rDpNPX6fPr1oKS+/GDgSv69qST075oU+DBYyv4bpwL4z11491Ox5vOcDNL+HmBK/TJddPWGSlj558ja/YbTBvreudT1rL5s8W+I4vwyJPL79O3c9KCqCvq/TOb9mlcK+cmhiPZjsaD3RxTu/dzc+vgQRZz0/G1++S7k8v0Nmw77MN1U9gIa8PYStPr/p1BO/TMJcPdwnzT5qokG/NizEvnqVfT2kfQA+nphDv4uAQb587oM9g5MWvkyQRL8gnak7KtF7Pfdj1r6EiUS/v1ZDvruDWT2Yydu9jINFv/NraTsbuVA9c/zDvuF+Rb/020S+hF0xPX+JmL3ceka/wKLGvolDKz2xGW0+XndIv2gaRr5aOz49KDpDvfB0Sb9+Rn06ylM6PcGSpr6sc0m/kHJHvvKsHz0y5Zi893JKv/tbcrmJJR495vuYvkVzSr/F+0Y+UKsFPRg9FL+SdEm/hwujujR3rDzuc42+M3ZJvzFhSb7FZn48pTG8PPd3Sr9lNOe68PaCPAqQh75Hekq/T9dJvkorLzwTzgY9onxLv/Z3DLsZ9Dk8IkSDvnJ/S784cUU+5uLLO1GjC7+4gkq/ruYcuzqWmbv9a4C+24VKvwYwSr6/+x68TWIlPaiIS7+zdg+7r8ARvHW9gr6Gi0u/UOpJviJtZbyFXA09+o1Mv64u9LoOHlq8uW2GvmuQTL/xg0m+axOYvAsp1Dxckk2/sLS4uijVk7wJkIu+C5tYO9L6Pr6jQ/I831WHPqrJ3rkDFgY8KMkOPYWxmrwE+Yi58adPPiQ9DT24yZm+mq14O2l4yz75Q+k8j/wUv/tjQDxIvU4+BOqJPOqYj74JRoE8PyvWOwztNzzpNI88LFiCPLtUTj58pz08OhSLvn1bozx++8o+j0nJO6WGD7/OT+Q8eRJOPp0jprvsNYi+Q6QCPbP5yj6PPiq8wHAPvyIeIz36S04+j+ywvGq2ir4XnzM99w3aO+RP3bwFpHM8pSo0PUQBTz4r4Nq8ZYqSvhu6RD0WN/M7Y+IEvYXvB7vEVUU9hLY/vuMNBb03do8+fP81PWPPCDxjM9y8obq4vJquNj1j2T6+M+XfvJHnhT4Caic9XoEVPL0LtbwJYiK9YCkoPVt3UT6Kiru85a2tvj/rOD0LCiA8WB7zvACIXL0YuDk93DpSPpbw+7w1I7a+molKPVUuLjylHBu9FlGVvY5oSz2fVzy+phUhvX+SVD5OVzw9Rz5APCsUEL3xIse9YE09PVpRVD5UCxi9QTbNvqVJTj18JVE8z+A4vdHd9b1aVU890Qo6vnu2Qr162SE+NXNAPSePwL7NwzW9AenePvujIT0tszi+ZxkSvR4iBD5U3RI9U/u/vlCHB70eDdI+g0voPNOxN74618u85s7bPWLnyjy7uoM8jkG6vAq8Rb7Wic08i/82vsPk2byyFL09O0KwPPN9iTxjxMq8GKNVvjACszxGPza+9vLsvBLxmz1Y2ZU8yM6+vkd54LyFGLg+v5QxPEVsNb4mkKW8fBFvPSoN7zuLYZU8FwCcvCRmdr59APs7Ndk0vphsw7wGYjw9RkKHO8ZimjyN47u8KxyCvhmckzvrKTS+KoblvFPr/zxHOIE6rlOgPNxn4LwHUYq+QYa0Ol1aM75gVQa9GphhPEJPC7uYVL2+nTQFvbaDlz63/xu8dn4Qvynt2bx/AxQ/wXmqvGHmvL41ZXa8xfGNPm/s5rxSxzG+CY0bvHMXpbwfrwG9rLi8vpInIrx5/ok+JuEfvaN+Mb5CrZO76DDXvD4ULr3Wnby+X+Sku6mshz75QUy941oxvu1WjDmn1e+8NXJavfiUvL4Spqa5QOiGPoWeeL2zPRC/BkSiO9NaDj9fY5O9vzVCv2qsgzwRrlk/NnayvdZQEL+Mfgc93A4QP2ONyb25YUK/yZc1PeyaXT9Dp+i9iJcQv62BfD1bVBY/v8n/vfu/vb5WTpY9shChPup7B77+1jS+9DCjPSMSPD3QGQu+u9u+vmoSpT2Zq7k+NLwSvocsN77x7LM9J3fFPQ1mFr5dyHM899+3PelRK74LGBa+nJ1VPqYFsT03Wdy+VNIRvloIzj7pZJ89S9Exv46UCb6QP1M+gfGCPVHGwb73WgW+tTcpPP/hZj1P14e90CQFvkyCUT78cmE9SHGuviH0AL7dKw880olFPZd//7xQxgC+mYc/vr77Qj2/eI0+8poEvq2H8Ttsnlk9Fh+Buk10BL5piE4+w4lZPfKJjb7aUgC+2sLKPlXkQj00Sg2/KG3wvT0RTT7erRU9iH96vkQ56L3i75c7qqMBPdodcz2mCOi9roFDvmyABj3NRLk+otrvvTyFczsHJSQ9GDujPayz7721/0o+g6wqPcPfTL73lOe9nhXJPrFIGj0ZavW+wX7XvT/XST4ICeY8dTYzvudrz731bdE6e1zJPLRtAT4mW8+9YRJJPt8R3jw6PCK+LVDHvSBlYTq3HMQ8/BkSPilHx71ZREe+CH3bPLa44j6nP8+9SrsOOQgFEj2PgSI+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.7421392202377319, 0.6653075814247131, 0.7267931699752808, 0.6819790601730347, 0.2869735658168793, 0.9037841558456421, 0.7229446768760681, 0.67928147315979, 0.2761118710041046, 0.9095240831375122, 0.7475538849830627, 0.35844656825065613, 0.8867701292037964, 0.35553956031799316, 0.8888583779335022, 0.6565616130828857, 0.7402483224868774, 0.34571242332458496, 0.892118513584137, 0.6718780398368835, 0.722542941570282, 0.32426518201828003, 0.8984048962593079, 0.6982182860374451, 0.6918348670005798, 0.7080796360969543, 0.6820647716522217, 0.7158459424972534, 0.6739458441734314, 0.7218491435050964, 0.3326181173324585, 0.11096330732107162, 0.950690746307373, 0.9050228595733643, 0.7428180575370789, 0.6320573687553406, 0.22915831208229065, 0.9250655174255371, 0.8140720725059509, 0.4927603006362915, 0.8256897330284119, 0.4635116755962372, 0.14022856950759888, 0.9471401572227478, 0.1091386154294014, 0.955327033996582, 0.9178064465522766, 0.789749801158905, 0.4668661952018738, 0.8222082853317261, 0.542384684085846, 0.7849184274673462, 0.3939862847328186, 0.8846608400344849, 0.31964564323425293, 0.905807614326477, 0.7489412426948547, 0.41083580255508423, 0.15646404027938843, 0.06817567348480225, 0.9601849913597107, 0.9351122975349426, 0.8529565334320068, 0.5891408324241638, 0.7689388990402222, 0.41001275181770325, 0.85971999168396, 0.3872164487838745, 0.8695835471153259, 0.6449781060218811, 0.712408721446991, 0.33459702134132385, 0.8875104188919067, 0.7030553221702576, 0.6440524458885193, 0.7298330664634705, 0.6086268424987793, 0.7550589442253113, 0.5702375173568726, 0.22106748819351196, 0.9190675616264343, 0.8118273615837097, 0.4568728506565094, 0.8366149663925171, 0.399812787771225, 0.858665943145752, 0.6573303937911987, 0.6431314945220947, 0.698617160320282, 0.5970705151557922, 0.2663327157497406, 0.8998541831970215, 0.7739797234535217, 0.5198624134063721, 0.7329416275024414, 0.5576813817024231, 0.70754075050354, 0.4125094413757324, 0.832767128944397, 0.6260417699813843, 0.6459827423095703, 0.6510291695594788, 0.3772663176059723, 0.8166376948356628, 0.6205098032951355, 0.6700549125671387, 0.6124781966209412, 0.6753755807876587, 0.39235028624534607, 0.8146733045578003, 0.6239195466041565, 0.6599335670471191, 0.6341677904129028, 0.6481558680534363, 0.6476911902427673, 0.6325516104698181, 0.4194265902042389, 0.8801625967025757, 0.6284494996070862, 0.24167421460151672, 0.9156409502029419, 0.76093590259552, 0.6439739465713501, 0.24204489588737488, 0.9186106324195862, 0.22265751659870148, 0.9252332448959351, 0.8064363598823547, 0.54755038022995, 0.8280000686645508, 0.5013532638549805, 0.8330576419830322, 0.5301783680915833, 0.8211783170700073, 0.44648849964141846, 0.8661826848983765, 0.4002903997898102, 0.8819801807403564, 0.6532848477363586, 0.735798180103302, 0.3108850419521332, 0.9062906503677368, 0.7409862279891968, 0.3635069727897644, 0.8795132040977478, 0.389970988035202, 0.8734208345413208, 0.5946699976921082, 0.7963769435882568, 0.5623757243156433, 0.8139426112174988, 0.47416436672210693, 0.8425362706184387, 0.5129656195640564, 0.8352435231208801, 0.4811350703239441, 0.8488247990608215, 0.4439389109611511, 0.863001823425293, 0.5984823107719421, 0.22694513201713562, 0.9211893081665039, 0.7793283462524414, 0.6054004430770874, 0.7798303365707397, 0.6031936407089233, 0.783858597278595, 0.5954858660697937, 0.20878595113754272, 0.06996893137693405, 0.9638121128082275, 0.05880429223179817, 0.9675294160842896, 0.9519604444503784, 0.9039739370346069, 0.2654876410961151, 0.9250361323356628, 0.810467541217804, 0.5005314350128174, 0.19608362019062042, 0.9199252724647522, 0.7778365015983582, 0.6221711039543152, 0.7432409524917603, 0.3320571482181549, 0.9028380513191223, 0.7243411540985107, 0.35484206676483154, 0.8742329478263855, 0.6252550482749939, 0.22858230769634247, 0.9254031181335449, 0.8040980100631714, 0.4699220657348633, 0.83076012134552, 0.5142132639884949, 0.8305869698524475, 0.4852646291255951, 0.15715593099594116, 0.940685510635376], "advantages": [9.962183952331543, 8.465002059936523, 9.536214828491211, 8.101694107055664, 9.097320556640625, 10.37242317199707, 8.880951881408691, 7.589009761810303, 8.502288818359375, 9.741464614868164, 8.37849235534668, 7.243323802947998, 6.102297306060791, 6.795323848724365, 5.734568119049072, 6.275592803955078, 6.856711387634277, 5.912126064300537, 4.992595672607422, 5.31371545791626, 5.710473537445068, 4.871220588684082, 4.086394786834717, 4.187877655029297, 4.404812812805176, 3.656235694885254, 3.8152549266815186, 3.1065804958343506, 3.210392475128174, 2.5391974449157715, 2.5904624462127686, 3.211838722229004, 4.84514045715332, 3.16235089302063, 2.2554564476013184, 1.6845893859863281, 1.8318346738815308, 2.563819646835327, 1.7302453517913818, 1.2629121541976929, 0.7745552062988281, 0.7657491564750671, 0.8879890441894531, 1.7493469715118408, 0.93590247631073, 1.9783265590667725, 1.1151940822601318, 0.8954168558120728, 0.775594174861908, 0.3050791025161743, 0.46187418699264526, 0.13241349160671234, 0.1359269618988037, -0.1398974061012268, 0.032029103487730026, -0.26948264241218567, -0.053917258977890015, -0.0021848438773304224, -0.4351915717124939, -1.343087077140808, -2.5623273849487305, -1.8149359226226807, -1.5570367574691772, -1.8669129610061646, -2.5702195167541504, -2.30163311958313, -2.4784204959869385, -2.9450674057006836, -3.109083890914917, -3.6538853645324707, -4.486635684967041, -4.30039119720459, -4.466681003570557, -5.137042999267578, -6.019649982452393, -5.9258131980896, -6.823407173156738, -6.780771732330322, -7.698112964630127, -7.696495532989502, -7.8424973487854, -8.747162818908691, -9.762453079223633, -9.752503395080566, -10.830366134643555, -10.793676376342773, -11.949895858764648, -12.873753547668457, -13.030129432678223, -13.986685752868652, -14.163536071777344, -14.018338203430176, -15.399600982666016, -16.513521194458008, -17.135520935058594, -17.702316284179688, -18.332582473754883, -18.960052490234375, -19.045076370239258, -20.366775512695312, -21.085317611694336, -21.73343849182129, -22.483348846435547, -22.410316467285156, -23.800878524780273, -24.59133529663086, -25.32060432434082, -26.15968894958496, -26.890260696411133, -26.700393676757812, -28.359477996826172, -29.394014358520508, -30.01532745361328, -31.133323669433594, -31.70237159729004, -32.913211822509766, 16.74986457824707, 17.005142211914062, 16.46055030822754, 16.680124282836914, 17.812210083007812, 16.40336036682129, 15.946146011352539, 16.11150360107422, 17.23311424255371, 15.871072769165039, 17.063596725463867, 15.668039321899414, 15.206764221191406, 15.446927070617676, 14.943414688110352, 15.255030632019043, 14.6339750289917, 14.896696090698242, 14.321704864501953, 14.723248481750488, 14.059152603149414, 14.552416801452637, 13.8137845993042, 13.92780876159668, 13.52879524230957, 14.27650260925293, 13.32952880859375, 13.286197662353516, 14.286084175109863, 12.876018524169922, 13.755963325500488, 12.480130195617676, 12.534699440002441, 12.164260864257812, 12.34476089477539, 11.869222640991211, 12.453166007995605, 11.520573616027832, 11.986315727233887, 11.262653350830078, 11.902769088745117, 11.045820236206055, 11.89995002746582, 10.888172149658203, 11.019709587097168, 12.389384269714355, 10.562883377075195, 10.405895233154297, 10.220525741577148, 10.210987091064453, 9.893479347229004, 10.020485877990723, 9.576271057128906, 10.529535293579102, 12.597681045532227, 10.005120277404785, 12.116042137145996, 9.55986213684082, 8.058792114257812, 8.062898635864258, 7.5842671394348145, 7.372654438018799, 8.368803024291992, 10.201176643371582, 12.549418449401855, 9.340702056884766, 6.770698070526123, 8.483539581298828, 6.023033142089844, 4.886241912841797, 5.2277727127075195, 6.650463104248047, 8.920669555664062, 5.800436019897461, 3.7454683780670166, 3.0560812950134277, 2.9703521728515625, 3.977858304977417, 5.986214637756348, 3.1176180839538574, 1.4986090660095215, 2.2507383823394775, 0.7635002732276917, 0.5527420043945312]} +{"type": "SampleBatch", "eps_id": [1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1940147822, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAKc/z71Kuw45CAUSPY+BIj46Ps+9SCtIvigFHz0Hwew+9T/XvcqKcrqZ5kQ995U6PqhJ173bHkY+4NNTPbxowr3nXM+9dccZuyENTD2evVo+gnXPvUqjRD7wjF09qOqAvfCXx71/q3q71WRYPag1fD4LwMe9jBNDPhWSbD1cuu+8dvK/vUX+xD5hLGo9SjqbvgowsL2iZUE+QVZRPY135Duoc6i9fM3ju3noUT05c6E+jbyovdLmPz54vWs9RrsgPXsPob1VaMM+avRuPT5gcL6IbZG9A24TP4O5Wz3wKgK/n61zvYaXwj4tEjI9rChMviOLVL3QFBM/AL0hPZrA9L49eiW9YOJEP+Mn9TwdDUK/9fLMvLHUEj8S7XE8WYHpvmP0XbxSukQ/I/e4O4l6Pr8N2+46u8ESP3pUF7xmMua+o7RZPLmtwT73U5W8r64jvnjUqjwI0zs+ZYSvvGcUAj7B4cg85vfBPlS0mrw/eDC+0nkDPdYAEz+H8La8nxvxvlSEMj3pRMI+CwwCvQLNPb6YmVE9izY9Pio7Eb2R9cY9rrxgPX3Cwj7SRQm9oHZTvgnmfz2oOD4+mDAaveJ0mj3ijoc9USwQvPUCFL1rRLY+nTKHPfJMPz6Vsu28YoxVPYjZjj1swsM+2CflvFWMf76vgp49YSJAPo0FB70r9ws9JDKmPRg0xD7tOAS9n5aJvmPktT21FkE+jDwavTbaXjyenb09b7bEPksfGb1A2ZS+SlrNPVJwFD8i8DC99k4Yv1ca5T09TMU+Pq1hvRzjob4A4/Q9C6NDPiSUe72/eCm9U7b8PZ4kxj7W9369daC0viVIBj4KcUU+J++NvciVpL0MOwo+NsejutQ5kb3UOUI+fzQKPql8Rz71dIm9oP7+vd4xDj6pqjo6h46OvfHGFT6aNQ4+/npJPtCQiL07jyu+Lz0SPq28LTuVbY+9jLvTPRVLEj7TeEs+gjGLvSOZV77cXBY+Cj2XOzzRk73Ei3U9DnUWPnLyQb6jXJG9gXSoPgyUEj6CFdo7reKDvWDMcjzxthI+lYBPPklHg72FTZi+Wt0WPiELCzxzdo+9HEfSvNkJFz7OclE+moOQva7grb45Ohs+WtkrPJ5snr2EUo+9N3EbPq/YO75tSqG9ORBKPnGvFz7RNMG+TDWZvfGm7T4C9Q8+tEMSvy8yhr0KfDs/g0EEPh0ywL6iZVC9mO/WPtoi+T094ze+3AEuvadj5D3Yx/E9ahuEPCffJL3g60a+8XDyPc6tNr4QyTS9QAWvPU8i6z2W9I082sgtvYsdYr4D2Os9qCRaPrLfP72XzQa/zJH0PY9amDzEAmu9Nf1+vs9U9T0n4DO+8Wh/vS0WzTziIu495UimPOtbfb0GxJK+uvfuPQ0QMr65a4q9BwhnvF7Y5z1iUbU8lf+KvXeWp750wOg9c2NfPsdnmL0K/yO/86/xPUDHxTwRpbK9wpi+vhut8j2+tS2+fOTBvfnO3b1Ruus9iQC6viVUxr3Zvhw+/9jcPcyQDr8SD8C9u9DTPoIJxj3gnbi+GB2vvRZxvj2PRLc9v+sNvwhOq72SDLc+eo+gPXhit74wqZy9HYMiPcPjkT21Vw2/KAmbvXFTnT68jHY9TUa2vh5zjr00xQe8xGJZPQHRDL8Dyo69+fiFPhZTLD0WQrW+QRKEvYSuVb3AUg89QlQMv0g1hr1PwmA++tXEPJAGPr9+b3q9s0r7PltwFjw4vm+/lTpSvdh7Qz+lbhy87549v42sE72PE+k+zZLHvNtvb799w9y8OYA8PxNoML2kZD2/Aj9IvGne3j4pA2297mULv79wZrtcHA4+VM+MvSFWPb+KJ0K6C1fcPo0aq73uYQu/CuIAPNa7DD6oZ8G9ueuyvvLqLTz1axy+9bfPvTpzC7/Cufc7A7ISPtUH5r27DbO+Kc4qPFWPFr7aWvS9D4QLv5dA9Tsdfxg+tlYFvot+Pb/QbCs8WE7jPo5/FL7VlAu/UXOePAFNHj4uqh++omezvlHHtzwpGAe+Sdcmvum7C7/aKaI86cUrPgoFMr6fwT2/raW9PFPn7j4/M0G+D8pvv1UMBT0vOUQ/ImJUvu/skL/z1kM9QdyIP0iSa77dEXC/E7eNPTm7Sj/pxn6+3GA+v/Mmrj2SkwU/7gCHvhzCDL8+hsM9f52DPkuijL6kX7a+uQ3OPY2mhLsMSJC+ZY4mvkbjzT1E2Ia+bvKRvhAl/TynGcM9eNEFv22hkb6uZCm+dbCtPQRzTr4SU5O+FCq5vmtupT2kvu49HQeXvmjMK77KNKo9/CgZvuu+mL6HW7q+bhSkPYcqLD4ReZy+Kiwuvmn3qj1rMsm98zaevh2Mu75J8aY9i9ZgPjH3ob7vkjC+oO+vPWLDPb05u6O+WPOuPNUJrj2XFp++PYOjvtVVXD60T6E98E8Tvy5Pob689ps8y72JPReehL5FHaG+Kjk1vpNDfj2UlF09vtqhvGWUv7ykTBs86xdkPKGvpbzl41++etwfPEqBnj4sgsm8JNXBvPmmgjygwqM8mGLNvIJWLz5t7YU8QwWJvsFUsbyKncW8Zik0PIcz9zyLSLW8n6pgvr4MPjzUEac+4jrZvPBCyLy3fJQ8n80YPTk83bxVgC4+bJmaPCqYf76mUMG8NQO7Po1oYzxRSQm/lXiFvO8FLj4tuk47P/50vjBBU7wen868W7rVuuHsXj3+hFu8RwAuPoocDrqagHS+09YjvOpdzrzVPq67iB5cPQUYLLyyhWG+vwaLu9SBsD7NQnS8Iz/NvDrOLTufwk89h3h8vCkcLj7ySXA7w+h2vnDBRLw37ro+iYKXuoRbCL96PZq7phIuPvB5QbyJGHa+h1Wrukt1zLwOHYi8+hdHPb7C7LrMfC4+ViaAvBZBf75h7dE6Ra7IvI39qLxLbx09nbWROjIELz5rsaK8RniFvhZwlDul8MO8RGfNvPVJ0jw8w4Q71asvPpYyybzZtIy+NjH1Ox4hvrxCOfa8yyckPFz75TvjHV++BZX0vDQNlj7TXy47QR23vMuQxLy1dRG8dxMRO2FVXr4sBca8CWKNPsiCC7s4ebG8FceYvOQsxbwZ6Ce7sOMxPp64nLzhKKW+0CpvOsUMrbyAktG8yWgTvVJqADoTDV2++XfXvIV6fj5v13q7jvmmvIfArrxsala9XMeKu2JbXL4lVLe8TyJvPlPnC7xY26G8NxGRvAVuh729YBK8/9MzPtHmm7xoi7q+b6qxu8nzvT6MmNe8Obgpv0PyAjsNfDQ+tBsivRzjwb65+7Q7xseVvFMhQb0LOMq9OQCpO9rNWb4LOEm9vv02PpTZbDp1bdC+Y5Q6vUI26T5ZLu27DmxYvgZEFb2XYBg+edg7vHnSgLxWEwm9694evpz/QLyqZle+A8kVvQfWAT6i9oK8wEZxvPhlC72jczW+TWCFvJNeVr4Z6hm9vCHWPd+sp7wpzc6+ZFkRvY4yxT4L2um8C01VvlyY47yZ5qY9bf0FvaiYUbw8Pta8OhlhvrUJB73mEzs+RUL6vJ5IBb+2JPC8GGBEvLvHJ71ncXO+bxvyvJB0U75yQTu91gArPVL4Cb2bR82+6NU3vQ+roz6U0Cq9YiFSvg+mHb3vz1c8DaA7va2szL7SkRy9W0GWPoVfXL0PAlG+XYcEvUqVNLwBGG29UynMvoJuBb0i5Yo+OuGGvc4OUL68at68ngMBvb0zj70Lusu+15PjvEdHgT4QgJ+9/0BPvlo1urzD9Ee9WMqnvQTv47vmNMK8qXmyvkgTqL0Qk06+mFH7vMP8gb2aVrC9L/XKvt7bAr2FpGA+LpPAvT5PF79fxuG843YAP9HI2L3+hcq+w46PvHBjTT5//Oi9Zf1MvhtkXbxW2se9mC/xvQBMyr4UXn28WWJDPlGvAL7LChe/Ndg+vBMH9T6pxAy+QxnKvoQbCLutnDo+JtoUvrNcTL4Dg8066ofjvXzwGL5QGMq+pHQruq9yOj7wBSG+Dv8Wvx3KQzvT+PI+VxotvqIcyr4Oc0w8zjM7PvcvNb5iDRe/Vi2EPPh09T6ERUG+FFLKviu50jxRdUQ+SF1Jvgo0F78ZKPI8Ei78Pux1Vb5Mucq+V20hPSJXVj7RkV2+6T1OvgiTMj2+v5C9xrFhvuZS47vNyCw9mtayviXWYb4Nf0++lisQPUejMr2H/GW+VObLvvWYDD0MGoU+dSRuvluDUL7O5CE9g76xvAxQcr79b8y+yB0gPTf+kD58fXq+pKlRvrJQNz1SVko7lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 0, 0, 1, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 0, 0, 1, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 0, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAADo+z71IK0i+KAUfPQfB7D71P9e9yopyupnmRD33lTo+qEnXvdseRj7g01M9vGjCvedcz711xxm7IQ1MPZ69Wj6Cdc+9SqNEPvCMXT2o6oC98JfHvX+rervVZFg9qDV8PgvAx72ME0M+FZJsPVy677x28r+9Rf7EPmEsaj1KOpu+CjCwvaJlQT5BVlE9jXfkO6hzqL18zeO7eehRPTlzoT6NvKi90uY/Pni9az1GuyA9ew+hvVVowz5q9G49PmBwvohtkb0DbhM/g7lbPfAqAr+frXO9hpfCPi0SMj2sKEy+I4tUvdAUEz8AvSE9msD0vj16Jb1g4kQ/4yf1PB0NQr/18sy8sdQSPxLtcTxZgem+Y/RdvFK6RD8j97g7iXo+vw3b7jq7wRI/elQXvGYy5r6jtFk8ua3BPvdTlbyvriO+eNSqPAjTOz5lhK+8ZxQCPsHhyDzm98E+VLSavD94ML7SeQM91gATP4fwtryfG/G+VIQyPelEwj4LDAK9As09vpiZUT2LNj0+KjsRvZH1xj2uvGA9fcLCPtJFCb2gdlO+CeZ/Pag4Pj6YMBq94nSaPeKOhz1RLBC89QIUvWtEtj6dMoc98kw/PpWy7bxijFU9iNmOPWzCwz7YJ+W8VYx/vq+Cnj1hIkA+jQUHvSv3Cz0kMqY9GDTEPu04BL2flom+Y+S1PbUWQT6MPBq9NtpePJ6dvT1vtsQ+Sx8ZvUDZlL5KWs09UnAUPyLwML32Thi/VxrlPT1MxT4+rWG9HOOhvgDj9D0Lo0M+JJR7vb94Kb1Ttvw9niTGPtb3fr11oLS+JUgGPgpxRT4n7429yJWkvQw7Cj42x6O61DmRvdQ5Qj5/NAo+qXxHPvV0ib2g/v693jEOPqmqOjqHjo698cYVPpo1Dj7+ekk+0JCIvTuPK74vPRI+rbwtO5Vtj72Mu9M9FUsSPtN4Sz6CMYu9I5lXvtxcFj4KPZc7PNGTvcSLdT0OdRY+cvJBvqNckb2BdKg+DJQSPoIV2jut4oO9YMxyPPG2Ej6VgE8+SUeDvYVNmL5a3RY+IQsLPHN2j70cR9K82QkXPs5yUT6ag5C9ruCtvjk6Gz5a2Ss8nmyevYRSj703cRs+r9g7vm1Kob05EEo+ca8XPtE0wb5MNZm98abtPgL1Dz60QxK/LzKGvQp8Oz+DQQQ+HTLAvqJlUL2Y79Y+2iL5PT3jN77cAS69p2PkPdjH8T1qG4Q8J98kveDrRr7xcPI9zq02vhDJNL1ABa89TyLrPZb0jTzayC29ix1ivgPY6z2oJFo+st8/vZfNBr/MkfQ9j1qYPMQCa701/X6+z1T1PSfgM77xaH+9LRbNPOIi7j3lSKY861t9vQbEkr669+49DRAyvrlrir0HCGe8XtjnPWJRtTyV/4q9d5anvnTA6D1zY18+x2eYvQr/I7/zr/E9QMfFPBGlsr3CmL6+G63yPb61Lb585MG9+c7dvVG66z2JALq+JVTGvdm+HD7/2Nw9zJAOvxIPwL270NM+ggnGPeCduL4YHa+9FnG+PY9Etz2/6w2/CE6rvZIMtz56j6A9eGK3vjCpnL0dgyI9w+ORPbVXDb8oCZu9cVOdPryMdj1NRra+HnOOvTTFB7zEYlk9AdEMvwPKjr35+IU+FlMsPRZCtb5BEoS9hK5VvcBSDz1CVAy/SDWGvU/CYD761cQ8kAY+v35ver2zSvs+W3AWPDi+b7+VOlK92HtDP6VuHLzvnj2/jawTvY8T6T7Nkse8229vv33D3Lw5gDw/E2gwvaRkPb8CP0i8ad7ePikDbb3uZQu/v3Bmu1wcDj5Uz4y9IVY9v4onQroLV9w+jRqrve5hC78K4gA81rsMPqhnwb2567K+8uotPPVrHL71t8+9OnMLv8K59zsDshI+1QfmvbsNs74pzio8VY8Wvtpa9L0PhAu/l0D1Ox1/GD62VgW+i349v9BsKzxYTuM+jn8UvtWUC79Rc548AU0ePi6qH76iZ7O+Uce3PCkYB75J1ya+6bsLv9opojzpxSs+CgUyvp/BPb+tpb08U+fuPj8zQb4Pym+/VQwFPS85RD8iYlS+7+yQv/PWQz1B3Ig/SJJrvt0RcL8Tt409ObtKP+nGfr7cYD6/8yauPZKTBT/uAIe+HMIMvz6Gwz1/nYM+S6KMvqRftr65Dc49jaaEuwxIkL5ljia+RuPNPUTYhr5u8pG+ECX9PKcZwz140QW/baGRvq5kKb51sK09BHNOvhJTk74UKrm+a26lPaS+7j0dB5e+aMwrvso0qj38KBm+676Yvodbur5uFKQ9hyosPhF5nL4qLC6+afeqPWsyyb3zNp6+HYy7vknxpj2L1mA+Mfehvu+SML6g7689YsM9vTm7o75Y86481QmuPZcWn749g6O+1VVcPrRPoT3wTxO/Lk+hvrz2mzzLvYk9F56EvkUdob4qOTW+k0N+PZSUXT007aK+j+6MPAhZgT0bnV++oa+lvOXjX7563B88SoGePiyCybwk1cG8+aaCPKDCozyYYs28glYvPm3thTxDBYm+wVSxvIqdxbxmKTQ8hzP3PItItbyfqmC+vgw+PNQRpz7iOtm88ELIvLd8lDyfzRg9OTzdvFWALj5smZo8Kph/vqZQwbw1A7s+jWhjPFFJCb+VeIW87wUuPi26Tjs//nS+MEFTvB6fzrxbutW64exePf6EW7xHAC4+ihwOupqAdL7T1iO86l3OvNU+rruIHlw9BRgsvLKFYb6/Bou71IGwPs1CdLwjP828Os4tO5/CTz2HeHy8KRwuPvJJcDvD6Ha+cMFEvDfuuj6Jgpe6hFsIv3o9mrumEi4+8HlBvIkYdr6HVau6S3XMvA4diLz6F0c9vsLsusx8Lj5WJoC8FkF/vmHt0TpFrsi8jf2ovEtvHT2dtZE6MgQvPmuxorxGeIW+FnCUO6Xww7xEZ8289UnSPDzDhDvVqy8+ljLJvNm0jL42MfU7HiG+vEI59rzLJyQ8XPvlO+MdX74FlfS8NA2WPtNfLjtBHbe8y5DEvLV1Ebx3ExE7YVVeviwFxrwJYo0+yIILuzh5sbwVx5i85CzFvBnoJ7uw4zE+nricvOEopb7QKm86xQytvICS0bzJaBO9UmoAOhMNXb75d9e8hXp+Pm/XeruO+aa8h8CuvGxqVr1cx4q7YltcviVUt7xPIm8+U+cLvFjbobw3EZG8BW6Hvb1gErz/0zM+0eabvGiLur5vqrG7yfO9PoyY17w5uCm/Q/ICOw18ND60GyK9HOPBvrn7tDvGx5W8UyFBvQs4yr05AKk72s1Zvgs4Sb2+/TY+lNlsOnVt0L5jlDq9QjbpPlku7bsObFi+BkQVvZdgGD552Du8edKAvFYTCb3r3h6+nP9AvKpmV74DyRW9B9YBPqL2grzARnG8+GULvaNzNb5NYIW8k15WvhnqGb28IdY936ynvCnNzr5kWRG9jjLFPgva6bwLTVW+XJjjvJnmpj1t/QW9qJhRvDw+1rw6GWG+tQkHveYTOz5FQvq8nkgFv7Yk8LwYYES8u8cnvWdxc75vG/K8kHRTvnJBO73WACs9UvgJvZtHzb7o1Te9D6ujPpTQKr1iIVK+D6Ydve/PVzwNoDu9razMvtKRHL1bQZY+hV9cvQ8CUb5dhwS9SpU0vAEYbb1TKcy+gm4FvSLlij464Ya9zg5Qvrxq3ryeAwG9vTOPvQu6y77Xk+O8R0eBPhCAn73/QE++WjW6vMP0R71Yyqe9BO/ju+Y0wrypebK+SBOovRCTTr6YUfu8w/yBvZpWsL0v9cq+3tsCvYWkYD4uk8C9Pk8Xv1/G4bzjdgA/0cjYvf6Fyr7Djo+8cGNNPn/86L1l/Uy+G2RdvFbax72YL/G9AEzKvhRefbxZYkM+Ua8AvssKF7812D68Ewf1PqnEDL5DGcq+hBsIu62cOj4m2hS+s1xMvgODzTrqh+O9fPAYvlAYyr6kdCu6r3I6PvAFIb4O/xa/HcpDO9P48j5XGi2+ohzKvg5zTDzOMzs+9y81vmINF79WLYQ8+HT1PoRFQb4UUsq+K7nSPFF1RD5IXUm+CjQXvxko8jwSLvw+7HVVvky5yr5XbSE9IldWPtGRXb7pPU6+CJMyPb6/kL3GsWG+5lLju83ILD2a1rK+JdZhvg1/T76WKxA9R6MyvYf8Zb5U5su+9ZgMPQwahT51JG6+W4NQvs7kIT2DvrG8DFByvv1vzL7IHSA9N/6QPnx9er6kqVG+slA3PVJWSjv1rn6+c7EovHKRNz3n34y+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.13719937205314636, 0.9455202221870422, 0.883267879486084, 0.3241580128669739, 0.897824227809906, 0.2754460275173187, 0.9104159474372864, 0.7693707942962646, 0.5555956363677979, 0.2020452618598938, 0.9278361797332764, 0.8304445147514343, 0.5634552836418152, 0.7574928998947144, 0.5905201435089111, 0.2537570595741272, 0.9037778973579407, 0.2368222177028656, 0.9135667085647583, 0.7957863807678223, 0.48212572932243347, 0.8273695707321167, 0.488566517829895, 0.8391197323799133, 0.5705462694168091, 0.7803183794021606, 0.618432879447937, 0.25096505880355835, 0.9131265878677368, 0.7292927503585815, 0.6890793442726135, 0.6930261254310608, 0.7302887439727783, 0.6463812589645386, 0.2283569574356079, 0.9265924096107483, 0.8258713483810425, 0.4770001769065857, 0.8642448782920837, 0.621516764163971, 0.7447679042816162, 0.6967484951019287, 0.677482008934021, 0.7603904604911804, 0.5959271788597107, 0.8122093081474304, 0.49755772948265076, 0.8166764974594116, 0.42987048625946045, 0.8757345080375671, 0.3428109884262085, 0.8995289206504822, 0.7375515103340149, 0.39223575592041016, 0.14446179568767548, 0.9382781386375427, 0.8464528918266296, 0.5299351215362549, 0.8251839280128479, 0.476645827293396, 0.15287822484970093, 0.939720869064331, 0.8739405870437622, 0.33450618386268616, 0.893109917640686, 0.27183032035827637, 0.09085450321435928, 0.9544187784194946, 0.9248446226119995, 0.8414086103439331, 0.6262593865394592, 0.6810675263404846, 0.689680278301239, 0.618844211101532, 0.7388423681259155, 0.5580809712409973, 0.7772582173347473, 0.5001614689826965, 0.8078552484512329, 0.5543692111968994, 0.25420576333999634, 0.8978145122528076, 0.25054264068603516, 0.9040773510932922, 0.7698517441749573, 0.544794499874115, 0.7876192927360535, 0.4834475815296173, 0.8147203326225281, 0.5000714659690857, 0.8085223436355591, 0.4827721118927002, 0.8317233324050903, 0.561135470867157, 0.7762807607650757, 0.4067894518375397, 0.13016876578330994, 0.05212301015853882, 0.9690552353858948, 0.9586198329925537, 0.9320699572563171, 0.8502817749977112, 0.6093135476112366, 0.27321985363960266, 0.8930248022079468, 0.6584535837173462, 0.7683393955230713, 0.5655114650726318, 0.82439124584198, 0.4634256362915039, 0.8659628629684448, 0.638332724571228, 0.29768359661102295, 0.8830918669700623, 0.6542655229568481, 0.7500438690185547, 0.3773428797721863, 0.8888494968414307, 0.6487245559692383, 0.7514327168464661, 0.34327083826065063, 0.898598849773407, 0.6842237114906311, 0.2794755697250366, 0.9078225493431091, 0.7340961694717407, 0.6731790900230408, 0.7401441931724548, 0.33594703674316406, 0.8980317711830139, 0.6753561496734619, 0.2637201249599457, 0.914222240447998, 0.7618116140365601, 0.6292572617530823, 0.7794030904769897, 0.600063681602478, 0.7989006042480469, 0.563621461391449, 0.8197431564331055, 0.4809440076351166, 0.8431661128997803, 0.5049905180931091, 0.8345053195953369, 0.47711068391799927, 0.8547343015670776, 0.5638241767883301, 0.804248571395874, 0.5858399271965027, 0.7930123209953308, 0.39634421467781067, 0.11876644194126129, 0.9499401450157166, 0.9006449580192566, 0.7197067141532898, 0.3354465365409851, 0.8889331817626953, 0.6396569013595581, 0.772725522518158, 0.599037766456604, 0.7978551387786865, 0.4455600380897522, 0.8524134755134583, 0.5339066386222839, 0.17154662311077118, 0.937429666519165, 0.855592668056488, 0.5831952095031738, 0.7778592109680176, 0.6179322004318237, 0.7569401264190674, 0.6454053521156311, 0.7383816838264465, 0.6670306324958801, 0.7223247289657593, 0.3160020411014557, 0.9010210633277893, 0.7179820537567139, 0.3329188823699951, 0.8901341557502747, 0.6690647006034851, 0.7321866750717163, 0.33952000737190247, 0.8916082382202148, 0.6777668595314026, 0.720153272151947, 0.31574514508247375, 0.9015423655509949, 0.28536882996559143, 0.9118154048919678, 0.24589796364307404, 0.9231727123260498, 0.7983414530754089, 0.4533107578754425, 0.8427982926368713, 0.5093666315078735, 0.844536542892456, 0.45235079526901245, 0.8668922781944275, 0.6098235249519348], "advantages": [-0.6832361221313477, -0.7623199224472046, -1.4601459503173828, -1.237363576889038, -2.2146105766296387, -2.1727120876312256, -2.9458534717559814, -3.0773112773895264, -2.4560582637786865, -3.9368882179260254, -4.33083963394165, -4.756577014923096, -4.537454128265381, -3.4821078777313232, -5.467580318450928, -4.592899322509766, -2.3672873973846436, -5.548250675201416, -3.343376398086548, -6.375543117523193, -7.880308628082275, -8.514739036560059, -8.657002449035645, -7.978157043457031, -9.394831657409668, -9.994200706481934, -10.15439510345459, -10.747709274291992, -10.633482933044434, -11.532987594604492, -11.713919639587402, -12.289694786071777, -12.477373123168945, -13.052756309509277, -13.241198539733887, -12.435039520263672, -13.938089370727539, -14.553775787353516, -14.68388557434082, -15.347989082336426, -15.184497833251953, -16.23296546936035, -16.1124267578125, -17.134666442871094, -17.06755828857422, -18.051389694213867, -18.051103591918945, -17.014686584472656, -19.17130470275879, -20.0139102935791, -20.198976516723633, -20.94585609436035, -21.242944717407227, -20.518190383911133, -18.819480895996094, -16.77690315246582, -20.551326751708984, -23.55690574645996, -25.070968627929688, -24.85271453857422, -26.2314453125, -26.191831588745117, -27.26449966430664, -27.321903228759766, -28.340612411499023, -28.585344314575195, -29.323650360107422, -28.092811584472656, -29.969758987426758, -30.671058654785156, -30.21611213684082, -28.77077865600586, -31.73383140563965, -30.56482696533203, -33.048648834228516, -32.22014617919922, -34.1186408996582, -33.69059753417969, -34.894290924072266, -34.924461364746094, -33.991641998291016, -32.6479606628418, -35.719425201416016, -34.5211296081543, -37.23918533325195, -38.16288375854492, -38.46779251098633, -38.91407775878906, -37.56731033325195, -39.473140716552734, -38.001060485839844, -39.9852294921875, -41.25992965698242, -40.68833923339844, -39.39668655395508, -41.33430480957031, -42.91082000732422, -43.489784240722656, -43.4068489074707, -44.65269088745117, -45.27902603149414, -45.3729362487793, -45.4395751953125, -45.58658981323242, -45.498233795166016, -47.11289978027344, -48.49867248535156, -48.88654327392578, -50.209110260009766, -50.75872039794922, -52.01993942260742, -52.7369384765625, -53.20253372192383, -53.08909606933594, -54.99459457397461, -56.402679443359375, 12.071743965148926, 12.108894348144531, 11.676773071289062, 12.166440963745117, 11.300980567932129, 11.372553825378418, 10.886247634887695, 11.29710865020752, 12.803985595703125, 10.92104721069336, 10.127022743225098, 10.529043197631836, 9.733109474182129, 9.845380783081055, 9.296360969543457, 9.679487228393555, 11.218320846557617, 9.326764106750488, 8.513236999511719, 8.960230827331543, 8.111382484436035, 8.611445426940918, 7.714823246002197, 8.28856086730957, 7.329068183898926, 7.288589000701904, 6.898428916931152, 6.796186447143555, 6.4617791175842285, 7.272472381591797, 6.081640720367432, 5.8660783767700195, 5.659859657287598, 5.372581958770752, 5.236370086669922, 6.365539073944092, 9.139998435974121, 6.333460330963135, 4.722204685211182, 4.18825101852417, 4.754950046539307, 3.7013492584228516, 4.178575038909912, 3.3239128589630127, 4.012305736541748, 2.994929790496826, 3.1596670150756836, 2.612265110015869, 3.751471996307373, 6.525382041931152, 3.936882495880127, 2.39058256149292, 2.0504608154296875, 2.2928028106689453, 1.7113906145095825, 2.264075994491577, 1.441030502319336, 2.295166492462158, 1.2389525175094604, 2.3730499744415283, 4.885766983032227, 2.6935598850250244, 1.2506216764450073, 0.9317762851715088, 1.1635041236877441, 2.935565948486328, 1.2256314754486084, 0.4357769191265106, 1.126481294631958, 2.8726069927215576, 1.122775912284851, 0.028035970404744148, 0.9063622951507568, -0.2586348056793213, 0.5556990504264832, -0.6199492812156677, 0.05718197301030159, 1.1061116456985474, 2.177203893661499, 0.6777712106704712, -0.7804844975471497, 0.03493981063365936, -1.3528562784194946, -0.6984062194824219]} +{"type": "SampleBatch", "eps_id": [1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1267110314, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAPWufr5zsSi8cpE3PeffjL7x5H6+APlSvjgHIT2OUgA9j46BvqZFPLw7mCM9Gap+vq+sgb5ibjs+tDgPPfRMB7+4mX++775NvMHZxzwObWa+j9t/vh4DVb6B+6I8tx2aPRcPgr62FFi8zU+vPHonWL6qMYK+0J1VviC6jDzVxbQ9hVSEvqbVzr5aMJs82cvFPoR3iL4RJVa+03vaPBAgzD26m4q+TSrPvk/Q6jygIc0+acCOviX0Vr5WOhY9ruXvPbHmkL5ko8++4tIfPY+i1z7NDZW+MQ9YvkpTQj1wahA+6jaXvhDbh7zt4E09NZYLvmNil77YDDY+MLZCPexQ075WkJW+0EiTvK/mID0h89e9eL+VvsbLND5bQxg9kF3FvqLwk753Opy8iF7xPKeGpr2gIpS+2NQzPhUM5Dztprq+QVaSvqnvoryLUag8fnyBvWWKkr5QITM+qvWdPMbesr7Sv5C+vJmnvDxxSTwyfE+9dPWQvunuXL7w1zg8P8h7Pgsrk76YXKq8+LSEPJIJMb2PYZO+nEQyPjZAezwVWKm+MZmRvjAOrrzh3g48H0cIveTQkb426zE+6fcDPCV5pb5rCZC+RwWwvDmG0DqXMeW8v0GQvmbHMT6sLoc6yuujvqJ6jr74U7C8+wWwu05q3rwOs46+o9cxPgzRwbuynqS+yOuMvpgBr7zZQ0q8nZb7vMgjjb5Gf12+HVRUvEIfhD7RWo++3gqsvHSK/7u1eh693pGPvsdpMj7icgy8CeuqviHJjb7EK70+G9Z5vCMTIb+TAIq+IcIyPooB5LxFxq6+9DaIvi8KpbyI9w29kNdrvcRriL6r6Fu+C68SvYdYZT68noq+9MqcvAtWAL0IaaO96NCKvhfuWr5d3wa9nrZPPl4Bjb78RZW8yoLsvATizL0jMY2+Rwdavsrm/LwZzDs+Sl+PvqFIjryd2t689m7zvdKMj749Llm+IVTyvJQTKT7NuJG+QKSHvMJG17zYCAy+NeSRvmBdWL6Tru285hAXPhoOlL4GLYG86YLVvCreHb5wN5S+W49Xvi3F7rzwTQU+RV+WvqdxdbwGcdm80qsvvouGlr7Z1zg+hIz1vE/q8b5YrZS+/T5ovBx7Ib0c80G+gNKUvn24Vb4y/zC9SoW5PaD1lr599FS8d5MpvaqUXL6zF5e+9ntUvvU4O7367YI9qTeZviKjQLw9/DW90qR4vntWmb69KlO+d+BJvbiFET0Sc5u+YRvNvmT3Rr3W4J8+N42fvnG8Ub7HYi29CNeYOySmob6fYhW89QAtvZwlmr4LvqG+8H9QvtGqRb1MOLS8zdOjvhXIy74ueEe9OpmCPirnp74IFU++3ZIyvd5KV71L+am+VU7VuynhNr05pLe+XAqqvknCQT4eQ1S9sUMqv1YaqL6LpKi7l1+FvS5Bx77UJ6i+zBFMvlNQlb2EEfG9PzKqvp5dyb6YIpq9vbwaPj05rr4D6Um+FfKTvTxHKL4hPrC+xErIvkGtmr3Xc9Y9nz+0vj/CR75CY5a9bt5XvgE/tr4DNce+wQWfvWrlbD3yOrq+749Fvk2nnL3cP4S+tDS8vo4Vxr7FO6e92IkYPOUqwL6zqxS/JdqmvebgjD5JHca+J+XEvvSUm71iVCy9Yw3Kvp0ZFL8eTp298B9nPu75z772x8O+Zg+UvbmwuL1V5NO+944TvwPBl71QIzc+VMvZviQ4Rb+ubZC9vf/kPtuu4b4jCBO/kDd8vW1oCD52kOe+C73Bvu1Ncb3xhTa+Z3DrvpPcOr4B6H+9KlP6vsRO7b7T2sC+qPqTvbLBXb4vKvG+nhgSv3HZnL08+lY9NQL3vtjFv74as5q9KdWGvhbY+r6/ihG/eXylvTBgljs4VQC/fqO+vlpMpb0L+p++QT0CvzD0EL+tGLK9iY09vWojBb/ubL2+7v2zvebrur5YCAe/Y1EQvxXywr0kks+9QesJv6bjQb/ZGMe9rq4jPvfLDb/Hc3O/vYzAva831z5xqhK/KYSSvxZVr726lS4/xYYYvyfOcr8XZpO9OCu6Pu5hHb+znEC/WoGEvbk8ST0bPCG/sU9yvy9+gr04NaQ+vRQmv64kQL9ptmq995UNPIPsKb+D/g2/LwFqvZRlmr6Gwyy/vbk/v6Nagb1cqOC8KJkwv8SODb8zeoK9676tvu9tM79CQz+/g2CQvTARir0zQTe/IvBwv2sjk71zEk8+zRI8v99Nkb8A24q9elvxPrfiQb/7bXC/ABhvvQzwIT63sUa/1xORv4QjYr30G90+T39Mv+YDcL/iwj69a3P6PTBMUb/B5j2/RL40vcAlPb57GFW/Ia9vvwLgQ72Y9L89quNZv+mQPb9iMjy9L8davj6uXb94V2+/9bJNvTJ3gz2sd2K/w4yQv8BwSL1jZa4+3T9ov7P6br99iSy9CNAGPXAHbb+CYpC/P9cpvUnAnz7wzXK/IUipv9dHEL3eYRc/DJnluwOWSLwBWc287wiZuwyf7bvh9lO+4xzOvBtcjz6lozq8ftI8vNs8oLyiB6i8amo+vPpIPD4rmaO8cHujvhoqAryzkDO8oOnXvJsYB716wQW8d6FSvgNR3bxopoA+WChJvBkRJ7wAJrS8egZMvbp/TLzNrT0+OE+8vAbesr47zQ+8m4kcvAeM9bxFFIO9tO4SvKtyPj5EBAC9ImO7vmr6q7sVNg68pf8dveGlqr2pqrG7EhtQvhXTJL2Hq0k+Wm0bvACM97vhsBS9d4HdvRPnHbz7lEA+GY0dvZ/60r6FjcC7JZjUu8xOP7366ga+AM7Eu4u7Tb7qGUq9f1UVPp08JLxstqe7jic+vcTiJb716SW8v1VMvuVsS72J7Ow9FE1nvMCtdbvL8kG9bdtEvoyHaLwb60q+brJRvQRbrj1Pu5S8lrYZuwe5Sr3sm2S+sB2VvKRySb7xAl29KM1aPf1Ytby2Nci+rqJYvV4+qD4tavW83eJHvmy3Pb0QiqE8wrIKvYV9x77hGTy9jUmYPuKdKr2JiUa+MbwjvTCBGbztfzq9wN3Gvq6AJL3gdoo+fFFavV07Fb8uWQ69I9kMP0gJhb3JDUe/yo3CvL/rVD+H4qS9igUVv1Ij6bs/Gwg/dLq8vQgXxr7HlmY7dG5yPlaTzL09bUS+rjkHPI6MYL2/btS9By3GvtSF6juQOXY+Y0nkvf+mRL6oDUQ83aVMvRsn7L2+UMa+d64zPPtifD6bBPy98fxEvv04gjwTCC+94vIBvuWCxr5VcXY8UYWCPqPjCb7hcEW+7fykPGETB72J1g2+tv0DO8GVnzw8c6O++ssNvmAFRr7Wj1Y8E76nvNjBEb4BAse+J9pPPHB4jT6vtxm+AWZGvlYylTwQREq8fK8dviUtkTqJLJM8QjSZvq2pHb6a7Ua+G0xEPNlwdLowpCG+2N4yOuL9QzwlYpS+nKAhvqxTSD6BDco7hSUUv/CeHb53Bcg+1DOxuxKTXr+4nhW+RVBIPpC/uryNFRS/HZ0RvjHjVzrWwgy9JgmWvsyYEb7ZIUk+TcQkvfiqGL8Akw2+e5/2OtqeVb0iG6K+IokNvksURb62jm+9GDYnvS56Eb5KEWI71eZyvczms74YaBG+I1xDvs3Xh70Aqp+9VlAVvqYQxb5ICYu9mLZFPkgyHb7AZkG+siCDvbg39r2AECG+TvfoO1UNiL1zY92+OesgvhKAP75ew5m9uT8lvrS/JL49EMO+hF+gvRWK2j0ojSy+sywTv5gAnL3uVL8+SlM4vjDURL8fsoy9KDQkP1gSSL6LphK/rNhkvQjlpz7AzVO+4gPBvrX7Sb0wt5Q8OYZbvsKJOb7+fki9wU+Uvi48X74ZTMC+1jlgvdkGUrxO7Wa+beMRv6tGYb05MoY+GZlyvs5+v77+zUu9jC9CvQJCer7ugRG/ObBPvXeraj4B84K+2cK+vjDqPL2a7aG9s8OGvq8nEb9VZEO9GnxLPhaSjL5u7EK/9hwzvUKz8z4aXpS+TNIQvwAfDL0q6i0+EymavqJ/vb51avy8uKMIvk/znb6JlRC/myMJverzGD7Zu6O+Wwe9vkTO+bwQYB2+rYOnvrpZEL8tfgm91FIEPtNJrb6yLEK/YtD9vHt/0j4sDrW+UR0Qv2R0urwt6N496NG6vtgpvL5Bn6i8Y4RDvk6Vvr6/9A+/o+fHvBXwwj1qV8S+ytBBv1BPuLx+mMI+FhjMvtLID78RFHS8j52kPXDY0b4irkG/a71ZvAWVvD65l9m+aa4Pvz8YwruPZJI9lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 1], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAPHkfr4A+VK+OAchPY5SAD2PjoG+pkU8vDuYIz0Zqn6+r6yBvmJuOz60OA899EwHv7iZf77vvk28wdnHPA5tZr6P23++HgNVvoH7ojy3HZo9Fw+CvrYUWLzNT688eidYvqoxgr7QnVW+ILqMPNXFtD2FVIS+ptXOvlowmzzZy8U+hHeIvhElVr7Te9o8ECDMPbqbir5NKs++T9DqPKAhzT5pwI6+JfRWvlY6Fj2u5e89seaQvmSjz77i0h89j6LXPs0Nlb4xD1i+SlNCPXBqED7qNpe+ENuHvO3gTT01lgu+Y2KXvtgMNj4wtkI97FDTvlaQlb7QSJO8r+YgPSHz1714v5W+xss0PltDGD2QXcW+ovCTvnc6nLyIXvE8p4amvaAilL7Y1DM+FQzkPO2mur5BVpK+qe+ivItRqDx+fIG9ZYqSvlAhMz6q9Z08xt6yvtK/kL68mae8PHFJPDJ8T7109ZC+6e5cvvDXODw/yHs+CyuTvphcqrz4tIQ8kgkxvY9hk76cRDI+NkB7PBVYqb4xmZG+MA6uvOHeDjwfRwi95NCRvjbrMT7p9wM8JXmlvmsJkL5HBbC8OYbQOpcx5by/QZC+ZscxPqwuhzrK66O+onqOvvhTsLz7BbC7TmrevA6zjr6j1zE+DNHBu7KepL7I64y+mAGvvNlDSrydlvu8yCONvkZ/Xb4dVFS8Qh+EPtFaj77eCqy8dIr/u7V6Hr3ekY++x2kyPuJyDLwJ66q+IcmNvsQrvT4b1nm8IxMhv5MAir4hwjI+igHkvEXGrr70Noi+LwqlvIj3Db2Q12u9xGuIvqvoW74LrxK9h1hlPryeir70ypy8C1YAvQhpo73o0Iq+F+5avl3fBr2etk8+XgGNvvxFlbzKguy8BOLMvSMxjb5HB1q+yub8vBnMOz5KX4++oUiOvJ3a3rz2bvO90oyPvj0uWb4hVPK8lBMpPs24kb5ApIe8wkbXvNgIDL415JG+YF1YvpOu7bzmEBc+Gg6UvgYtgbzpgtW8Kt4dvnA3lL5bj1e+LcXuvPBNBT5FX5a+p3F1vAZx2bzSqy++i4aWvtnXOD6EjPW8T+rxvlitlL79Pmi8HHshvRzzQb6A0pS+fbhVvjL/ML1Khbk9oPWWvn30VLx3kym9qpRcvrMXl772e1S+9Tg7vfrtgj2pN5m+IqNAvD38Nb3SpHi+e1aZvr0qU7534Em9uIURPRJzm75hG82+ZPdGvdbgnz43jZ++cbxRvsdiLb0I15g7JKahvp9iFbz1AC29nCWavgu+ob7wf1C+0apFvUw4tLzN06O+FcjLvi54R706mYI+KuenvggVT77dkjK93kpXvUv5qb5VTtW7KeE2vTmkt75cCqq+ScJBPh5DVL2xQyq/VhqovoukqLuXX4W9LkHHvtQnqL7MEUy+U1CVvYQR8b0/Mqq+nl3Jvpgimr29vBo+PTmuvgPpSb4V8pO9PEcoviE+sL7ESsi+Qa2avddz1j2fP7S+P8JHvkJjlr1u3le+AT+2vgM1x77BBZ+9auVsPfI6ur7vj0W+Taecvdw/hL60NLy+jhXGvsU7p73YiRg85SrAvrOrFL8l2qa95uCMPkkdxr4n5cS+9JSbvWJULL1jDcq+nRkUvx5Onb3wH2c+7vnPvvbHw75mD5S9ubC4vVXk0773jhO/A8GXvVAjNz5Uy9m+JDhFv65tkL29/+Q+267hviMIE7+QN3y9bWgIPnaQ574LvcG+7U1xvfGFNr5ncOu+k9w6vgHof70qU/q+xE7tvtPawL6o+pO9ssFdvi8q8b6eGBK/cdmcvTz6Vj01Ave+2MW/vhqzmr0p1Ya+Ftj6vr+KEb95fKW9MGCWOzhVAL9+o76+WkylvQv6n75BPQK/MPQQv60Ysr2JjT29aiMFv+5svb7u/bO95uu6vlgIB79jURC/FfLCvSSSz71B6wm/puNBv9kYx72uriM+98sNv8dzc7+9jMC9rzfXPnGqEr8phJK/FlWvvbqVLj/Fhhi/J85yvxdmk704K7o+7mEdv7OcQL9agYS9uTxJPRs8Ib+xT3K/L36CvTg1pD69FCa/riRAv2m2ar33lQ08g+wpv4P+Db8vAWq9lGWavobDLL+9uT+/o1qBvVyo4LwomTC/xI4NvzN6gr3rvq2+720zv0JDP7+DYJC9MBGKvTNBN78i8HC/ayOTvXMSTz7NEjy/302RvwDbir16W/E+t+JBv/ttcL8AGG+9DPAhPrexRr/XE5G/hCNivfQb3T5Pf0y/5gNwv+LCPr1rc/o9MExRv8HmPb9EvjS9wCU9vnsYVb8hr2+/AuBDvZj0vz2q41m/6ZA9v2IyPL0vx1q+Pq5dv3hXb7/1sk29MneDPax3Yr/DjJC/wHBIvWNlrj7dP2i/s/puv32JLL0I0AY9cAdtv4JikL8/1ym9ScCfPvDNcr8hSKm/10cQvd5hFz9ik3m/zz6Qvzatv7yNVpM+DJ/tu+H2U77jHM68G1yPPqWjOrx+0jy82zygvKIHqLxqaj68+kg8PiuZo7xwe6O+GioCvLOQM7yg6de8mxgHvXrBBbx3oVK+A1HdvGimgD5YKEm8GREnvAAmtLx6Bky9un9MvM2tPT44T7y8Bt6yvjvND7ybiRy8B4z1vEUUg7207hK8q3I+PkQEAL0iY7u+avqruxU2Dryl/x294aWqvamqsbsSG1C+FdMkvYerST5abRu8AIz3u+GwFL13gd29E+cdvPuUQD4ZjR29n/rSvoWNwLslmNS7zE4/vfrqBr4AzsS7i7tNvuoZSr1/VRU+nTwkvGy2p7uOJz69xOIlvvXpJby/VUy+5WxLvYns7D0UTWe8wK11u8vyQb1t20S+jIdovBvrSr5uslG9BFuuPU+7lLyWthm7B7lKveybZL6wHZW8pHJJvvECXb0ozVo9/Vi1vLY1yL6uoli9Xj6oPi1q9bzd4ke+bLc9vRCKoTzCsgq9hX3HvuEZPL2NSZg+4p0qvYmJRr4xvCO9MIEZvO1/Or3A3ca+roAkveB2ij58UVq9XTsVvy5ZDr0j2Qw/SAmFvckNR7/KjcK8v+tUP4fipL2KBRW/UiPpuz8bCD90ury9CBfGvseWZjt0bnI+VpPMvT1tRL6uOQc8joxgvb9u1L0HLca+1IXqO5A5dj5jSeS9/6ZEvqgNRDzdpUy9Gyfsvb5Qxr53rjM8+2J8PpsE/L3x/ES+/TiCPBMIL73i8gG+5YLGvlVxdjxRhYI+o+MJvuFwRb7t/KQ8YRMHvYnWDb62/QM7wZWfPDxzo776yw2+YAVGvtaPVjwTvqe82MERvgECx74n2k88cHiNPq+3Gb4BZka+VjKVPBBESrx8rx2+JS2ROokskzxCNJm+rakdvprtRr4bTEQ82XB0ujCkIb7Y3jI64v1DPCVilL6coCG+rFNIPoENyjuFJRS/8J4dvncFyD7UM7G7EpNev7ieFb5FUEg+kL+6vI0VFL8dnRG+MeNXOtbCDL0mCZa+zJgRvtkhST5NxCS9+KoYvwCTDb57n/Y62p5VvSIbor4iiQ2+SxRFvraOb70YNie9LnoRvkoRYjvV5nK9zOazvhhoEb4jXEO+zdeHvQCqn71WUBW+phDFvkgJi72YtkU+SDIdvsBmQb6yIIO9uDf2vYAQIb5O9+g7VQ2IvXNj3b456yC+EoA/vl7Dmb25PyW+tL8kvj0Qw76EX6C9FYraPSiNLL6zLBO/mACcve5Uvz5KUzi+MNREvx+yjL0oNCQ/WBJIvoumEr+s2GS9COWnPsDNU77iA8G+tftJvTC3lDw5hlu+wok5vv5+SL3BT5S+LjxfvhlMwL7WOWC92QZSvE7tZr5t4xG/q0ZhvTkyhj4ZmXK+zn6/vv7NS72ML0K9AkJ6vu6BEb85sE+9d6tqPgHzgr7Zwr6+MOo8vZrtob2zw4a+rycRv1VkQ70afEs+FpKMvm7sQr/2HDO9QrPzPhpelL5M0hC/AB8MvSrqLT4TKZq+on+9vnVq/Ly4owi+T/OdvomVEL+bIwm96vMYPtm7o75bB72+RM75vBBgHb6tg6e+ulkQvy1+Cb3UUgQ+00mtvrIsQr9i0P28e3/SPiwOtb5RHRC/ZHS6vC3o3j3o0bq+2Cm8vkGfqLxjhEO+TpW+vr/0D7+j58e8FfDCPWpXxL7K0EG/UE+4vH6Ywj4WGMy+0sgPvxEUdLyPnaQ9cNjRviKuQb9rvVm8BZW8PrmX2b5prg+/PxjCu49kkj0FV9++vGq7vsA/k7sma2S+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.756100594997406, 0.6515371203422546, 0.27392688393592834, 0.9032630920410156, 0.7177965044975281, 0.6907975077629089, 0.7001656293869019, 0.29242339730262756, 0.9091400504112244, 0.2592725455760956, 0.9178235530853271, 0.22246022522449493, 0.9267336130142212, 0.8150317072868347, 0.5043784379959106, 0.8050561547279358, 0.5391262769699097, 0.7899214029312134, 0.5634126663208008, 0.7793945670127869, 0.5781745314598083, 0.7740053534507751, 0.41584810614585876, 0.8650074601173401, 0.6059991121292114, 0.7561507821083069, 0.6087162494659424, 0.7580081820487976, 0.6039019823074341, 0.765137791633606, 0.5913336277008057, 0.7772562503814697, 0.4296134114265442, 0.8560630679130554, 0.5659237504005432, 0.20581138134002686, 0.9263558387756348, 0.8268990516662598, 0.5296928882598877, 0.8078779578208923, 0.5639780163764954, 0.7897661924362183, 0.5955390930175781, 0.7707171440124512, 0.625027060508728, 0.7503579258918762, 0.6530634760856628, 0.7281825542449951, 0.6801934242248535, 0.7035447955131531, 0.2931455671787262, 0.9074767231941223, 0.7535005807876587, 0.6125473976135254, 0.7869656085968018, 0.5590490698814392, 0.8182992935180664, 0.5019524097442627, 0.8211316466331482, 0.4568959176540375, 0.8609375953674316, 0.6036350131034851, 0.7615033388137817, 0.3553500771522522, 0.10620827972888947, 0.9551283717155457, 0.9165684580802917, 0.7830103039741516, 0.5512691736221313, 0.8262907862663269, 0.46774110198020935, 0.8609398603439331, 0.38346102833747864, 0.8886261582374573, 0.6966401934623718, 0.653684675693512, 0.7513241767883301, 0.5858906507492065, 0.7949452996253967, 0.4842629134654999, 0.8081994652748108, 0.4692431092262268, 0.15298905968666077, 0.9438210725784302, 0.877147376537323, 0.3255527913570404, 0.8979164958000183, 0.2635405957698822, 0.9152228832244873, 0.2084365040063858, 0.9296020269393921, 0.8384625315666199, 0.6073265671730042, 0.30342909693717957, 0.8704909682273865, 0.6670771241188049, 0.6860288977622986, 0.6314540505409241, 0.285156786441803, 0.899448573589325, 0.2472555935382843, 0.9114115238189697, 0.7900082468986511, 0.5240451693534851, 0.7550365328788757, 0.5520618557929993, 0.7385621666908264, 0.4295436441898346, 0.8309191465377808, 0.40049228072166443, 0.8433223962783813, 0.6298428773880005, 0.6709883213043213, 0.6448736786842346, 0.3432188332080841, 0.8522629737854004, 0.4863392412662506, 0.8424087762832642, 0.49456992745399475, 0.8485194444656372, 0.5477163195610046, 0.8129344582557678, 0.4291001558303833, 0.8722109794616699, 0.38296541571617126, 0.8875553607940674, 0.6704058051109314, 0.7221771478652954, 0.29367727041244507, 0.9112568497657776, 0.7575416564941406, 0.6152802109718323, 0.7929198145866394, 0.556701123714447, 0.8237192630767822, 0.4931412637233734, 0.8505368232727051, 0.5740993022918701, 0.78038090467453, 0.6180276274681091, 0.7533022165298462, 0.6537297368049622, 0.2725464999675751, 0.09283673763275146, 0.9558241367340088, 0.9155663251876831, 0.7621207237243652, 0.6239885687828064, 0.7753652930259705, 0.6035820841789246, 0.7908192276954651, 0.5774814486503601, 0.8080897331237793, 0.4550636410713196, 0.850283145904541, 0.5335582494735718, 0.8301497101783752, 0.49839186668395996, 0.8329988718032837, 0.509502649307251, 0.17002059519290924, 0.06413768231868744, 0.9635510444641113, 0.9431396722793579, 0.12658293545246124, 0.9492011666297913, 0.896753191947937, 0.2916020452976227, 0.9133553504943848, 0.7691136002540588, 0.5800385475158691, 0.19020146131515503, 0.936056911945343, 0.852791428565979, 0.598386824131012, 0.25553426146507263, 0.9033126831054688, 0.7208718061447144, 0.32151052355766296, 0.900627851486206, 0.7274596095085144, 0.6292542219161987, 0.7590486407279968, 0.5876310467720032, 0.7853286862373352, 0.45330071449279785, 0.8328307867050171, 0.5316824913024902, 0.8127815127372742, 0.5038561820983887, 0.8262620568275452, 0.5247547626495361, 0.8002399206161499, 0.47011104226112366, 0.8391430377960205, 0.5495445728302002, 0.7885800004005432, 0.5474036931991577, 0.7929500937461853, 0.46395400166511536], "advantages": [10.71619701385498, 9.496904373168945, 10.224000930786133, 11.308004379272461, 9.940113067626953, 8.838640213012695, 9.475197792053223, 8.429200172424316, 7.431064605712891, 7.785325527191162, 6.861108303070068, 7.049731731414795, 6.1984124183654785, 6.217492580413818, 6.367193698883057, 6.959926605224609, 5.7406110763549805, 6.309293270111084, 5.112285614013672, 5.6639580726623535, 4.486131191253662, 5.027531623840332, 3.867297649383545, 3.000225067138672, 3.0318500995635986, 3.514237403869629, 2.37080454826355, 2.852041006088257, 1.7219473123550415, 2.2080397605895996, 1.0927358865737915, 1.5896426439285278, 0.491803914308548, -0.33763387799263, -0.32301193475723267, 0.155025452375412, 1.5050715208053589, -0.1369173675775528, -1.1637365818023682, -1.9489446878433228, -1.8420066833496094, -2.592057943344116, -2.5179710388183594, -3.2329299449920654, -3.190302848815918, -3.8703176975250244, -3.8571112155914307, -4.502234935760498, -4.515966892242432, -5.12598180770874, -5.163969039916992, -4.646975040435791, -5.509710788726807, -6.043408393859863, -6.056879997253418, -6.542786598205566, -6.569504261016846, -7.002768516540527, -7.467788219451904, -7.6845598220825195, -7.778643608093262, -8.108591079711914, -8.441168785095215, -8.750797271728516, -8.899496078491211, -8.093663215637207, -8.989055633544922, -9.177718162536621, -9.253134727478027, -9.658561706542969, -9.641874313354492, -10.104632377624512, -10.016752243041992, -10.50851058959961, -10.380828857421875, -10.469677925109863, -10.940350532531738, -10.937837600708008, -11.504593849182129, -11.446361541748047, -11.814696311950684, -12.157753944396973, -12.842756271362305, -13.141839027404785, -13.263139724731445, -13.28312873840332, -13.844392776489258, -13.972935676574707, -14.40645694732666, -14.686785697937012, -14.926937103271484, -15.401124000549316, -15.889636993408203, -16.464384078979492, -17.194381713867188, -17.680950164794922, -18.257383346557617, -18.896621704101562, -19.4682674407959, -19.60881805419922, -20.5964412689209, -20.547060012817383, -21.720386505126953, -22.682628631591797, -23.39211082458496, -24.158687591552734, -24.874181747436523, -25.668039321899414, -25.89198875427246, -27.09056854248047, -27.21060562133789, -28.5135440826416, -29.35718536376953, -30.072494506835938, -30.93113136291504, -31.531635284423828, 6.526088714599609, 6.3266754150390625, 6.083540916442871, 7.01475715637207, 5.698334693908691, 5.386723518371582, 5.2717084884643555, 6.424638748168945, 4.9197516441345215, 6.250074863433838, 4.62456750869751, 4.0756940841674805, 4.311397075653076, 6.0907416343688965, 4.151798725128174, 3.344545602798462, 4.013464450836182, 3.028205394744873, 3.9763708114624023, 2.794191360473633, 4.070122241973877, 2.6745963096618652, 2.4652323722839355, 2.5488569736480713, 2.09835147857666, 2.5073435306549072, 1.8085483312606812, 2.355356216430664, 3.9631409645080566, 1.6893315315246582, 0.9215003848075867, 1.937110185623169, 0.599037766456604, 1.710857629776001, 0.26438188552856445, 1.4385112524032593, -0.09456945210695267, 1.1069295406341553, 3.181992769241333, 0.8805528879165649, -0.7504717111587524, 0.5060690641403198, 2.4393715858459473, 0.234903022646904, 2.11749005317688, 4.167758464813232, 6.837975025177002, 4.385025978088379, 2.4337315559387207, 4.606293678283691, 2.8317625522613525, 1.2482519149780273, 3.125519275665283, 1.7339537143707275, 0.29299408197402954, 2.014551877975464, 3.6011769771575928, 2.557168483734131, 1.4866931438446045, 0.22868794202804565, -1.05729079246521, 0.45399731397628784, 1.6226900815963745, 2.4563777446746826, 1.9006891250610352, 1.168404221534729, 1.9297596216201782, 1.3716800212860107, 1.921997308731079, 1.5300136804580688, 0.7787219882011414, 1.419045329093933, 1.5690706968307495, 1.4618823528289795, 1.4684877395629883, 1.4843612909317017, 1.0726405382156372, 1.2647889852523804, 0.9960322380065918, 1.2049814462661743, 0.9722687005996704, 0.9141455888748169, 0.7524077892303467, 0.5713729858398438]} +{"type": "SampleBatch", "eps_id": [1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1995249552, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAAVX3768aru+wD+TuyZrZL6YFuO+p6QPv+y3Erznqos9gNXovq2PQb9Gvvi7L1G3PpGT8L4QfHO/vo7huVz4JT/ZUPq+KIhBv6NkTTxuBLY+TwcBv3mgD78t8aA8082IPa3mA78UpkG/7OKrPOIxuz4pxge/g8YPv+zJ5zxxEKM9SqYKv9fTQb971fQ85x7DPrCGDr/1/A+/4qIZPYGryD3oZxG/WRJCv76pIT0Q9c0+jUkVv/BEEL/BnUI9mmz6PTYsGL/hYkK/GaJMPcrv2z54Dxy/+p8Qv7TSbz1Uthw+8/MevxfHQr8sXHw9d13tPjbZIr8GEBG/UyuRPdKKQz7uvyW/1Ma+vq39mD2d3aC9Uagnv5yXEb8LxpU9TWtyPsCRKr9UykO/anifPTMzDT8yfC6/0iMSv/gPtj36m5E+b2gxv5MUwb4MtsE9ei+sPLhWM7+z3Tu+cZLCPQIodr4wRzS/vnXCvs25uD1EP6U9ATk2v7SMPr7eB7w9tZo6vuksN7/HyMO+CpG0PX9BDT4eIjm/7yIUv4A3uj2lB+o+kxg8v7gVxb5u8Mw9kQ9HPh0RPr8t90O+0ObUPVRdhb3zCz+/bcAKO/070j3ahKW+LAk/v/f3Rr4p/sQ9b4uRuNoHQL+RkDO6bv3EPYeXhb7ACEC/g5JFPnhNuj2ikwW/3As/vw76xT4r7qQ9K9NIvwoRPb9SohQ/YcyEPat/hr8IGDq/HPXEPn6EMz2sCz2/0h84v+2EQT7CC+489Vfdvh4oN78DcsQ+SjenPMI3N783MTW/fs1APozTxzsWVNW+bjo0v0Li4rvadxK7iPn5vYFDNL+13E6+2zmZu0D/LT5JTDW/Pabfu7R4p7ohb/69PFU1v3DGTr7Vkna7xBMsPuhdNr8pQ8u+loTSuYlI6z5CZji/yLZOvq4AEDzUuio+2m45v8NSy77dokY8raHsPlx3O7/vBE++SgqvPHd+MT5YgDy/chfxu3Fwyzxld+a9/Yk8v8XlPz5+ALk8CmjLvlyUO79OpQO84NJvPFjOx73knju/7ls/PtLaTzwibMW+86k6v5cLCrxHAqM7eiO2vf60Or+KGj8+93JROyWVwr5hwDm/4vULvFhXkLsF27C9lMs5v8t8UL5Z78i7ZdpRPnHWOr9afgm8CUIFuzSot71x4Tq/Sl5QvmjMerueOE8+J+w7v/gOzL4tHWc5S9j8Pov2Pb8z+xe/kG4lPPhdST+wAEG/vyHMvkOX0zz/g/4+QwtDv13XUL6XhBI9xcVZPpUWRL9Jk8y+kvAjPeozBD9LIka/FPJRvqA+Tj0OSnI+Bi9Hv3A5zb61oGE9LnQLP2Y8Sb+eelO+YiCHPcAxij4XS0q/xi1LvJoukj1yVKS6WFtKvzh5Vb51IZI9gkqgPpdsS79Jemy8OPSePUcNMz2Cf0u/qKFXvpe+oD30PLg+hZNMvwBuiLzJe689dkO+PVmpTL+zZDU+8EmzPaiuL74qwUu/x+e9PvNCrD1j7N6+AdtJv2nlMj56bZo90n7wvQX2SL8Zwrw+JJ6VPXpTxb7MEke/SLswPufUhT2pm5C9lTBGv0fCuz6D8II9BhivvutPRL9nlQ8/Ld1pPV4kHb/FcEG/tVFBPxqUNz01VGO/+pI9v0U4Dz+gqt08G+IUv7G1Or84Wbo+MsN8PLWcj76j2Di/massPsjZIDwsZYs8n/s3vzEquj4xbSY8GoyLvgofNr8D/w4/iDuaOzngD7/mQjO/Yw66PvYW1rsJJIm+mGYxv1NnLD6g0EK8bHe6POuJML/FMLo+NVs7vK4bjL5FrS6/eBYPP0ODirwI5hG/qdArv0AZQT9I4+e8Oi1ev//zJ7+2Pw8/Wgo7veuPFb+QFiW/A/C6Pn3mar2jz5y+ADgjv9+dDz+8/oG9U9Edv69YIL+Px0E/8j6bvT2vbb+IeBy/LxwQP4FGwb0aGCm/sJYZv+4MvT6aVNy9xUfMvriyF79c0hA/Q6zsvfA6Ob88zRS/7qW+PqUnBb6UUPC+LOUSvyWYNz54xA6+EZ5hviz6Eb+TrF68oUcTvj1/zzz9CxK/9qo7PtXCEr4RPJ6+xhsRv2d6HLwoFxm+RquEvUooEb8jGk++ymoavpCBNT5hMRK/zjPKvnrJFr70b9Y+BTcUv87PSr6jNQ6+LiyrPZ46Fb/YdaO6cH8Mvgd7f75BPBW/i9RGvn+bEb7pMke7wToWv3Qkxr5vqxG+ZGp3PgA2GL/GbxS/qrgMvoD3+D7/LRu/dNRFvz/DAr53kDs/4iIfv5hGd7/bg+e9VIt7P/AUJL+v/ES/mkS/vUQaKD+CBSi/qbwSvx9fpL35Eqw+zvQqv1QJwb4Lm5a9M1qdPPriLL9KLhK/otGVvS9ckz5szy+/Wve/vrIHir1iUN2827oxv9GrEb/7Iou9GoN5PrGkNL+jW0O/+ieBvYGgAz/tjDi/zzERvxExWL0FJ08+U/8iPQUw+bvWVCo9tQSTPNhfIj3yYT8+NM0rPTfNhb5crzE9aSQQvLFkFj2UXzU93PYwPSpVUb5SBRo9xlayPro3ID1ZhyG8Fo42PeOzij34aB89V4ZSvmcaPD0Ujr8+bJEOPZ97zb5+wFo9nVEuP69h2zzj9VO+WESJPbuhzz7Id7k8LQ9TvKXgmT0z8Qk+eVu3PBJvOT4sZZ89Xo4IvtgG1TzUtHa81u6ZPQQ4Oz5Gj9I85jQ3PvVroT0qi669a9/vPHjivj5L7p09Sz60vlh6Fj2lFxE/6YKPPcOTHr9P6EQ9oNG9Pi1HbD2Db5y+SkdjPa8KMz6TP1M9oiGcOxGacT2jBr0+gKNTPRHcir5J7Ic9TogxPs9rPT3l0xg9NwaPPYRat7xJekA9fj2wPoYbjj2GKzA+Fq1cPeiiiD2BJ5U9VZK7Pj4kYj0rf1W++CikPYiTLj7UD1E9NQTPPaIkqz06zLo+rVdZPcJFM75AFro9VCUPPy0ASz31rua+gf3QPYUMuj5iFyY97BcSvsnf3z0XuSs+ZGcaPRJtJj46vuY9QXq5Psy3Jz1CuvG9z5T1PauIDj+BDB49sYfLvn8xBj5t5bg+1Pf6PJhPvr3Wlg0+SY0pPkO+6zw2SVY+8foQPjh2uD61Awc9gPiXvdRbGD6pDg4/h+8APQJutr4puSM+iP23Pmp+xzxcpFy9ORUrPr7TJz4Kq748w098PoBwLj6MPQG9uAnnPFLcCz8Syy0+3xMnPj5GID0UfIY+giIxPs8aBb28yjU9BUMRPyJ4MD4F3SU+lUZkPaT+kz5byTM+zWe2PnL0ez1j0Gw8MBU7PjrqDD+RI309XDOEviBbRj4/gbU+ofxnPVZfWj28nU0+C3wMP7JabD12R2K+3NpYPoY2Pj+AQFo9N2X9vmsSaD6FFAw/bLUxPQljPr5DR3M+pt09P016Ij2P4e2+2zuBPurGCz9i1fg83YYjvizThj4WnT0/U6vePPOh4r7RaI4+EpELP5Ulljya6RC++/2TPlkZsz7u6308YaggPveSlz5scQs/gaqYPIIABr7dJp0+HlM9P8k5gzyX0NW+jLmkPvlQCz9MOPs7SJf1vSZMqj47PT0/faGsOwoC0r713bE+2kQLP9dbQLupOe29EnC3Phw7PT9pF6y7I6LRvssBvz6BTAs/CDZcvOaE8r03lMQ+d0w9P8+BgbyQo9S+oibMPt5nCz8wjcW8ebcCvia60T4QF7M+WHfavOcZIT4XT9U+Y5cLP6SwwLz+GBO+geTaPnOgPT/COdi8Yifjvkh64j4Nxgs/GnUQvRE9I76REeg+dtk9PzqEHb2eD+2+oKnvPivvbz8/ckO97X5Ev41C+T7DKD4/lymBvTME+77kbgA/GXMMP2g+lb2kV1++/j0DPxeRtT5vLZ69U8dPPc4OBT/A/gw/hRmcvQvUh76z4Ac/1DM/P0j3pr3UvBS/qLMLPwKRDT+Tw76991GhvnqIDj/397c+aavLvX3BWb1vXxA/07kpPt7Yzb3ahFM+rzgRPxdsuT7qYsW9nrDtvV0TEz9MlCw+4yPKvaIzFD5E8BM/eCzNvE02xL2tNs8+cM8TPwJoLz6RorO9/dyqPfWvFD/cDLi8vzewvVXEsT6DkhQ/4u8xPhT/ob0c5eo8RXYVP0hnvT5q0qC9DGWTviRbFz/AOzQ+D52svX+1q7zXQRg/veSRvNh4rb0v13k+gCoYPyeyNj56eqO9NeWXvVkUGT+VmH28LoSmvazqRD4QABk/fl9WvsGjnr0qouo+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 0, 1, 1, 0, 0, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 0, 1, 1, 0, 0, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAJgW476npA+/7LcSvOeqiz2A1ei+rY9Bv0a++LsvUbc+kZPwvhB8c7++juG5XPglP9lQ+r4oiEG/o2RNPG4Etj5PBwG/eaAPvy3xoDzTzYg9reYDvxSmQb/s4qs84jG7PinGB7+Dxg+/7MnnPHEQoz1Kpgq/19NBv3vV9DznHsM+sIYOv/X8D7/iohk9gavIPehnEb9ZEkK/vqkhPRD1zT6NSRW/8EQQv8GdQj2abPo9NiwYv+FiQr8Zokw9yu/bPngPHL/6nxC/tNJvPVS2HD7z8x6/F8dCvyxcfD13Xe0+NtkivwYQEb9TK5E90opDPu6/Jb/Uxr6+rf2YPZ3doL1RqCe/nJcRvwvGlT1Na3I+wJEqv1TKQ79qeJ89MzMNPzJ8Lr/SIxK/+A+2PfqbkT5vaDG/kxTBvgy2wT16L6w8uFYzv7PdO75xksI9Aih2vjBHNL++dcK+zbm4PUQ/pT0BOTa/tIw+vt4HvD21mjq+6Sw3v8fIw74KkbQ9f0ENPh4iOb/vIhS/gDe6PaUH6j6TGDy/uBXFvm7wzD2RD0c+HRE+vy33Q77Q5tQ9VF2FvfMLP79twAo7/TvSPdqEpb4sCT+/9/dGvin+xD1vi5G42gdAv5GQM7pu/cQ9h5eFvsAIQL+DkkU+eE26PaKTBb/cCz+/DvrFPivupD0r00i/ChE9v1KiFD9hzIQ9q3+GvwgYOr8c9cQ+foQzPawLPb/SHzi/7YRBPsIL7jz1V92+Hig3vwNyxD5KN6c8wjc3vzcxNb9+zUA+jNPHOxZU1b5uOjS/QuLiu9p3EruI+fm9gUM0v7XcTr7bOZm7QP8tPklMNb89pt+7tHinuiFv/r08VTW/cMZOvtWSdrvEEyw+6F02vylDy76WhNK5iUjrPkJmOL/Itk6+rgAQPNS6Kj7abjm/w1LLvt2iRjytoew+XHc7v+8ET75KCq88d34xPliAPL9yF/G7cXDLPGV35r39iTy/xeU/Pn4AuTwKaMu+XJQ7v06lA7zg0m88WM7HveSeO7/uWz8+0tpPPCJsxb7zqTq/lwsKvEcCozt6I7a9/rQ6v4oaPz73clE7JZXCvmHAOb/i9Qu8WFeQuwXbsL2Uyzm/y3xQvlnvyLtl2lE+cdY6v1p+CbwJQgW7NKi3vXHhOr9KXlC+aMx6u544Tz4n7Du/+A7Mvi0dZzlL2Pw+i/Y9vzP7F7+QbiU8+F1JP7AAQb+/Icy+Q5fTPP+D/j5DC0O/XddQvpeEEj3FxVk+lRZEv0mTzL6S8CM96jMEP0siRr8U8lG+oD5OPQ5Kcj4GL0e/cDnNvrWgYT0udAs/ZjxJv556U75iIIc9wDGKPhdLSr/GLUu8mi6SPXJUpLpYW0q/OHlVvnUhkj2CSqA+l2xLv0l6bLw49J49Rw0zPYJ/S7+ooVe+l76gPfQ8uD6Fk0y/AG6IvMl7rz12Q749WalMv7NkNT7wSbM9qK4vvirBS7/H570+80KsPWPs3r4B20m/aeUyPnptmj3SfvC9BfZIvxnCvD4knpU9elPFvswSR79IuzA+59SFPambkL2VMEa/R8K7PoPwgj0GGK++609Ev2eVDz8t3Wk9XiQdv8VwQb+1UUE/GpQ3PTVUY7/6kj2/RTgPP6Cq3Twb4hS/sbU6vzhZuj4yw3w8tZyPvqPYOL+Zqyw+yNkgPCxlizyf+ze/MSq6PjFtJjwajIu+Ch82vwP/Dj+IO5o7OeAPv+ZCM79jDro+9hbWuwkkib6YZjG/U2csPqDQQrxsd7o864kwv8Uwuj41Wzu8rhuMvkWtLr94Fg8/Q4OKvAjmEb+p0Cu/QBlBP0jj57w6LV6///Mnv7Y/Dz9aCju9648Vv5AWJb8D8Lo+feZqvaPPnL4AOCO/350PP7z+gb1T0R2/r1ggv4/HQT/yPpu9Pa9tv4h4HL8vHBA/gUbBvRoYKb+wlhm/7gy9PppU3L3FR8y+uLIXv1zSED9DrOy98Do5vzzNFL/upb4+pScFvpRQ8L4s5RK/JZg3PnjEDr4RnmG+LPoRv5OsXryhRxO+PX/PPP0LEr/2qjs+1cISvhE8nr7GGxG/Z3ocvCgXGb5Gq4S9SigRvyMaT77Kahq+kIE1PmExEr/OM8q+eskWvvRv1j4FNxS/zs9KvqM1Dr4uLKs9njoVv9h1o7pwfwy+B3t/vkE8Fb+L1Ea+f5sRvukyR7vBOha/dCTGvm+rEb5kanc+ADYYv8ZvFL+quAy+gPf4Pv8tG7901EW/P8MCvneQOz/iIh+/mEZ3v9uD571Ui3s/8BQkv6/8RL+aRL+9RBooP4IFKL+pvBK/H1+kvfkSrD7O9Cq/VAnBvgublr0zWp08+uIsv0ouEr+i0ZW9L1yTPmzPL79a97++sgeKvWJQ3bzbujG/0asRv/sii70ag3k+saQ0v6NbQ7/6J4G9gaADP+2MOL/PMRG/ETFYvQUnTz5SdDu/EB++vpaeR72/Otq92F8iPfJhPz40zSs9N82FvlyvMT1pJBC8sWQWPZRfNT3c9jA9KlVRvlIFGj3GVrI+ujcgPVmHIbwWjjY947OKPfhoHz1XhlK+Zxo8PRSOvz5skQ49n3vNvn7AWj2dUS4/r2HbPOP1U75YRIk9u6HPPsh3uTwtD1O8peCZPTPxCT55W7c8Em85Pixlnz1ejgi+2AbVPNS0drzW7pk9BDg7PkaP0jzmNDc+9WuhPSqLrr1r3+88eOK+PkvunT1LPrS+WHoWPaUXET/pgo89w5Mev0/oRD2g0b0+LUdsPYNvnL5KR2M9rwozPpM/Uz2iIZw7EZpxPaMGvT6Ao1M9EdyKvknshz1OiDE+z2s9PeXTGD03Bo89hFq3vEl6QD1+PbA+hhuOPYYrMD4WrVw96KKIPYEnlT1Vkrs+PiRiPSt/Vb74KKQ9iJMuPtQPUT01BM89oiSrPTrMuj6tV1k9wkUzvkAWuj1UJQ8/LQBLPfWu5r6B/dA9hQy6PmIXJj3sFxK+yd/fPRe5Kz5kZxo9Em0mPjq+5j1Berk+zLcnPUK68b3PlPU9q4gOP4EMHj2xh8u+fzEGPm3luD7U9/o8mE++vdaWDT5JjSk+Q77rPDZJVj7x+hA+OHa4PrUDBz2A+Je91FsYPqkODj+H7wA9Am62vim5Iz6I/bc+an7HPFykXL05FSs+vtMnPgqrvjzDT3w+gHAuPow9Ab24Cec8UtwLPxLLLT7fEyc+PkYgPRR8hj6CIjE+zxoFvbzKNT0FQxE/IngwPgXdJT6VRmQ9pP6TPlvJMz7NZ7Y+cvR7PWPQbDwwFTs+OuoMP5EjfT1cM4S+IFtGPj+BtT6h/Gc9Vl9aPbydTT4LfAw/slpsPXZHYr7c2lg+hjY+P4BAWj03Zf2+axJoPoUUDD9stTE9CWM+vkNHcz6m3T0/TXoiPY/h7b7bO4E+6sYLP2LV+DzdhiO+LNOGPhadPT9Tq94886HivtFojj4SkQs/lSWWPJrpEL77/ZM+WRmzPu7rfTxhqCA+95KXPmxxCz+Bqpg8ggAGvt0mnT4eUz0/yTmDPJfQ1b6MuaQ++VALP0w4+ztIl/W9JkyqPjs9PT99oaw7CgLSvvXdsT7aRAs/11tAu6k57b0ScLc+HDs9P2kXrLsjotG+ywG/PoFMCz8INly85oTyvTeUxD53TD0/z4GBvJCj1L6iJsw+3mcLPzCNxbx5twK+JrrRPhAXsz5Yd9q85xkhPhdP1T5jlws/pLDAvP4YE76B5No+c6A9P8I52LxiJ+O+SHriPg3GCz8adRC9ET0jvpER6D522T0/OoQdvZ4P7b6gqe8+K+9vPz9yQ73tfkS/jUL5PsMoPj+XKYG9MwT7vuRuAD8Zcww/aD6VvaRXX77+PQM/F5G1Pm8tnr1Tx089zg4FP8D+DD+FGZy9C9SHvrPgBz/UMz8/SPemvdS8FL+osws/ApENP5PDvr33UaG+eogOP/f3tz5pq8u9fcFZvW9fED/TuSk+3tjNvdqEUz6vOBE/F2y5Pupixb2esO29XRMTP0yULD7jI8q9ojMUPkTwEz94LM28TTbEva02zz5wzxM/AmgvPpGis7393Ko99a8UP9wMuLy/N7C9VcSxPoOSFD/i7zE+FP+hvRzl6jxFdhU/SGe9PmrSoL0MZZO+JFsXP8A7ND4Pnay9f7WrvNdBGD+95JG82HitvS/XeT6AKhg/J7I2Pnp6o7015Ze9WRQZP5WYfbwuhKa9rOpEPhAAGT9+X1a+waOevSqi6j6q7Rc/OaFYvHbei71+nxE+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.8257688283920288, 0.5626216530799866, 0.23873643577098846, 0.9067564606666565, 0.7856099605560303, 0.49766361713409424, 0.8052246570587158, 0.4569495916366577, 0.826154887676239, 0.40981125831604004, 0.8473221659660339, 0.357986181974411, 0.8677178621292114, 0.30410391092300415, 0.8865046501159668, 0.7486224174499512, 0.5095136761665344, 0.2117650955915451, 0.9142415523529053, 0.8285714387893677, 0.6418746709823608, 0.6049680709838867, 0.6971065998077393, 0.5416725277900696, 0.2561173737049103, 0.892861008644104, 0.7902103662490845, 0.602611780166626, 0.6124669909477234, 0.6567437052726746, 0.442424476146698, 0.26543065905570984, 0.15198110044002533, 0.9127352237701416, 0.8533020615577698, 0.26722103357315063, 0.859377920627594, 0.7457646131515503, 0.537477433681488, 0.7126670479774475, 0.5382471084594727, 0.2885003089904785, 0.8654332160949707, 0.27987420558929443, 0.8691206574440002, 0.7341107130050659, 0.5083977580070496, 0.7070846557617188, 0.5116556286811829, 0.7083638906478882, 0.5089176893234253, 0.7143175601959229, 0.49994274973869324, 0.7362181544303894, 0.5003710389137268, 0.26436078548431396, 0.1265174001455307, 0.9294705390930176, 0.8794173002243042, 0.23314137756824493, 0.8861583471298218, 0.2124416083097458, 0.8937957286834717, 0.8099349737167358, 0.3407258987426758, 0.8290203809738159, 0.302369087934494, 0.8464890718460083, 0.7347143292427063, 0.5731024742126465, 0.5978360772132874, 0.6071500182151794, 0.5675453543663025, 0.6360226273536682, 0.46020692586898804, 0.29985934495925903, 0.8216624855995178, 0.7090388536453247, 0.5391409397125244, 0.6603631973266602, 0.46193012595176697, 0.7209643125534058, 0.5536842346191406, 0.6486561298370361, 0.43428632616996765, 0.2470548003911972, 0.8723310828208923, 0.7935767769813538, 0.3426437973976135, 0.16905328631401062, 0.9184088706970215, 0.877114474773407, 0.2011764645576477, 0.9117933511734009, 0.8626859188079834, 0.7626271843910217, 0.42835715413093567, 0.8348175883293152, 0.691239058971405, 0.4503331482410431, 0.7731600999832153, 0.44171008467674255, 0.8382903337478638, 0.6743131279945374, 0.4115300178527832, 0.19679485261440277, 0.09697017073631287, 0.9424310326576233, 0.8917253017425537, 0.7317364811897278, 0.5915199518203735, 0.6917995810508728, 0.6348872780799866, 0.3492182791233063, 0.8458874225616455, 0.6246107220649719, 0.633935809135437, 0.7183441519737244, 0.3329625129699707, 0.8886919021606445, 0.28432315587997437, 0.09692191332578659, 0.9518459439277649, 0.9200588464736938, 0.8263546824455261, 0.42920032143592834, 0.8587599992752075, 0.6513775587081909, 0.3166543245315552, 0.8752435445785522, 0.6515479683876038, 0.7288286685943604, 0.610114574432373, 0.24187780916690826, 0.9096800088882446, 0.7953376770019531, 0.49959492683410645, 0.8199946284294128, 0.5503228902816772, 0.7696138024330139, 0.42422568798065186, 0.849068284034729, 0.6130354404449463, 0.7291445136070251, 0.37080830335617065, 0.8672952651977539, 0.6579036116600037, 0.6932432055473328, 0.3313560485839844, 0.11992057412862778, 0.9412961006164551, 0.10656625032424927, 0.9456089735031128, 0.9074525833129883, 0.7886638641357422, 0.4975608289241791, 0.8132179975509644, 0.5485888123512268, 0.7597054839134216, 0.5700709819793701, 0.7503189444541931, 0.5799199342727661, 0.7484939694404602, 0.4210088849067688, 0.8413461446762085, 0.5897952914237976, 0.7458730936050415, 0.579553484916687, 0.7573979496955872, 0.5591249465942383, 0.7746632099151611, 0.5280197858810425, 0.7963343262672424, 0.5141072273254395, 0.7851057052612305, 0.45569196343421936, 0.8341508507728577, 0.40394654870033264, 0.14309191703796387, 0.9374598264694214, 0.8845937848091125, 0.7365036010742188, 0.5504610538482666, 0.2136368602514267, 0.9179798364639282, 0.8351991176605225, 0.6383849382400513, 0.6346977949142456, 0.7075672149658203, 0.4456018805503845, 0.7734357118606567, 0.5113584995269775, 0.7308218479156494, 0.4294901192188263, 0.826607346534729, 0.6374543905258179, 0.6162071228027344, 0.6886051893234253, 0.44197309017181396, 0.763580858707428], "advantages": [23.95086097717285, 22.991289138793945, 21.537565231323242, 19.930086135864258, 21.27425765991211, 22.73177146911621, 21.32071304321289, 22.721357345581055, 21.31170654296875, 22.64296531677246, 21.235586166381836, 22.485576629638672, 21.085309982299805, 22.241727828979492, 20.85965919494629, 21.908767700195312, 22.75075912475586, 21.792234420776367, 20.482646942138672, 21.293676376342773, 21.96241569519043, 22.550926208496094, 21.83930206298828, 22.339046478271484, 21.627050399780273, 20.795997619628906, 21.122188568115234, 21.445466995239258, 22.096115112304688, 21.277395248413086, 21.846481323242188, 23.073759078979492, 24.91632652282715, 27.05020523071289, 25.446693420410156, 23.597171783447266, 25.451757431030273, 23.49586296081543, 21.63089942932129, 20.332618713378906, 20.778114318847656, 19.474611282348633, 18.725296020507812, 18.40410614013672, 17.706119537353516, 17.378328323364258, 17.69611167907715, 18.8607120513916, 17.083532333374023, 18.235816955566406, 16.39839744567871, 17.552953720092773, 15.65087890625, 14.328669548034668, 14.592450141906738, 13.27332592010498, 12.489974975585938, 11.317055702209473, 11.115880012512207, 10.679228782653809, 10.003006935119629, 9.593965530395508, 8.885342597961426, 8.538631439208984, 8.45246696472168, 7.66885232925415, 7.508100509643555, 6.728886604309082, 6.521304607391357, 6.7088446617126465, 7.3170695304870605, 5.9806108474731445, 6.441709041595459, 5.097110748291016, 5.425076961517334, 6.047134876251221, 6.868538856506348, 5.49683141708374, 3.9302456378936768, 2.3060507774353027, 2.482945680618286, 2.9832944869995117, 1.3077181577682495, -0.47211751341819763, -0.2787628769874573, 0.25314730405807495, 1.0077744722366333, -0.6019788980484009, -2.5261166095733643, -1.8333390951156616, -0.9061185717582703, -2.638707160949707, -4.73069953918457, -3.705681800842285, -5.840037822723389, -8.381463050842285, -11.129940032958984, -10.104172706604004, -12.916311264038086, -15.562175750732422, -17.688013076782227, -17.878828048706055, -16.883323669433594, -19.563915252685547, -21.731664657592773, -23.403392791748047, -25.000246047973633, -27.16853904724121, -27.425058364868164, -28.188426971435547, -28.61065673828125, -30.048154830932617, -30.426204681396484, -31.88494110107422, -33.06418228149414, -33.91288757324219, 6.634699821472168, 6.084136486053467, 6.241830825805664, 7.4102091789245605, 5.696282863616943, 6.8732075691223145, 9.031584739685059, 6.211906433105469, 4.569305896759033, 3.941854238510132, 4.253909587860107, 3.5661749839782715, 3.669088840484619, 4.778134346008301, 3.494128704071045, 3.2404539585113525, 3.251777172088623, 3.1045773029327393, 4.017563343048096, 2.887772560119629, 2.7893154621124268, 2.906893014907837, 2.7356178760528564, 3.635089635848999, 2.930985927581787, 3.4188101291656494, 3.0205795764923096, 3.7671284675598145, 3.3917672634124756, 4.333363056182861, 3.613539934158325, 4.097507953643799, 4.1393208503723145, 5.5793843269348145, 8.023341178894043, 5.890435218811035, 8.471368789672852, 6.36739444732666, 5.106582164764404, 4.97397518157959, 6.078526496887207, 5.716768264770508, 6.614887237548828, 6.722918510437012, 7.381338119506836, 7.771437644958496, 8.224390029907227, 8.801892280578613, 10.495489120483398, 9.615983009338379, 9.84536075592041, 10.4797945022583, 10.697399139404297, 11.139212608337402, 11.398507118225098, 11.568729400634766, 11.90092658996582, 11.765414237976074, 12.209936141967773, 11.789682388305664, 12.265124320983887, 11.655421257019043, 12.231867790222168, 14.454611778259277, 12.044990539550781, 10.694457054138184, 10.092092514038086, 9.982656478881836, 11.03602123260498, 9.159738540649414, 8.151397705078125, 7.595863342285156, 7.089807510375977, 6.472036838531494, 5.983863353729248, 5.3483476638793945, 4.860368728637695, 4.220131874084473, 3.800036907196045, 3.0389490127563477, 2.4920849800109863, 1.820901870727539, 1.2539622783660889, 0.7226181030273438]} +{"type": "SampleBatch", "eps_id": [1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 1747731318, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAKrtFz85oVi8dt6LvX6fET5V3Bc/EjI7PkgLhr1+Ui++8ssYP3TywT6VDo29PrX3vnO8Gj9/Jj0+pN+gvd65Wr6Qrhs/SAHDPmWfqb16rQe/xqEdP7i5Ez/AVL+9+PpYvyGWID+bQsQ+RAzivdLuFb+OjCI/G3ZCPoQJ+r1mzqi+d4UjP9Z6V7tWxQO+hHyfvSiBIz+4EEm+n10FvgG4Lz7LfyI/ll7HvvHZAb5DT9c+aIEgP4lZRb5XevK9Gca5PcyEHz+SncW+LcPuvfAIsD7nih0/SvFBvvut4L0ZWIk8qJIcP978w74u/t+9KbyLPu2cGj9wgRO/adDUvdx5Bz+zqRc/53HCvlAjv73rmlI+67cVP3HNEr+3tra9+13vPkrIEj9qZ0S/epCjvYIVOz+02g4/3jESv4Shhb3uAtQ+L+4LP9MLwL4KV2m9ygHQPYwCCj85gDe+DQVhvXGJVb6qFwk/FICIPEkacr1cnQS/gS0JP2XXNb4KRY69eGh6vsBECD8fSL6+OEmYvVZ50DyhXQY/csAzvl8+l72AWJS+jHcFP5gzvb5+HKO9zOmtvDCTAz/PhjG+GvujvYAGrb70rwI/kMS6PKnSsb1Snie/1s0CPz0eL75QpMy9ywDIvrDtAT8Iq7q+YKTcvcsxBr7RDwA/KCIsvocC4r1tT+m++Wb+PsQWub65rPS9W4RMvlGz+j7QAg6/+tr8vbUjWT0gBfU+hVW3vhmv+r0JR42+dFrxPigfDb87/gK+9nHGvF616z4miLW+PH0Dvndwtb7sE+g+ETIMvy2/Cr6Rlta9UnjiPv+WPb+F5Ay+MrARPuzi2j6X+W6/mfoJvhX0xj7RU9E+SjCQv1AFAr42zCI/1srFPjrpqL9u/um9RdZiP4lHuD46w4+/MLPFvZJNDz9Hx6w+joyov3/Frr3ZD1I/ZUufPtxcwb9dKY29lOiKP1PTjz6iR6i/9GtBvTuQRT/0XII+GCvBv48zAr3kS4Y/vNFlPgAkqL+5ATG82Sc/P7LqSj7RGMG/UliHO4OahD9yBSw+3iCov56RyzztnD4/6B4RPqIyj7/TxyI9wG3rPgxr9D0JPqi//HJIPWvHQz+ilL49t0zBv5mMgz16RYk/h7mAPZ9wqL/fea89zvJMP9ulFT0/jcG/j0TQPWY5jz+uMs47rcOov30Z/j1vHFw/CHikvM0HkL+eqBA+RcsbP/hpLr0brm6/SB8dPo8duj6kynq92Fw9vxqRJD6oofs9nrGbvUYVDL9HFSc++sDrvWsbsr1ehj6/wLkkPpeoZT5Rl9C9nz4Nv5pRKT6WwV+8szDnvR+xP78ACik+1Q+nPjPuAr5aInK/trgvPqeIKj8bTRa+CONAvz5dPT41M90+brslvle2D79VNkY+Ur1PPqg6Mb6WK72+9V1KPq3bvLzCyzi+tuo1vhflST4HZn2+LG88vlL7v76w00Q+zovNPREdRL5pbzu+4+FGPjMUAb683Ee+0OUQPAFNRD68FbS+Xq5HvnvvQL7vGD0+4O+ruzOKS75IcsW+bP08PobYrT4McFO+ATUVv5vxQz62LC8/z19fvjEZyL4u9VE+GTLqPjcoPr27GZO8WCSQPPfrGz3KoD+9jCc1PvtgljxG/X6+vyIxvZRPl7w8KVs81FpKPRqmMr2a6Vq+dllrPBxLsT5uKUS9Co+avJtorjxdNm49GrVFvTpuW77n77c8tQe3PgpDV722rZ+8uoHyPJFgkz3R21i94GUzPgNM/jxkTli+v4FKvQdevT4csNs8Xun8vkM1LL2ajTI+l8GKPCCWRb5/7B29iQ29PudIVjz+4vW+x1n/vEUdMj5Rq2M74d87vjfa4rwe7q28lfZMueTn4T29VOa8BREyPgDFAztL0Tq+ItfJvMHrvD7Atta69e/yvrxijbyJDzI+slE2vE6yOr68ymG8gKKsvOAPcrw3yto9hLJovGlwMj49Dk+8NwxDvsiYL7w0bam8RryGvJcZyT22Xza8K90yPn5LbbyTbEy+W0b6u11wvT72Wpe8B2P+vk9gebk7WDM+VMLovEIWV76j+VU750KgvCWWBb10nJY9VVU8Oy48ND7GH/+8z8BqvmGE0TtKzZi8olcSvQ/sWj3/SsU7xFxavsD2Db3tVqs+BirmOiWQkLxdGeW8Fur/PGvntzoSdFm+FfvfvKhDoT5rYzq7NhaKvFBgrLyQ8mE8dHtQu7LYNj7jHaq8USeSvm59zDkAMYW8yuLYvNK8HjoWBYg4EQ5YvmTJ2LzM05E+XSaIuwP8fbw3H6q8E9KDvCpPkru7Ojg+I8KsvOdoob4hOGO66C50vNNo4LxI/O+8z62YujzvOD6NNeW8djapvm9gIDucOGe8u60NvWaLP70H4Q07NbtVvnCCEb2WcXA+cLIDuyTPVrxNjPy8NgmNvbvhFLuow1S+XOoDvX4RWz4qnNK7nArOvq7H5Lw6Gf4+6SttvOniU77td5O845xHPtJ8mLxvKT28kg9nvOq0070TYZq87GpTvop3hLyLQz0+vjS8vF8TNryVXky8WT7nvdsGvrwbmzw+UV5xvICq0b6P2Z+8MbYvvADHu7zf1fi9YpuhvMAmPT4mr8+8NbjXvr5Xg7xsbCS8clsKvWkGDL6r/IS81ZFRvimPFb2wkxQ+pISmvEr9E7xOrAm9h7Eivn7/p7ydi1C+RLAWveft+z2DXcm8i4kDvIOcDL09ZDm+P67KvLQRQD5UcRu9zP33vhjzq7zIVuW7Dx9DvSvEUL6mGK28/T1OvpfSU70HWJY9UxjOvAmgyr4Sz029jNyyPrB3B717wEy+5zAxvQjwKD0C2Re9QuGKu/HPLb2Sj4e+5DEYvct/S76DgEO9tzppPI55KL33x0G7+1VCvWoTlr6Qtyi9uRtKvhZZWr2xKYG8vuI4vZWLyL6+o1u9de6EPhP5WL2OBBa/315GvTrxCD9EfYS9XM3Hvo+MGr3Y2Gg+N3mUvY+1Fb/Z6we9FwgCP05trL3sSMe+Up+8vBbnUT6oXry9f29GvrUJm7zvTMK9o07EvY/6xr77lKq8u2FEPrk51L2g2kW+LymLvCH0272/I9y95rPGvtbBnLx4MTg+LgnsvVpTRb6Pkn68nUPzvcvt8739CC07VL+SvJ4o1b4c0vO9LNZEvkn11ryXcgS+t7H7vZQfxr5XJuy8b6oePqTFBb5+5xS/ZMPSvLl14D4zrxG+fbnFvqTvirzEBQ0+5pcZvhLAFL+3vmi81JnZPk5+Jb4gf8W+9/W6u4/xAj6sZC2+5hpDvlZQTrvzmSq+nEsxvoxuxb5/V9S7WBYAPk8xOb7e9kK+rV2Cu9C0Lb6HFz2+kcGcO7eJ8bsZWe2+cv48vn7MQr77VYi8kWAxvtHjQL6KLcW+WrekvDnQ6T3rxki+CUNCvtoCkrz2OT2+iqlMvojmxL6RSbC8m1bRPc2JVL4kUhS/UoqfvDesxj5pZ2C+vpnEvhnuP7wN1bY9mkRovkA1FL9LrSK8XqjBPuYfdL4uccS+rPEau6rWqD14+3u+cwtBvqiMO7oJDVi+29d/vmhrxL5dt6G7tdmmPZnZg75Q9kC+9aVYu1LfWb6Ux4W+91vEvi/D97vRh6E98LSJvlYaFL+aEsS7jwK9PoOhj76TQsS+xXC3Og/GmD1djpO+HRYUv9N+PTviRbw+xXqZvppKxL5a3ic8IoybPchnnb5XIhS/k8FAPPVivj6tVKO+rSBGv0FNnTxUECs/gEGrvgE/FL8sZAU9mmbDPosvsb6H1cS+yqckPZmTyz1VH7W+5nlCvmnMLD3SqDi+MRG3vqtvxb6WBh49R10APhAEu76Vp0O+fUsoPb+gHr7x+Ly+HohhO8qaGz2CY96+6++8vlzPRL5+C/A8Iw8FvsHnvr6NYCM7ZMHaPMeT07444b6+DKFFvvwMlzym5uW9JtvAvj0+8zqcqIQ8TFLMvkjWwL7KIka+MY0GPIGBz72D0cK+uAfHvoGzyjuOmEY+i8zGvrtYRr675iQ8ljbGvVDIyL5/kKc66i8FPMLFxb72xMi+7JxGvuab0znvcrq9acHKvvnNlzqfwLm6k2bEvmC+yr4omUa+XuoUvJoau73Jusy+YCLHviTaMrxuNEs+WrbQvlt6Fb82p+O7ya75PgKx1r7uAMe+rOE3O6VsRT7oq9q+mnYVv9lK2juzBPk+aabgvpUSx75MQoY8qHxIPqmh5L6PlUa+PVamPM7Cu70Knua+K1fHvuVQlzy2UFQ+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [1, 1, 0, 1, 1, 0, 0, 0, 0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 0, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 1, 1, 1, 0, 1, 0, 0, 1, 1, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [1, 1, 1, 0, 1, 1, 0, 0, 0, 0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 0, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 1, 1, 1, 0, 1, 0, 0, 1, 1, 1, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 0], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAFXcFz8SMjs+SAuGvX5SL77yyxg/dPLBPpUOjb0+tfe+c7waP38mPT6k36C93rlavpCuGz9IAcM+ZZ+pvXqtB7/GoR0/uLkTP8BUv734+li/IZYgP5tCxD5EDOK90u4Vv46MIj8bdkI+hAn6vWbOqL53hSM/1npXu1bFA76EfJ+9KIEjP7gQSb6fXQW+AbgvPst/Ij+WXse+8dkBvkNP1z5ogSA/iVlFvld68r0Zxrk9zIQfP5Kdxb4tw+698AiwPueKHT9K8UG++63gvRlYiTyokhw/3vzDvi7+370pvIs+7ZwaP3CBE79p0NS93HkHP7OpFz/nccK+UCO/veuaUj7rtxU/cc0Sv7e2tr37Xe8+SsgSP2pnRL96kKO9ghU7P7TaDj/eMRK/hKGFve4C1D4v7gs/0wvAvgpXab3KAdA9jAIKPzmAN74NBWG9cYlVvqoXCT8UgIg8SRpyvVydBL+BLQk/Zdc1vgpFjr14aHq+wEQIPx9Ivr44SZi9VnnQPKFdBj9ywDO+Xz6XvYBYlL6MdwU/mDO9vn4co73M6a28MJMDP8+GMb4a+6O9gAatvvSvAj+QxLo8qdKxvVKeJ7/WzQI/PR4vvlCkzL3LAMi+sO0BPwirur5gpNy9yzEGvtEPAD8oIiy+hwLivW1P6b75Zv4+xBa5vrms9L1bhEy+UbP6PtACDr/62vy9tSNZPSAF9T6FVbe+Ga/6vQlHjb50WvE+KB8Nvzv+Ar72cca8XrXrPiaItb48fQO+d3C1vuwT6D4RMgy/Lb8KvpGW1r1SeOI+/5Y9v4XkDL4ysBE+7OLaPpf5br+Z+gm+FfTGPtFT0T5KMJC/UAUCvjbMIj/WysU+Oumov27+6b1F1mI/iUe4PjrDj78ws8W9kk0PP0fHrD6OjKi/f8WuvdkPUj9lS58+3FzBv10pjb2U6Io/U9OPPqJHqL/0a0G9O5BFP/Rcgj4YK8G/jzMCveRLhj+80WU+ACSov7kBMbzZJz8/supKPtEYwb9SWIc7g5qEP3IFLD7eIKi/npHLPO2cPj/oHhE+ojKPv9PHIj3Abes+DGv0PQk+qL/8ckg9a8dDP6KUvj23TMG/mYyDPXpFiT+HuYA9n3Cov995rz3O8kw/26UVPT+Nwb+PRNA9ZjmPP64yzjutw6i/fRn+PW8cXD8IeKS8zQeQv56oED5Fyxs/+GkuvRuubr9IHx0+jx26PqTKer3YXD2/GpEkPqih+z2esZu9RhUMv0cVJz76wOu9axuyvV6GPr/AuSQ+l6hlPlGX0L2fPg2/mlEpPpbBX7yzMOe9H7E/vwAKKT7VD6c+M+4Cvloicr+2uC8+p4gqPxtNFr4I40C/Pl09PjUz3T5uuyW+V7YPv1U2Rj5SvU8+qDoxvpYrvb71XUo+rdu8vMLLOL626jW+F+VJPgdmfb4sbzy+Uvu/vrDTRD7Oi809ER1EvmlvO77j4UY+MxQBvrzcR77Q5RA8AU1EPrwVtL5erke+e+9Avu8YPT7g76u7M4pLvkhyxb5s/Tw+htitPgxwU74BNRW/m/FDPrYsLz/PX1++MRnIvi71UT4ZMuo+0WBnvk7fS75YU1s+e15xPsqgP72MJzU++2CWPEb9fr6/IjG9lE+XvDwpWzzUWko9GqYyvZrpWr52WWs8HEuxPm4pRL0Kj5q8m2iuPF02bj0atUW9Om5bvufvtzy1B7c+CkNXvbatn7y6gfI8kWCTPdHbWL3gZTM+A0z+PGROWL6/gUq9B169Phyw2zxe6fy+QzUsvZqNMj6XwYo8IJZFvn/sHb2JDb0+50hWPP7i9b7HWf+8RR0yPlGrYzvh3zu+N9rivB7urbyV9ky55OfhPb1U5rwFETI+AMUDO0vROr4i18m8weu8PsC21rr17/K+vGKNvIkPMj6yUTa8TrI6vrzKYbyAoqy84A9yvDfK2j2Esmi8aXAyPj0OT7w3DEO+yJgvvDRtqbxGvIa8lxnJPbZfNrwr3TI+fkttvJNsTL5bRvq7XXC9PvZal7wHY/6+T2B5uTtYMz5Uwui8QhZXvqP5VTvnQqC8JZYFvXSclj1VVTw7Ljw0PsYf/7zPwGq+YYTRO0rNmLyiVxK9D+xaPf9KxTvEXFq+wPYNve1Wqz4GKuY6JZCQvF0Z5bwW6v88a+e3OhJ0Wb4V+9+8qEOhPmtjOrs2Foq8UGCsvJDyYTx0e1C7stg2PuMdqrxRJ5K+bn3MOQAxhbzK4ti80rweOhYFiDgRDli+ZMnYvMzTkT5dJoi7A/x9vDcfqrwT0oO8Kk+Su7s6OD4jwqy852ihviE4Y7roLnS802jgvEj877zPrZi6PO84Po015bx2Nqm+b2AgO5w4Z7y7rQ29Zos/vQfhDTs1u1W+cIIRvZZxcD5wsgO7JM9WvE2M/Lw2CY29u+EUu6jDVL5c6gO9fhFbPiqc0rucCs6+rsfkvDoZ/j7pK2286eJTvu13k7zjnEc+0nyYvG8pPbySD2e86rTTvRNhmrzsalO+ineEvItDPT6+NLy8XxM2vJVeTLxZPue92wa+vBubPD5RXnG8gKrRvo/Zn7wxti+8AMe7vN/V+L1im6G8wCY9Piavz7w1uNe+vleDvGxsJLxyWwq9aQYMvqv8hLzVkVG+KY8VvbCTFD6khKa8Sv0TvE6sCb2HsSK+fv+nvJ2LUL5EsBa95+37PYNdybyLiQO8g5wMvT1kOb4/rsq8tBFAPlRxG73M/fe+GPOrvMhW5bsPH0O9K8RQvqYYrbz9PU6+l9JTvQdYlj1TGM68CaDKvhLPTb2M3LI+sHcHvXvATL7nMDG9CPAoPQLZF71C4Yq78c8tvZKPh77kMRi9y39LvoOAQ723Omk8jnkovffHQbv7VUK9ahOWvpC3KL25G0q+FllavbEpgby+4ji9lYvIvr6jW7117oQ+E/lYvY4EFr/fXka9OvEIP0R9hL1czce+j4wavdjYaD43eZS9j7UVv9nrB70XCAI/Tm2svexIx75Sn7y8FudRPqhevL1/b0a+tQmbvO9Mwr2jTsS9j/rGvvuUqry7YUQ+uTnUvaDaRb4vKYu8IfTbvb8j3L3ms8a+1sGcvHgxOD4uCey9WlNFvo+SfrydQ/O9y+3zvf0ILTtUv5K8nijVvhzS870s1kS+SfXWvJdyBL63sfu9lB/Gvlcm7Lxvqh4+pMUFvn7nFL9kw9K8uXXgPjOvEb59ucW+pO+KvMQFDT7mlxm+EsAUv7e+aLzUmdk+Tn4lviB/xb739bq7j/ECPqxkLb7mGkO+VlBOu/OZKr6cSzG+jG7Fvn9X1LtYFgA+TzE5vt72Qr6tXYK70LQtvocXPb6RwZw7t4nxuxlZ7b5y/jy+fsxCvvtViLyRYDG+0eNAvootxb5at6S8OdDpPevGSL4JQ0K+2gKSvPY5Pb6KqUy+iObEvpFJsLybVtE9zYlUviRSFL9Sip+8N6zGPmlnYL6+mcS+Ge4/vA3Vtj2aRGi+QDUUv0utIrxeqME+5h90vi5xxL6s8Rq7qtaoPXj7e75zC0G+qIw7ugkNWL7b13++aGvEvl23obu12aY9mdmDvlD2QL71pVi7Ut9ZvpTHhb73W8S+L8P3u9GHoT3wtIm+VhoUv5oSxLuPAr0+g6GPvpNCxL7FcLc6D8aYPV2Ok74dFhS/0349O+JFvD7Fepm+mkrEvlreJzwijJs9yGedvlciFL+TwUA89WK+Pq1Uo76tIEa/QU2dPFQQKz+AQau+AT8UvyxkBT2aZsM+iy+xvofVxL7KpyQ9mZPLPVUftb7meUK+acwsPdKoOL4xEbe+q2/FvpYGHj1HXQA+EAS7vpWnQ759Syg9v6AevvH4vL4eiGE7ypobPYJj3r7r77y+XM9Evn4L8DwjDwW+wee+vo1gIztkwdo8x5PTvjjhvr4MoUW+/AyXPKbm5b0m28C+PT7zOpyohDxMUsy+SNbAvsoiRr4xjQY8gYHPvYPRwr64B8e+gbPKO46YRj6LzMa+u1hGvrvmJDyWNsa9UMjIvn+QpzrqLwU8wsXFvvbEyL7snEa+5pvTOe9yur1pwcq++c2XOp/AubqTZsS+YL7KviiZRr5e6hS8mhq7vcm6zL5gIse+JNoyvG40Sz5attC+W3oVvzan47vJrvk+ArHWvu4Ax76s4Tc7pWxFPuir2r6adhW/2UraO7ME+T5ppuC+lRLHvkxChjyofEg+qaHkvo+VRr49VqY8zsK7vQqe5r4rV8e+5VCXPLZQVD6pmuq+ySZHvlRJuTwWvKK9lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.5154235363006592, 0.2407984882593155, 0.8930491209030151, 0.20471283793449402, 0.0959024727344513, 0.942366361618042, 0.9147151112556458, 0.8594895005226135, 0.7571074962615967, 0.6010606288909912, 0.579083263874054, 0.661003828048706, 0.5139110088348389, 0.7095549702644348, 0.5481089949607849, 0.6325888633728027, 0.593794584274292, 0.40812650322914124, 0.7593300938606262, 0.5699806213378906, 0.3511321544647217, 0.19731508195400238, 0.8849126100540161, 0.8208212852478027, 0.2828945517539978, 0.8361754417419434, 0.2529817819595337, 0.14979971945285797, 0.9049645662307739, 0.8643420934677124, 0.19566157460212708, 0.8765650987625122, 0.8275020718574524, 0.24515573680400848, 0.8461055159568787, 0.21365191042423248, 0.8620225787162781, 0.8132255673408508, 0.7391813397407532, 0.6320911645889282, 0.49179768562316895, 0.6634432077407837, 0.5171874165534973, 0.35053154826164246, 0.7950902581214905, 0.3392367362976074, 0.8140704035758972, 0.3081779181957245, 0.8440852761268616, 0.7421359419822693, 0.42034468054771423, 0.2058154195547104, 0.9088913202285767, 0.1409517526626587, 0.9379616975784302, 0.9134494066238403, 0.8646001815795898, 0.7633389234542847, 0.5767053961753845, 0.6558250188827515, 0.7236817479133606, 0.5068705081939697, 0.16148389875888824, 0.944336473941803, 0.9127411842346191, 0.8364307880401611, 0.6637890934944153, 0.5920984745025635, 0.8096596002578735, 0.6075601577758789, 0.6472212672233582, 0.239411860704422, 0.07584802806377411, 0.9601652026176453, 0.944404125213623, 0.6352842450141907, 0.7237421274185181, 0.35462144017219543, 0.8774682879447937, 0.32610297203063965, 0.8871390223503113, 0.7081003189086914, 0.35540705919265747, 0.8692069053649902, 0.35613197088241577, 0.8721439242362976, 0.6560870409011841, 0.7107724547386169, 0.34210875630378723, 0.8800112009048462, 0.6835890412330627, 0.6796474456787109, 0.6989685297012329, 0.6628954410552979, 0.28434285521507263, 0.9001704454421997, 0.7520711421966553, 0.5877953767776489, 0.7782660722732544, 0.4555862843990326, 0.8298412561416626, 0.4791680872440338, 0.8212074637413025, 0.5041744112968445, 0.8197965025901794, 0.5323110818862915, 0.7942827343940735, 0.4500851035118103, 0.8427754640579224, 0.41224443912506104, 0.8581467866897583, 0.6321430802345276, 0.7220175266265869, 0.6597849130630493, 0.3012312650680542, 0.8910598754882812, 0.6973886489868164, 0.680000901222229, 0.6864230632781982, 0.309643417596817, 0.8883459568023682, 0.2807754874229431, 0.8978433609008789, 0.7537546753883362, 0.5817849040031433, 0.776976466178894, 0.5429124236106873, 0.20139668881893158, 0.9196030497550964, 0.8297119736671448, 0.5751257538795471, 0.7580527067184448, 0.389935702085495, 0.8625457882881165, 0.34107786417007446, 0.8785499930381775, 0.7080844640731812, 0.36989837884902954, 0.8616166114807129, 0.3872119188308716, 0.8587493300437927, 0.6092347502708435, 0.7477098703384399, 0.5950357913970947, 0.7570720314979553, 0.5818967819213867, 0.23456957936286926, 0.9069869518280029, 0.7869290709495544, 0.47435855865478516, 0.8227680921554565, 0.4748462438583374, 0.8261275887489319, 0.5362269282341003, 0.7895361185073853, 0.5338279604911804, 0.20864850282669067, 0.9144184589385986, 0.8064946532249451, 0.494231641292572, 0.8152246475219727, 0.5249249339103699, 0.7980294823646545, 0.5219731330871582, 0.8025006055831909, 0.4904705882072449, 0.8125725984573364, 0.48787981271743774, 0.8144630789756775, 0.5175417065620422, 0.8051778078079224, 0.5016656517982483, 0.8153951168060303, 0.47676408290863037, 0.17097119987010956, 0.9315175414085388, 0.8540765047073364, 0.6266636848449707, 0.7042249441146851, 0.6643496751785278, 0.3308687210083008, 0.8660773634910583, 0.3480274975299835, 0.8616894483566284, 0.3568207323551178, 0.8604270815849304, 0.6429173946380615, 0.7116952538490295, 0.3702296018600464, 0.8559997081756592, 0.36393237113952637, 0.8605296015739441, 0.6505918502807617, 0.3036905527114868, 0.8867617845535278, 0.290048211812973, 0.8919240236282349, 0.7318577170372009, 0.588309645652771, 0.7470623850822449], "advantages": [6.995183944702148, 6.387371063232422, 6.1750898361206055, 5.257962226867676, 5.12085485458374, 6.14564323425293, 4.199608325958252, 3.053567409515381, 2.3646302223205566, 1.8006842136383057, 1.200494408607483, 0.36818861961364746, -0.19588571786880493, -1.0585012435913086, -1.594315528869629, -2.2148935794830322, -3.1124563217163086, -3.7041049003601074, -4.381546974182129, -5.305499076843262, -6.269021987915039, -7.244521617889404, -8.026840209960938, -8.60002326965332, -9.063530921936035, -10.0787353515625, -10.535754203796387, -11.564391136169434, -12.162471771240234, -12.858068466186523, -13.282214164733887, -14.26744270324707, -14.668869018554688, -14.946483612060547, -16.243879318237305, -16.416399002075195, -17.739198684692383, -17.798625946044922, -17.940139770507812, -18.42546272277832, -19.29154396057129, -20.429227828979492, -21.683347702026367, -22.71961784362793, -23.920515060424805, -25.268037796020508, -26.362838745117188, -27.77870750427246, -28.72662353515625, -30.143386840820312, -31.615825653076172, -32.11310958862305, -32.814266204833984, -34.17627716064453, -34.845916748046875, -36.362552642822266, -38.207576751708984, -40.74843978881836, -44.385318756103516, -48.55998229980469, -47.21519088745117, -51.216514587402344, -50.100730895996094, -49.44249725341797, -53.1641731262207, -56.30058670043945, -58.3718376159668, -59.67190170288086, -60.38432312011719, -61.772830963134766, -62.7506103515625, -63.938621520996094, -64.62114715576172, -64.35713195800781, -66.89244842529297, 15.066789627075195, 15.24238395690918, 14.751054763793945, 15.03076457977295, 14.266862869262695, 14.544379234313965, 13.76075267791748, 13.799358367919922, 14.878401756286621, 13.524380683898926, 14.668540954589844, 13.277851104736328, 12.860018730163574, 12.87961483001709, 14.106449127197266, 12.672066688537598, 12.262079238891602, 12.309367179870605, 11.899307250976562, 11.95831298828125, 13.383491516113281, 11.849344253540039, 11.427570343017578, 11.58701229095459, 11.152270317077637, 11.654766082763672, 10.685210227966309, 11.14793872833252, 10.22532844543457, 10.474597930908203, 9.987019538879395, 10.419861793518066, 9.556634902954102, 9.893877983093262, 9.364506721496582, 9.789912223815918, 9.23778247833252, 9.656144142150879, 8.917020797729492, 9.295891761779785, 10.578024864196777, 8.669842720031738, 8.067720413208008, 8.319108963012695, 7.764983654022217, 8.458747863769531, 7.758238792419434, 8.604207038879395, 7.878355979919434, 8.139037132263184, 7.812491416931152, 8.045218467712402, 7.825507640838623, 9.143359184265137, 8.328527450561523, 8.552919387817383, 9.509198188781738, 8.428767204284668, 8.644680976867676, 8.75616455078125, 9.148237228393555, 9.199009895324707, 9.889113426208496, 10.870268821716309, 9.511799812316895, 10.345779418945312, 9.130401611328125, 8.706911087036133, 9.063611030578613, 8.719443321228027, 8.979187965393066, 8.71177864074707, 9.556770324707031, 9.068345069885254, 9.08609676361084, 9.329458236694336, 8.664011001586914, 8.830598831176758, 8.207993507385254, 8.2406644821167, 8.007857322692871, 8.130121231079102, 9.222833633422852, 8.401140213012695, 7.885528564453125, 8.33476734161377, 7.7013654708862305, 7.285739898681641, 7.217772960662842, 6.732208251953125, 6.71940803527832, 7.326486587524414, 6.518040180206299, 7.225553512573242, 6.329440116882324, 5.516270637512207, 5.814112186431885, 4.95138692855835, 5.2736663818359375, 4.37452507019043, 3.90032958984375, 3.535935401916504, 3.72714900970459, 4.267701148986816, 3.352041006088257, 3.9003024101257324, 4.855912685394287, 3.9139962196350098, 4.891931533813477, 3.907681703567505, 4.9063568115234375, 3.884683609008789, 2.57891845703125, 3.3908727169036865, 4.380886554718018, 3.3223631381988525, 4.32577657699585, 3.2390105724334717, 1.7951972484588623, 0.121520034968853, 0.8420259952545166, -0.7820897102355957, -0.1367400735616684, 0.6520874500274658, -0.7575302124023438]} +{"type": "SampleBatch", "eps_id": [627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 627353008, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAKma6r7JJke+VEm5PBa8or19mOy+8aPHvoZErDz6jmE+ppbwvkPKR75mW9A85Y+GvRyW8r64+se+kZfFPPeHcD4Blva+5wYWv7kT7Dxp+wg/SJb8vmtdyL5v3yE9fNSAPhNMAL9khkm+T3w2Pbkn6LwHTgG/RMAZu/4pND3U/5y+GlEBv+PRSr5NCxs9L8rZubZUAr9elMm+lgIbPXGhmz7BWAS/qt8WvzjpMz0qvBs/Ol0Hv0gsyr4Iv2U9s9yoPspiCb8aZ02+0GGAPemfYj21aQq/tAjLvvilgj0Z9rs+eXEMv3o/T75rr5E9pPvCPcB6Db9zAsy+u5WVPWOk0T4EhQ+/J1lRvjRbpj29ChA++5AQv3UzLbwxHqw9X1UAvtaeEL+Apjs+DvymPbfcx76lrg+/eMlTvOH+lj0gwJW9lr8PvzBpOT4oAJQ9Me6uvkPSDr9y3XW8lQGGPdny3bzu5Q6/GRFYvn3lhD13LJI+f/oPvwwlirwgl5A9vshkPJkQEL9bbjU+jCmRPZPigr5eKA+/2qmavAWxhj1Rw289HUEPv0ByMz7QFok9UcxZvmxbDr/cKKq8kGCAPWN6zT2mdg6/Su9dvpx8hD2IANM+uZIPvzUEubzuXZU93eAPPlSwD79161++Px+bPSQa6T7yzhC/0F3Uvi3FrT3aVUU/mu4Sv944Yr4JWM09P4oBPysQFL8v/9+8/hHiPWF0fD4CNBS/NgcqPiEr7D0X1Bm8X1oTvz37tz6tyOs9xPOGvmGDEb80ryY+2/zgPUUygT0GrhC/ilu2PleS4z2NpkW+MNsOv5CuDD9mqts9ALrlvuUKDL9fxrQ+mUnJPSNi/r0cPAq/E/ILPygzxD0vyMS+l28Hvy1dsz4QdbQ9uiKBvWukBb/+vx0+49+xPXscgz5/2gS/UheyPgxdvD0BBwO8lhIDv2GhCj8wCbw9K2WKvsxMAL/MwLA+2/awPaFCTD2eEP2+MvwJP8MBsz2Xflu+p4v3viaXOz8jOqo90tr0vroK8L6TXQk/g6OWPehnJL4ajOq+wVSuPv8PkD2CGh4+hw/nvoHXCD/6YpY9Uh3svUSW4b5NgTo/EqqRPc+SxL50INq+KS5sPz/wgT01Qye/+K3Qvq8EOj9YWk49OMyuviQ9yb6o5Ac/oWIyPXtDCL2YzcO+hq05P/aoLz0QsZ++QWC8vr+TBz/9GxY9HcDEu/Lztr6XYzk/Ep4VPfHkkr6Pia++OU8HP486/DzW3Is8/h+qvu4kOT+nBv884g6Ivh24or5w+mo/xXzTPLVmDL/wUZm+yWqOP+BCczyqN1W/PO2NvvjUaj+3Qe260RoJv46IhL4h4zg/syZNvNlLeb6fRnq+N+JqP4Z2jrxWQQq/NHxnvgD/OD9D8ua88XyBvnyvWL4jDGs/9jAIvabpDb+24UW+Ozg5P3OaNb3JbYu+ahA3vvVvBz906Uu9AmDGO6g6LL7bjzk/f2pLvR6Umr5ZYh2+9MwHPwkmZL2QN8+8J4USvsvxOT+COGa9loervgKlA76JNQg/LtWAvT0feL3kfvG9SmA6P2BQg715rr6+7azTvXmLbD+IkZK9+0Ivv62997xvvoC8E1lEvRzc7DzYUPq8XClXvrb6Qb17uJw+7l4OvUwxa7xv5yi9HFmVuvqLD73w31W+Vf8ovS93jj4iqCC9auJXvPEzEr01bt68d7whvWDBVL5dbRS9+hOCPrLBMr27C0e8vDr/vNUaTL15wDO9H8ZTvmKyA71rc24+mrFEvZ45OLzSPeG8ZeyOvWmdRb2L51K+5KzsvPk7Wz68fFa9ISPNvg+Zybziu/4+KE93vY4fUr6CKnC8iuxJPj0PhL23tSG85owvvGv2zL28doS9g9o9PipYULyx0Mq+RL15vTg6HLy0Eqm8wBzcvT2Fer2WQlG+mq66vDXpNj5xoYW9bRASvJJqnbyqIvi97f6FvYulUL5iRLG8al4pPndXjr1VEMy+CSuWvMkA5z6xqp697g9Qvoh+GLyrcBw+QP2mvVn0Abzp3cy7UUISvmxQp70axj8+fzwVvFz9376opJ+94Lr8u3xLkrz8ORe+iPWfvaxvT765faq8R6gOPq1BqL0kpOq7eqqTvMOyI77DjKi9VN5Ovo/brbxsIQI+GNOwvVNJ2LtpCZm83Fowvk8Ysb32SE6+6UC1vC+E6j2qWLm9VTXFuwR+orwjgz2+xpe5vd6rTb5v0MC873DPPdnRwb0J8LC7CjiwvJ9+S753CsK9MQNNvi7H0LznXrI9yz3KvYw3yr4mgsK8O0a+PjNr2r2p+Ba/4J6FvJwfKD/8kvK9Cu3JvjIsULthybc+N10BvjsATL4cKYM7Iq+FPbNxBb6i7sm+g/CtO3/rtz58hQ2+GSNMvrGtTDzcs4s9qpoRvpSSkLvpB2M8FsthvsyxEb7UhUy+7MYaPKm2nD30yBW+jHObu+jZMzyDSFq+0+EVvlnSTL5JANw77+WpPYP6Gb49tqO7LC8JPNiUVL61FBq+3gpNvgJRijsQo7M9hi4evlt0yr7ezMM7d3HDPqhHJr4JMU2+2vtePJ46uj09Yiq+79myu8rHfDwMKkq+234qvs+dTb5/Fjw8rvjMPZybLr4bbL+7IOJcPIJ9Qb49ui6+TftNvmf3HjzOFN093dgyvrn2yr7kVkI84q7OPjf3Or4ETE6+7U6jPFcG6z10Fz++/EnZuzwctjzdry++ODo/vuTpTr4ZAJo8qx0DPp5dQ76wd+y7mvquPER1Ir50g0O+OIBPvlL8lDwAEhA+26lHvo31/rtxCaw8kbQVvqbSR77iElC+gxWUPLW1HD79+0u+SZ4IvFgorTwqGgm+tSdMvp2lUL6jOJc8M1spPvpTUL564RG8eFGyPNeo+L2pglC+dlw+PuxsnjxPdtC+A7RMvp/wwj5obzc8xpQxv9PnRL706j0+iHkvu86Dy75yG0G+d+HCPkIeLrxq5jC/3k85vvYbPj5vRsi8SKvNvoKCNb7aTxW8bAsFvQ5A771KsjW+MKVQvlidDr33aCk+jd45viOSBbzUDwG9ZFUNvkwJOr5ArU++V14MvcMGFD6aMD6+nGHsu8KGAL0IjCK+bFY+vry3Tr64hw29uLb9PdF4Qr74j827sWEDvRbON761mUK+d75NvgUWEr2vu9I9HrdGvvKAyr4cqAm9E6vEPsHQTr4Yu0y+KGHUvHDzpT366FK+oBLKvnsax7yxGLs+M/5avun+S76LO4u8LXuFPakSX74Cxsm+2Y2AvCt0tD7SJGe+qscWvyaeDbxe4CM/yzRzvqaYyb7FSYg7EoOwPiNFe76zwxa/nBw1PJGEIz9mqoO+zbTJvhE1wzyN+bI+IrOHvpXjFr+vevw8LVAmPzy8jb4mGsq+uHUzPXvSuz4Ax5G+iScXv+qCUT1hSiw/0tKXvvHJyr56UoQ9RTHLPhnhm778xE6+3JOUPZE7AD5u8p2+yZ8BvPa0mT3kKRO+KweevpjrUL4C0pM9TcovPgEeoL5K7CO8GdqaPbN4x707OKC+Dl88Ps7clj2rCLu+AFaevsrARrxa5oc9nSZOvc11nr7uWDo+m9aFPVuJpL7BmJy+mIVlvMtZcT1mRBC8er2cvsWLOD4ioXA9o42Qvgrlmr7ouL8+O4BZPeBQDr9sD5e+Ses2Prb1Kz2s63y+JjuVvuIKvz7luRc9V7EGvwNpkb5axDU+vD/ZPMhZY76wl4++DFOTvHPftDzamaE91caPvucONT4JzcE80LBTvlL3jb6WrZi8MO6fPGEfvz0uKI6+pytbvmA4rzxBwcg+QVmQvuR2nbw4du88YZLZPaWLkL5ArjM+C28APTxVNb6pv46+iYG9PrHa4zxRX+u+ZPWKvurQMj79iJg8wDEivp8rib6fDaq8BSt9PNN3Dz4KYom+3UhdvvaJlTwgEOA+h5iLvjn+rbw0Pd08LF4aPjTQi75r8F2+HPD1PCJW5z5eCI6+vKe0vJj7Hz0Fzyw+LkKOvvygMD64zi09MQzkvQJ+jL4ZLL68gq8kPcMTRz7duoy+52svPpycND1Qva69yfmKvp1Guz5Hny09g0C6vu86h762bQ8/ZdIPPTeOJL85foG+nba6PgRUtjwftK2+gYR7vgpOLT5vfH08MGUivS8NeL6GcLo+lH5wPCSip74LmHC+B90sPoc1BTxG2va8+yJtviJGuj6Eq/Y7gPajvomvZb5Uoiw+nzCTOtJgzrymO2K+gX/ZvD1MIjrmIok+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAH2Y7L7xo8e+hkSsPPqOYT6mlvC+Q8pHvmZb0Dzlj4a9HJbyvrj6x76Rl8U894dwPgGW9r7nBha/uRPsPGn7CD9Ilvy+a13Ivm/fIT181IA+E0wAv2SGSb5PfDY9uSfovAdOAb9EwBm7/ik0PdT/nL4aUQG/49FKvk0LGz0vytm5tlQCv16Uyb6WAhs9caGbPsFYBL+q3xa/OOkzPSq8Gz86XQe/SCzKvgi/ZT2z3Kg+ymIJvxpnTb7QYYA96Z9iPbVpCr+0CMu++KWCPRn2uz55cQy/ej9PvmuvkT2k+8I9wHoNv3MCzL67lZU9Y6TRPgSFD78nWVG+NFumPb0KED77kBC/dTMtvDEerD1fVQC+1p4Qv4CmOz4O/KY9t9zHvqWuD794yVO84f6WPSDAlb2Wvw+/MGk5PigAlD0x7q6+Q9IOv3LddbyVAYY92fLdvO7lDr8ZEVi+feWEPXcskj5/+g+/DCWKvCCXkD2+yGQ8mRAQv1tuNT6MKZE9k+KCvl4oD7/aqZq8BbGGPVHDbz0dQQ+/QHIzPtAWiT1RzFm+bFsOv9woqryQYIA9Y3rNPaZ2Dr9K712+nHyEPYgA0z65kg+/NQS5vO5dlT3d4A8+VLAPv3XrX74/H5s9JBrpPvLOEL/QXdS+LcWtPdpVRT+a7hK/3jhivglYzT0/igE/KxAUvy//37z+EeI9YXR8PgI0FL82Byo+ISvsPRfUGbxfWhO/Pfu3Pq3I6z3E84a+YYMRvzSvJj7b/OA9RTKBPQauEL+KW7Y+V5LjPY2mRb4w2w6/kK4MP2aq2z0AuuW+5QoMv1/GtD6ZSck9I2L+vRw8Cr8T8gs/KDPEPS/IxL6Xbwe/LV2zPhB1tD26IoG9a6QFv/6/HT7j37E9exyDPn/aBL9SF7I+DF28PQEHA7yWEgO/YaEKPzAJvD0rZYq+zEwAv8zAsD7b9rA9oUJMPZ4Q/b4y/Ak/wwGzPZd+W76ni/e+Jpc7PyM6qj3S2vS+ugrwvpNdCT+Do5Y96GckvhqM6r7BVK4+/w+QPYIaHj6HD+e+gdcIP/pilj1SHey9RJbhvk2BOj8SqpE9z5LEvnQg2r4pLmw/P/CBPTVDJ7/4rdC+rwQ6P1haTj04zK6+JD3JvqjkBz+hYjI9e0MIvZjNw76GrTk/9qgvPRCxn75BYLy+v5MHP/0bFj0dwMS78vO2vpdjOT8SnhU98eSSvo+Jr745Twc/jzr8PNbcizz+H6q+7iQ5P6cG/zziDoi+HbiivnD6aj/FfNM8tWYMv/BRmb7Jao4/4EJzPKo3Vb887Y2++NRqP7dB7brRGgm/joiEviHjOD+zJk282Ut5vp9Ger434mo/hnaOvFZBCr80fGe+AP84P0Py5rzxfIG+fK9YviMMaz/2MAi9pukNv7bhRb47ODk/c5o1vclti75qEDe+9W8HP3TpS70CYMY7qDosvtuPOT9/aku9HpSavlliHb70zAc/CSZkvZA3z7wnhRK+y/E5P4I4Zr2Wh6u+AqUDvok1CD8u1YC9PR94veR+8b1KYDo/YFCDvXmuvr7trNO9eYtsP4iRkr37Qi+/DdStvZzdOj9AnK69hqHUvthQ+rxcKVe+tvpBvXu4nD7uXg69TDFrvG/nKL0cWZW6+osPvfDfVb5V/yi9L3eOPiKoIL1q4le88TMSvTVu3rx3vCG9YMFUvl1tFL36E4I+ssEyvbsLR7y8Ov+81RpMvXnAM70fxlO+YrIDvWtzbj6asUS9njk4vNI94bxl7I69aZ1FvYvnUr7krOy8+TtbPrx8Vr0hI82+D5nJvOK7/j4oT3e9jh9SvoIqcLyK7Ek+PQ+Evbe1IbzmjC+8a/bMvbx2hL2D2j0+KlhQvLHQyr5EvXm9ODocvLQSqbzAHNy9PYV6vZZCUb6arrq8Nek2PnGhhb1tEBK8kmqdvKoi+L3t/oW9i6VQvmJEsbxqXik+d1eOvVUQzL4JK5a8yQDnPrGqnr3uD1C+iH4YvKtwHD5A/aa9WfQBvOndzLtRQhK+bFCnvRrGPz5/PBW8XP3fvqikn73guvy7fEuSvPw5F76I9Z+9rG9Pvrl9qrxHqA4+rUGovSSk6rt6qpO8w7IjvsOMqL1U3k6+j9utvGwhAj4Y07C9U0nYu2kJmbzcWjC+TxixvfZITr7pQLW8L4TqPapYub1VNcW7BH6ivCODPb7Gl7m93qtNvm/QwLzvcM892dHBvQnwsLsKOLC8n35LvncKwr0xA02+LsfQvOdesj3LPcq9jDfKviaCwrw7Rr4+M2vavan4Fr/gnoW8nB8oP/yS8r0K7cm+MixQu2HJtz43XQG+OwBMvhwpgzsir4U9s3EFvqLuyb6D8K07f+u3PnyFDb4ZI0y+sa1MPNyziz2qmhG+lJKQu+kHYzwWy2G+zLERvtSFTL7sxho8qbacPfTIFb6Mc5u76NkzPINIWr7T4RW+WdJMvkkA3Dvv5ak9g/oZvj22o7ssLwk82JRUvrUUGr7eCk2+AlGKOxCjsz2GLh6+W3TKvt7Mwzt3ccM+qEcmvgkxTb7a+148njq6PT1iKr7v2bK7ysd8PAwqSr7bfiq+z51Nvn8WPDyu+Mw9nJsuvhtsv7sg4lw8gn1Bvj26Lr5N+02+Z/cePM4U3T3d2DK+ufbKvuRWQjzirs4+N/c6vgRMTr7tTqM8VwbrPXQXP778Sdm7PBy2PN2vL744Oj++5OlOvhkAmjyrHQM+nl1DvrB37Lua+q48RHUivnSDQ744gE++UvyUPAASED7bqUe+jfX+u3EJrDyRtBW+ptJHvuISUL6DFZQ8tbUcPv37S75Jngi8WCitPCoaCb61J0y+naVQvqM4lzwzWyk++lNQvnrhEbx4UbI816j4vamCUL52XD4+7GyePE920L4DtEy+n/DCPmhvNzzGlDG/0+dEvvTqPT6IeS+7zoPLvnIbQb534cI+Qh4uvGrmML/eTzm+9hs+Pm9GyLxIq82+goI1vtpPFbxsCwW9DkDvvUqyNb4wpVC+WJ0OvfdoKT6N3jm+I5IFvNQPAb1kVQ2+TAk6vkCtT75XXgy9wwYUPpowPr6cYey7woYAvQiMIr5sVj6+vLdOvriHDb24tv090XhCvviPzbuxYQO9Fs43vrWZQr53vk2+BRYSva+70j0et0a+8oDKvhyoCb0Tq8Q+wdBOvhi7TL4oYdS8cPOlPfroUr6gEsq+exrHvLEYuz4z/lq+6f5Lvos7i7wte4U9qRJfvgLGyb7ZjYC8K3S0PtIkZ76qxxa/Jp4NvF7gIz/LNHO+ppjJvsVJiDsSg7A+I0V7vrPDFr+cHDU8kYQjP2aqg77NtMm+ETXDPI35sj4is4e+leMWv696/DwtUCY/PLyNviYayr64dTM9e9K7PgDHkb6JJxe/6oJRPWFKLD/S0pe+8cnKvnpShD1FMcs+GeGbvvzETr7ck5Q9kTsAPm7ynb7JnwG89rSZPeQpE74rB56+mOtQvgLSkz1Nyi8+AR6gvkrsI7wZ2po9s3jHvTs4oL4OXzw+ztyWPasIu74AVp6+ysBGvFrmhz2dJk69zXWevu5YOj6b1oU9W4mkvsGYnL6YhWW8y1lxPWZEELx6vZy+xYs4PiKhcD2jjZC+CuWavui4vz47gFk94FAOv2wPl75J6zY+tvUrPazrfL4mO5W+4gq/PuW5Fz1XsQa/A2mRvlrENT68P9k8yFljvrCXj74MU5O8c9+0PNqZoT3Vxo++5w41PgnNwTzQsFO+UveNvpatmLww7p88YR+/PS4ojr6nK1u+YDivPEHByD5BWZC+5HadvDh27zxhktk9pYuQvkCuMz4LbwA9PFU1vqm/jr6Jgb0+sdrjPFFf675k9Yq+6tAyPv2ImDzAMSK+nyuJvp8NqrwFK30803cPPgpiib7dSF2+9omVPCAQ4D6HmIu+Of6tvDQ93TwsXho+NNCLvmvwXb4c8PU8IlbnPl4Ijr68p7S8mPsfPQXPLD4uQo6+/KAwPrjOLT0xDOS9An6MvhksvryCryQ9wxNHPt26jL7nay8+nJw0PVC9rr3J+Yq+nUa7PkefLT2DQLq+7zqHvrZtDz9l0g89N44kvzl+gb6dtro+BFS2PB+0rb6BhHu+Ck4tPm98fTwwZSK9Lw14voZwuj6UfnA8JKKnvguYcL4H3Sw+hzUFPEba9rz7Im2+Ika6PoSr9juA9qO+ia9lvlSiLD6fMJM60mDOvKY7Yr6Bf9m8PUwiOuYiiT7ZxmK+y5ssPlzSwzuS4Mm8lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.563431441783905, 0.7628344297409058, 0.5356295704841614, 0.2208554744720459, 0.9120644927024841, 0.805221676826477, 0.5547137260437012, 0.7263613939285278, 0.41675612330436707, 0.16308242082595825, 0.9276096820831299, 0.8583623766899109, 0.3149293065071106, 0.8755647540092468, 0.26730912923812866, 0.8908889889717102, 0.7769559025764465, 0.5622466802597046, 0.6718210577964783, 0.6092480421066284, 0.632044792175293, 0.35083258152008057, 0.8472351431846619, 0.6944887042045593, 0.5346443057060242, 0.7282666563987732, 0.4901244044303894, 0.2427688091993332, 0.8860976099967957, 0.21124772727489471, 0.10345065593719482, 0.9379718899726868, 0.9069287776947021, 0.8476080894470215, 0.7441242337226868, 0.4099231958389282, 0.7811298966407776, 0.6485039591789246, 0.5273187160491943, 0.6908736824989319, 0.4775017201900482, 0.27513980865478516, 0.8451710343360901, 0.7583397030830383, 0.3778783679008484, 0.783466637134552, 0.6621564030647278, 0.5131763815879822, 0.31020182371139526, 0.8190524578094482, 0.718977689743042, 0.5592713356018066, 0.6369394063949585, 0.42603304982185364, 0.7499458193778992, 0.4087483584880829, 0.7619744539260864, 0.3944096863269806, 0.7722060084342957, 0.6174934506416321, 0.39916980266571045, 0.7937448620796204, 0.6316325068473816, 0.5874954462051392, 0.665071964263916, 0.5555748343467712, 0.7094094753265381, 0.4930602014064789, 0.7237469553947449, 0.5428276062011719, 0.6925302743911743, 0.6023407578468323, 0.6485161781311035, 0.3296322226524353, 0.8773176074028015, 0.5151809453964233, 0.7969620227813721, 0.5552796721458435, 0.7759831547737122, 0.5895014405250549, 0.7554612159729004, 0.6187127232551575, 0.735506534576416, 0.6438470482826233, 0.2839134633541107, 0.8970435857772827, 0.7174158692359924, 0.34208860993385315, 0.8786332011222839, 0.6870452165603638, 0.6770513653755188, 0.7033825516700745, 0.34064072370529175, 0.8805349469184875, 0.6641690731048584, 0.29057860374450684, 0.8934840559959412, 0.7339879870414734, 0.6211369633674622, 0.7483421564102173, 0.600459098815918, 0.762718677520752, 0.5777633190155029, 0.7773151993751526, 0.5524963140487671, 0.7922576665878296, 0.4758961796760559, 0.17312873899936676, 0.9298287630081177, 0.8380169868469238, 0.44677990674972534, 0.8466207385063171, 0.57846999168396, 0.7640720009803772, 0.5881628394126892, 0.7588244080543518, 0.595130443572998, 0.7550949454307556, 0.4003913998603821, 0.8638266921043396, 0.6254569888114929, 0.7293568849563599, 0.6374799013137817, 0.7203916311264038, 0.35219213366508484, 0.880338728427887, 0.6779037117958069, 0.6782917976379395, 0.6959462761878967, 0.6594412922859192, 0.7129769325256348, 0.6396936774253845, 0.7293673753738403, 0.6186374425888062, 0.7454165816307068, 0.4041803181171417, 0.15459243953227997, 0.930853545665741, 0.1416236311197281, 0.9363786578178406, 0.8771600723266602, 0.692272961139679, 0.6618179678916931, 0.7188828587532043, 0.6307380795478821, 0.743956446647644, 0.5969002842903137, 0.7679077982902527, 0.44033947587013245, 0.8384230732917786, 0.45649006962776184, 0.8333059549331665, 0.4636642634868622, 0.16825638711452484, 0.9304778575897217, 0.1564607173204422, 0.9345928430557251, 0.13891303539276123, 0.9397754669189453, 0.11874296516180038, 0.9452661871910095, 0.9010019302368164, 0.7714344263076782, 0.5068330764770508, 0.8109648823738098, 0.5675387382507324, 0.7181237936019897, 0.6213261485099792, 0.6761613488197327, 0.6662272214889526, 0.3659699261188507, 0.8406238555908203, 0.38610297441482544, 0.8361998200416565, 0.6061047911643982, 0.721760094165802, 0.5886551141738892, 0.26576125621795654, 0.8936172723770142, 0.7577317357063293, 0.471327543258667, 0.7948273420333862, 0.5243538618087769, 0.22516724467277527, 0.9051489233970642, 0.20626990497112274, 0.910885214805603, 0.8154552578926086, 0.4112367630004883, 0.8317334055900574, 0.6268150210380554, 0.33067744970321655, 0.859747052192688, 0.6731149554252625, 0.6428382396697998, 0.6720165610313416, 0.6439140439033508, 0.676583468914032, 0.36063018441200256, 0.8502982258796692], "advantages": [7.427423477172852, 6.072429656982422, 6.868880271911621, 5.495976448059082, 3.963945150375366, 4.469392776489258, 5.140400409698486, 5.844754695892334, 4.791353702545166, 3.366373300552368, 1.912610650062561, 2.1395699977874756, 2.5822439193725586, 1.2248811721801758, 1.5413694381713867, 0.23772397637367249, 0.3900194466114044, 0.6641584038734436, 1.2276068925857544, 0.1507691890001297, 0.62888503074646, -0.5168718099594116, -1.776077389717102, -1.7793984413146973, -1.5204557180404663, -2.7137339115142822, -2.5616979598999023, -3.7925326824188232, -5.086300373077393, -5.47756814956665, -6.67978572845459, -7.883882999420166, -8.724483489990234, -9.696170806884766, -10.332322120666504, -10.100647926330566, -11.655351638793945, -11.632172584533691, -10.544052124023438, -12.892107963562012, -12.00915241241455, -14.413633346557617, -16.12154197692871, -16.671293258666992, -16.10246467590332, -18.612600326538086, -18.214603424072266, -16.78709602355957, -20.068668365478516, -22.69741439819336, -22.63427734375, -21.470199584960938, -19.392126083374023, -23.42035675048828, -27.035579681396484, -26.09910011291504, -29.85240936279297, -28.99910545349121, -32.902523040771484, -32.13100051879883, -30.16973304748535, -27.392894744873047, -32.43961715698242, -37.49467849731445, -35.2943115234375, -40.5666389465332, -38.11614227294922, -43.57398986816406, -48.61722183227539, -47.002037048339844, -52.18333053588867, -50.229373931884766, -55.529361724853516, -53.21004104614258, -49.37760543823242, 21.45175552368164, 21.739347457885742, 21.289531707763672, 21.549474716186523, 21.172199249267578, 21.405364990234375, 21.09391975402832, 21.301456451416016, 21.051029205322266, 21.234195709228516, 22.120201110839844, 20.87684440612793, 20.704317092895508, 21.768112182617188, 20.989566802978516, 21.105453491210938, 21.041337966918945, 21.1368465423584, 21.872400283813477, 20.85971450805664, 20.84906578063965, 22.099565505981445, 21.24409294128418, 21.279481887817383, 21.388246536254883, 21.403446197509766, 21.566425323486328, 21.559871673583984, 21.78159523010254, 21.750450134277344, 22.0374813079834, 21.977205276489258, 22.470796585083008, 23.198339462280273, 21.98628044128418, 21.350448608398438, 21.765207290649414, 21.112199783325195, 21.33968162536621, 21.144468307495117, 21.377395629882812, 21.162199020385742, 21.403732299804688, 21.167354583740234, 21.484699249267578, 20.81749153137207, 21.015047073364258, 20.742582321166992, 20.935253143310547, 20.64582061767578, 20.91439437866211, 20.187877655029297, 20.30426597595215, 19.995586395263672, 20.08624839782715, 19.76849937438965, 19.830961227416992, 19.505708694458008, 19.536800384521484, 19.20586395263672, 19.201637268066406, 20.382747650146484, 23.039796829223633, 20.92194366455078, 23.806928634643555, 21.65060806274414, 20.373004913330078, 19.81646156311035, 20.247398376464844, 19.650461196899414, 20.168058395385742, 19.524856567382812, 20.13798713684082, 19.44279670715332, 19.147628784179688, 19.00004005432129, 18.697729110717773, 18.563947677612305, 18.253379821777344, 18.23143196105957, 17.51019859313965, 17.56808090209961, 16.742382049560547, 16.899930953979492, 15.932376861572266, 16.217252731323242, 15.069169998168945, 14.239531517028809, 14.072455406188965, 13.71423625946045, 13.422921180725098, 14.173169136047363, 13.0308837890625, 13.713277816772461, 12.56125259399414, 13.171510696411133, 15.170886993408203, 12.983121871948242, 15.015313148498535, 12.779891967773438, 11.530719757080078, 12.091137886047363, 10.830716133117676, 10.454005241394043, 9.675451278686523, 10.090161323547363, 11.965018272399902, 9.662449836730957, 8.382135391235352, 8.043291091918945, 7.081740856170654, 6.817946434020996, 5.75173807144165, 5.885012626647949, 4.7028350830078125, 4.740917205810547, 6.125810623168945, 9.087624549865723, 5.688118934631348, 3.326551675796509, 4.777929782867432, 2.37506365776062, 3.840435743331909, 1.3977327346801758, 0.07658004760742188]} +{"type": "SampleBatch", "eps_id": [700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 700489972, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAANnGYr7Lmyw+XNLDO5LgybwYU1++s0TavOqrszvZMoo+yd5fvgkrY75sSDI8s3cQP+NpZL6YK9y8x5m1PMvYjD7L9mS+4fMrPvSr4jyXUyy8ZYZhvoyruT7M8uA8P7SWviIZWr5NJSs+HrmwPMco4TverFa+RFC5PlLZsTyiz46+QkNPvheDKj49JoQ84h6oPDzaS77ECbk+BIOHPNC3iL5yc0S+q2gOPyeGNzxC8A2/6Q45vvDVuD5t3+s4LTqEvjKqMb5n3Sk+vJCluxIoDT19RC6+Ld+4Pvf6jrvqBYW+Zt8mvrMAKj79nxy8fv4APfx4I74s+bg+ME4SvAtEh77cEhy+YHgOPyPgaLz+Rg+/Eq0QvkAkuT6gItC80AKLvjhFCb7c2yo+Z578vB/rVTxs2gW+PNbjvMZ6+rwAepc+PWwGvvbBKz7QAcq88enOu9b8Ar4z9bk+qQrLvHIEnb5CGfe9MXssPoZJ/bz5fLO8DTPwvelcuj5AcAC9rvulvldK4b1oZC0+7P4ava80Kr3OWtq9hrHOvGBmHr1AsHQ+X2Pbve2CLj4n0wq9On+GvWBo1L1lPMa8ZzQQvYZQXT4eZtW9roUvPsP//LzoGLO9xWDOva2GvrzWqQW9+wRIPqRUz72iIF++4VLrvK6A9D53Qdi9UFu3vDgVnbyiKzQ+KizZvS0pMT5wQYC8v1X7vQkW0r1USbO8x1yUvKvwKD6G+9K9mqcxPvypcryZjwi+V+DLvTjWvD6HLo+88IHcvvbEvL1lIDI+gr7VvIEAE77ypLW95ya9PrVD7byRguO+HYOmvUDuMj6qCBu9yM0kvt1an71ror0+2TcovX4+7r4mL5C9CBU0PlNWTr0SUj6+HPuIvY2Xl7wXkF29RmC5PSa9ib36mjU+1yVWvf/6X76CeYK9rCqLvPYQaL3whWk9pCuDve9SWL5SZWO9jJ+pPszSi72RG3y8jkFIvZk7sTwmdIy93bo4Ptd7Rr36dpK+gxCFveVYZbwJ6129O0QUvEyjhb1jTVW+0ahevec4iD6DK469uiLOvifdSL1jhgo/K6mevYnVGL8wiRy9L2VRP0Edt71Jgs2+Dw+zvHl0Az8Ujse9o+dSvvPaPbxlJFs+vv3PvanqLrymde+77pGovbBt0L0OoVK+dLMSvF4NVT6I2ti97BsrvG8MnbswEbO9C0jZvbNtUr6dWda7FZ9QPtSy4b2cdCi8KaohuxViur2kHuK9DExSvpZ5jLu+t00+FYjqvePfJryoDI25Db6+veLy6r0IO1K+0bQLu8s/TD6lW/O9GVAmvB938zq6SsC9FcbzvQU6Ur7Pxaq3LylMPs4u/L1Wvia88v6BOyUbv72Fmfy9MFM9PhSvCTvs/MS+1Qb1vbOMwj58Tbe7RUAtv3V25b3YXz0+07ScvImNxb5D4929ploivFvs27wNRsu9K0vevX4KPj5mL+y8LfLMviax1r3RVBW8SuIWvWhE7725ENe95f0+PmJ0IL2ZgNe+92zPvVORA7xc70K9xDMQvivBz71ITE++n3hOvfe/Cz7nC9i9MVrZu4tKQ72/yy++dFHYvcfdTb7YWlG9O0HYPW7CSz3abzg9xGg+vT7p/jy/ck89aJB2PjLcO73nPY2+YSxjPfnVPT12dVK9QqgDO1f4Zj3kjRe+VEtSvU1cjj6C2Fo9O0mvvj6EO73m2Q0/y8w+Pf4jFr7LHw69dEt9PurJMj1enUg9o7jzvLWjZb0PzTY9pDMVviXo/LwViWg+ad0qPawpTD1+s9e8rfSZvbnyLj0iXhS+gQTkvN4aVj4nFCM9guCtvsTCwbxPWPw+JkIHPaSdE75WBWK8UnJFPvXl9jygm62+g9YivCRV9j7/V788zUQTvkLspblqxD0+38enPL8pUj1lKV47qBXcvfAvsDxOUBO+eZ2iOhvCPj74nZg8gJGtvjK9ojvabvU+gCZCPGxnE75Ycm48DMRAPir7Ejygta2+3BCWPKOR+D5rOg87NeITvp+b5TxNZks+WkA4uiwLrr5wEwM9jAEAP6rO9btNwhS+sQksPTXRXj6rgSq81p9JPf3cPT19Gny9Z2AavAAMFr440jg900d7Pj5kSryQPa++cexMPSVNDT/SRZ286r0Jv9kjej1TJ1w/cm31vHwIsL5rS6A9VkwWPwrhFr3Sdxm+oVe4PZPfoz4RKCO97qczPcJzxT2Nwmk9OpAfvXktcz4vysc9BZhQvvMbDL1zttw+LnK/PVGd7b4el9G8lfAfP9ZvrD0P1Tm/LHVWvJZv2z4ltI49qN/QvpgJlLsyIW4+1Px7PQ3PwL1Tsgs5TX3aPndGdD1jzbu+GQQOPEZcbD4VOlY9txJlvb+mWTz0rdk+O6VRPUrPqb6ue7I8CtlqPtN5Nj0qwb68Cw/YPLisCT1+kTQ9T/SQPtWQ3TxejWk+0sJLPR6DmDuRdwE9jzcEPW4kTD06DqA+hBwEPd4YaD5OwGU9yakTPd6tFj1lhNc+V7RoPY8kdL5zKTk9oX0dP0wsVT1tPQO/F49rPWm51j4fLSs9J+9Qvhn1hj14Jx0/JnYaPYVs974iGqA94iTWPlS/5TzqNze+zjuxPYQVZD60bsg809L6PWJbuj29wtU+kn/cPCBKJr40dcs9L1VjPlnkwTyr/A0+F43UPSdk1T4jnNg8mfoVvlif5T0ijBw/AJ3APEWD3L6Gq/49UWlOPzAZdDwiQze/EdkPPo4ngD8YXRg6U32Av0daJD4WW04/i7SfvGsBNr9x3DQ+nnkcPyoYCr3rVNm+DWFBPuJM1T4T3iy9KQgSvj/pST6fcGM+zow4vcG9Cz69dU4+APDVPuheLb2FKi6+dgRXPn0RHT/UTTu92Z3zvjiVYz5dlNY+XUhiveCiSr6EKmw+wzRmPlp+cr3obZ09K8VwPr+N+jxHMmy9z1qzPoZlcT7L6Wc+6H9PvfPeIz3rCHY+o40DPeM4TL1P8KE+T7F2PoUnJr7hTzK9PsQXP5lecz7e+wg9Kb8BvQzTkj7wDXQ+mX9qPneC1Lxs6YC8kb54Pk1W2T5+Fte88gqivg24gD4fQ2s+hXgFvYDlA71SEoM+yc0PPdQbCL1IA4A+W26DPqk5bD7ZQOe8ZONYvRjLhT6MYhM9yO3vvNc6bD5rKYY+HhFtPs8hyrysk5G9T4iIPiaIFj05x9W8AthaPqboiD5Lz20+YcOyvGdZsr1xSYs++VMZPfkHwbzXZks+k6uLPkd5bj6jfKC8PaTPvREOjj762Bs9IxmxvPd9PT7PcY4+mhNvPovHkrxAPOq92NWQPikoHj2vhKW8GL8wPhE7kT5Yom8+Jj2JvMZqAb6HoJM+/VAgPRLynbyK1CQ+IQeUPj4pcD6gkoO8iQoNvvFtlj7PYSI9rCOavPxuGT7e1ZY+yqtwPgmXgbx6Shi+/D2ZPjFoJD3f9Jm8akMOPjWnmT5SLXE+xDGDvER0I76fEJw+QJvcPtlYnbw1Jeq+IHqgPv+wcT4IRui8hNguvtvkoj4u89w+2x8CvUzJ8b4eUKc+n5FyPm3PKL0fSkK+GL2pPiyRLT16Wji9+qK3PS0sqj5RuRu+CQIxvW1SvD6Gnag+/LkyPV7gEr26PX096Q+pPmL+dD7Gzw29/MF3vhiDqz5j2TY92qEhvalBIj0e+Ks+vRJ2PhljHr1hz4e+EG6uPjdrOz3hHTS9yYd1PAPmrj7vRhi+muMyvZhClj4vYK0+bLWvvvLYGr1PcxI/jtypPlQXF75099e8AROJPsNZqD4ZQK++VRqsvEhODT962KQ+0n8Jv5dVI7z0hFY/fFifPvUGr745gN47UcsKP1nYmz4OWRa+GXSQPIjVgD51V5o+W8dEPTCuuTyoKRG9ZdWaPnbvFr6537M8RFOHPgBTmT6gKkI9kS3fPIYQr7xEz5k++6YXvjyt2zyhP48+CUuYPp4APz0UwgQ9Ga6Nu0fFmD6EEHc+Z2cEPR++kr7DPZs+fDY7PabZ2Txp1oM8lLWbPrRhGb6ofNw89FOiPuwsmj7FFDg9SDcIPYcNBz28opo+okEavsDqCj3VAqw+1xeZPsYmND1TcCY9LNhdPSOLmT4EOXQ+K+AqPfTWZr5Y/Js+h1UvPZJoGD2fE6Q9j2ycPpEScz63+B499GlNvtPanj7RJN0+14kOPT4q9r4VR6M+df9xPthNzjxTmjW+mLKlPmZVJz1kP7E8VUL8PbAdpj5qUHE+qG3FPBaCJr50h6g+8mDcPnzJqjxqKuW+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 0, 1, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 1, 0, 0, 0, 0, 1, 1, 0, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 0, 1, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 1, 0], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [1, 0, 0, 1, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 1, 1, 0, 0, 0, 0, 1, 1, 0, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 0, 1, 1, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 1], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAABhTX76zRNq86quzO9kyij7J3l++CStjvmxIMjyzdxA/42lkvpgr3LzHmbU8y9iMPsv2ZL7h8ys+9KviPJdTLLxlhmG+jKu5Pszy4Dw/tJa+Ihlavk0lKz4eubA8xyjhO96sVr5EULk+UtmxPKLPjr5CQ0++F4MqPj0mhDziHqg8PNpLvsQJuT4Eg4c80LeIvnJzRL6raA4/J4Y3PELwDb/pDjm+8NW4Pm3f6zgtOoS+MqoxvmfdKT68kKW7EigNPX1ELr4t37g+9/qOu+oFhb5m3ya+swAqPv2fHLx+/gA9/Hgjviz5uD4wThK8C0SHvtwSHL5geA4/I+BovP5GD78SrRC+QCS5PqAi0LzQAou+OEUJvtzbKj5nnvy8H+tVPGzaBb481uO8xnr6vAB6lz49bAa+9sErPtAByrzx6c671vwCvjP1uT6pCsu8cgSdvkIZ970xeyw+hkn9vPl8s7wNM/C96Vy6PkBwAL2u+6W+V0rhvWhkLT7s/hq9rzQqvc5a2r2Gsc68YGYevUCwdD5fY9u97YIuPifTCr06f4a9YGjUvWU8xrxnNBC9hlBdPh5m1b2uhS8+w//8vOgYs73FYM69rYa+vNapBb37BEg+pFTPvaIgX77hUuu8roD0PndB2L1QW7e8OBWdvKIrND4qLNm9LSkxPnBBgLy/Vfu9CRbSvVRJs7zHXJS8q/AoPob70r2apzE+/KlyvJmPCL5X4Mu9ONa8Pocuj7zwgdy+9sS8vWUgMj6CvtW8gQATvvKktb3nJr0+tUPtvJGC474dg6a9QO4yPqoIG73IzSS+3VqfvWuivT7ZNyi9fj7uviYvkL0IFTQ+U1ZOvRJSPr4c+4i9jZeXvBeQXb1GYLk9Jr2JvfqaNT7XJVa9//pfvoJ5gr2sKou89hBovfCFaT2kK4O971JYvlJlY72Mn6k+zNKLvZEbfLyOQUi9mTuxPCZ0jL3dujg+13tGvfp2kr6DEIW95VhlvAnrXb07RBS8TKOFvWNNVb7RqF695ziIPoMrjr26Is6+J91IvWOGCj8rqZ69idUYvzCJHL0vZVE/QR23vUmCzb4PD7O8eXQDPxSOx72j51K+89o9vGUkWz6+/c+9qeouvKZ177vukai9sG3QvQ6hUr50sxK8Xg1VPoja2L3sGyu8bwyduzARs70LSNm9s21Svp1Z1rsVn1A+1LLhvZx0KLwpqiG7FWK6vaQe4r0MTFK+lnmMu763TT4ViOq9498mvKgMjbkNvr694vLqvQg7Ur7RtAu7yz9MPqVb870ZUCa8H3fzOrpKwL0VxvO9BTpSvs/FqrcvKUw+zi78vVa+Jrzy/oE7JRu/vYWZ/L0wUz0+FK8JO+z8xL7VBvW9s4zCPnxNt7tFQC2/dXblvdhfPT7TtJy8iY3FvkPj3b2mWiK8W+zbvA1Gy70rS969fgo+PmYv7Lwt8sy+JrHWvdFUFbxK4ha9aETvvbkQ173l/T4+YnQgvZmA1773bM+9U5EDvFzvQr3EMxC+K8HPvUhMT76feE69978LPucL2L0xWtm7i0pDvb/LL750Udi9x91NvthaUb07Qdg9ho3gvWs7q7tltEi9OKJPvr9yTz1okHY+Mtw7vec9jb5hLGM9+dU9PXZ1Ur1CqAM7V/hmPeSNF75US1K9TVyOPoLYWj07Sa++PoQ7vebZDT/LzD49/iMWvssfDr10S30+6skyPV6dSD2juPO8taNlvQ/NNj2kMxW+Jej8vBWJaD5p3So9rClMPX6z17yt9Jm9ufIuPSJeFL6BBOS83hpWPicUIz2C4K2+xMLBvE9Y/D4mQgc9pJ0TvlYFYrxSckU+9eX2PKCbrb6D1iK8JFX2Pv9XvzzNRBO+QuyluWrEPT7fx6c8vylSPWUpXjuoFdy98C+wPE5QE755naI6G8I+PvidmDyAka2+Mr2iO9pu9T6AJkI8bGcTvlhybjwMxEA+KvsSPKC1rb7cEJY8o5H4Pms6Dzs14hO+n5vlPE1mSz5aQDi6LAuuvnATAz2MAQA/qs71u03CFL6xCSw9NdFePquBKrzWn0k9/dw9PX0afL1nYBq8AAwWvjjSOD3TR3s+PmRKvJA9r75x7Ew9JU0NP9JFnbzqvQm/2SN6PVMnXD9ybfW8fAiwvmtLoD1WTBY/CuEWvdJ3Gb6hV7g9k9+jPhEoI73upzM9wnPFPY3CaT06kB+9eS1zPi/Kxz0FmFC+8xsMvXO23D4ucr89UZ3tvh6X0byV8B8/1m+sPQ/VOb8sdVa8lm/bPiW0jj2o39C+mAmUuzIhbj7U/Hs9Dc/AvVOyCzlNfdo+d0Z0PWPNu74ZBA48RlxsPhU6Vj23EmW9v6ZZPPSt2T47pVE9Ss+pvq57sjwK2Wo+03k2PSrBvrwLD9g8uKwJPX6RND1P9JA+1ZDdPF6NaT7Swks9HoOYO5F3AT2PNwQ9biRMPToOoD6EHAQ93hhoPk7AZT3JqRM93q0WPWWE1z5XtGg9jyR0vnMpOT2hfR0/TCxVPW09A78Xj2s9abnWPh8tKz0n71C+GfWGPXgnHT8mdho9hWz3viIaoD3iJNY+VL/lPOo3N77OO7E9hBVkPrRuyDzT0vo9Ylu6Pb3C1T6Sf9w8IEomvjR1yz0vVWM+WeTBPKv8DT4XjdQ9J2TVPiOc2DyZ+hW+WJ/lPSKMHD8AncA8RYPcvoar/j1RaU4/MBl0PCJDN78R2Q8+jieAPxhdGDpTfYC/R1okPhZbTj+LtJ+8awE2v3HcND6eeRw/KhgKvetU2b4NYUE+4kzVPhPeLL0pCBK+P+lJPp9wYz7OjDi9wb0LPr11Tj4A8NU+6F4tvYUqLr52BFc+fREdP9RNO73ZnfO+OJVjPl2U1j5dSGK94KJKvoQqbD7DNGY+Wn5yvehtnT0rxXA+v436PEcybL3PWrM+hmVxPsvpZz7of0+9894jPesIdj6jjQM94zhMvU/woT5PsXY+hScmvuFPMr0+xBc/mV5zPt77CD0pvwG9DNOSPvANdD6Zf2o+d4LUvGzpgLyRvng+TVbZPn4W17zyCqK+DbiAPh9Daz6FeAW9gOUDvVISgz7JzQ891BsIvUgDgD5bboM+qTlsPtlA57xk41i9GMuFPoxiEz3I7e+81zpsPmsphj4eEW0+zyHKvKyTkb1PiIg+JogWPTnH1bwC2Fo+puiIPkvPbT5hw7K8Z1myvXFJiz75Uxk9+QfBvNdmSz6Tq4s+R3luPqN8oLw9pM+9EQ6OPvrYGz0jGbG89309Ps9xjj6aE28+i8eSvEA86r3Y1ZA+KSgePa+EpbwYvzA+ETuRPliibz4mPYm8xmoBvoegkz79UCA9EvKdvIrUJD4hB5Q+PilwPqCSg7yJCg2+8W2WPs9hIj2sI5q8/G4ZPt7Vlj7Kq3A+CZeBvHpKGL78PZk+MWgkPd/0mbxqQw4+NaeZPlItcT7EMYO8RHQjvp8QnD5Am9w+2VidvDUl6r4geqA+/7BxPghG6LyE2C6+2+SiPi7z3D7bHwK9TMnxvh5Qpz6fkXI+bc8ovR9KQr4Yvak+LJEtPXpaOL36orc9LSyqPlG5G74JAjG9bVK8PoadqD78uTI9XuASvbo9fT3pD6k+Yv50PsbPDb38wXe+GIOrPmPZNj3aoSG9qUEiPR74qz69EnY+GWMevWHPh74Qbq4+N2s7PeEdNL3Jh3U8A+auPu9GGL6a4zK9mEKWPi9grT5sta++8tgavU9zEj+O3Kk+VBcXvnT317wBE4k+w1moPhlAr75VGqy8SE4NP3rYpD7Sfwm/l1UjvPSEVj98WJ8+9QavvjmA3jtRywo/WdibPg5ZFr4ZdJA8iNWAPnVXmj5bx0Q9MK65PKgpEb1l1Zo+du8WvrnfszxEU4c+AFOZPqAqQj2RLd88hhCvvETPmT77phe+PK3bPKE/jz4JS5g+ngA/PRTCBD0Zro27R8WYPoQQdz5nZwQ9H76SvsM9mz58Njs9ptnZPGnWgzyUtZs+tGEZvqh83Dz0U6I+7CyaPsUUOD1INwg9hw0HPbyimj6iQRq+wOoKPdUCrD7XF5k+xiY0PVNwJj0s2F09I4uZPgQ5dD4r4Co99NZmvlj8mz6HVS89kmgYPZ8TpD2PbJw+kRJzPrf4Hj30aU2+09qePtEk3T7XiQ49Pir2vhVHoz51/3E+2E3OPFOaNb6YsqU+ZlUnPWQ/sTxVQvw9sB2mPmpQcT6obcU8FoImvnSHqD7yYNw+fMmqPGoq5b7L76w+LqhwPnboQjxR+Be+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.3526957333087921, 0.14588908851146698, 0.9279260039329529, 0.8642497658729553, 0.6950178146362305, 0.6099445819854736, 0.7067097425460815, 0.5995811223983765, 0.7148560285568237, 0.4066804349422455, 0.8370143175125122, 0.6136423349380493, 0.7033019661903381, 0.626803994178772, 0.6934657096862793, 0.3556269109249115, 0.8688167333602905, 0.6879609227180481, 0.36533549427986145, 0.8508206009864807, 0.6159877181053162, 0.7379282116889954, 0.5787248015403748, 0.7705727815628052, 0.4693235456943512, 0.8079946041107178, 0.5041186809539795, 0.7921157479286194, 0.5371452569961548, 0.22471413016319275, 0.9074897170066833, 0.7715845704078674, 0.5683070421218872, 0.7613142728805542, 0.41411346197128296, 0.8545545935630798, 0.37226608395576477, 0.87298184633255, 0.3204433023929596, 0.892001211643219, 0.7374461889266968, 0.59306401014328, 0.7798377871513367, 0.47273099422454834, 0.8156664967536926, 0.4795115292072296, 0.8371322154998779, 0.5875725746154785, 0.25028878450393677, 0.09666867554187775, 0.9492496252059937, 0.906002938747406, 0.750521719455719, 0.613213837146759, 0.7475426197052002, 0.6159107089042664, 0.7462193369865417, 0.6163398027420044, 0.7465018630027771, 0.6145813465118408, 0.7483493685722351, 0.6106396317481995, 0.7517302632331848, 0.395553857088089, 0.1433952897787094, 0.9358101487159729, 0.8732455372810364, 0.325672447681427, 0.8862518668174744, 0.2834150493144989, 0.8998479843139648, 0.7622450590133667, 0.5651187300682068, 0.7938673496246338, 0.5091937780380249, 0.5167255997657776, 0.822179913520813, 0.548430323600769, 0.22651681303977966, 0.9097132682800293, 0.7622870802879333, 0.6056082844734192, 0.7432963848114014, 0.6309908628463745, 0.2753540277481079, 0.8985282182693481, 0.2735895812511444, 0.9011515378952026, 0.7388026118278503, 0.6241142153739929, 0.2573023736476898, 0.9077515006065369, 0.23462699353694916, 0.9152626991271973, 0.2052493542432785, 0.9238213300704956, 0.8274386525154114, 0.4461662769317627, 0.14936786890029907, 0.06195663660764694, 0.9619399309158325, 0.9466362595558167, 0.9081125855445862, 0.799969494342804, 0.538242757320404, 0.24238047003746033, 0.8970289826393127, 0.7285022139549255, 0.6437066197395325, 0.6920600533485413, 0.6815974116325378, 0.6583442687988281, 0.28890737891197205, 0.8932166695594788, 0.25030797719955444, 0.9042867422103882, 0.7867713570594788, 0.49777668714523315, 0.7946602702140808, 0.5209549069404602, 0.7871285676956177, 0.4689597189426422, 0.8232669830322266, 0.44756776094436646, 0.8328958749771118, 0.5736610293388367, 0.23976729810237885, 0.09205390512943268, 0.952098548412323, 0.9221673607826233, 0.8314427733421326, 0.5761308073997498, 0.7505484819412231, 0.3738429546356201, 0.8780505657196045, 0.6957733035087585, 0.3589388132095337, 0.8582263588905334, 0.4041714668273926, 0.15794162452220917, 0.9275492429733276, 0.8371842503547668, 0.5470552444458008, 0.8010272979736328, 0.4985509216785431, 0.8025622963905334, 0.5279192924499512, 0.7874886393547058, 0.5539290308952332, 0.7726758122444153, 0.5771138072013855, 0.7580527067184448, 0.5980157256126404, 0.7434846758842468, 0.6171480417251587, 0.7287819385528564, 0.6349784731864929, 0.7137050628662109, 0.6519235372543335, 0.6979685425758362, 0.6683480143547058, 0.6812408566474915, 0.3154350519180298, 0.8865808248519897, 0.2822732925415039, 0.8968557715415955, 0.7556546330451965, 0.44990891218185425, 0.8108873963356018, 0.5221246480941772, 0.797235369682312, 0.47841259837150574, 0.8181584477424622, 0.569139301776886, 0.26748043298721313, 0.8876324892044067, 0.2711043357849121, 0.11027148365974426, 0.9452783465385437, 0.9027324914932251, 0.7765218615531921, 0.5241576433181763, 0.794934868812561, 0.49170902371406555, 0.8148983120918274, 0.5471991300582886, 0.770397424697876, 0.4310435950756073, 0.8436997532844543, 0.3905666470527649, 0.8622559309005737, 0.6562315225601196, 0.6873596906661987, 0.6870216727256775, 0.3421062231063843, 0.8696892857551575, 0.6479174494743347, 0.7160850167274475, 0.37141895294189453, 0.860981822013855], "advantages": [-5.805177688598633, -7.171714782714844, -7.50447940826416, -9.018877983093262, -9.317327499389648, -8.29710578918457, -10.692469596862793, -9.70886516571045, -12.131909370422363, -11.181135177612305, -8.599435806274414, -12.267099380493164, -14.794938087463379, -13.751933097839355, -16.2932186126709, -15.230537414550781, -12.382567405700684, -16.213668823242188, -18.843955993652344, -20.248062133789062, -20.56816291809082, -19.29852294921875, -21.902080535888672, -20.527822494506836, -23.156572341918945, -24.566425323486328, -24.704952239990234, -26.109140396118164, -26.212936401367188, -27.61455726623535, -28.09249496459961, -29.402257919311523, -29.52312469482422, -30.908702850341797, -31.026498794555664, -29.623170852661133, -32.17559051513672, -30.59762191772461, -33.21186828613281, -31.408525466918945, -34.087188720703125, -35.544185638427734, -35.202537536621094, -36.680118560791016, -37.20357894897461, -38.09786605834961, -37.558563232421875, -39.078208923339844, -39.642520904541016, -39.63841247558594, -39.3078498840332, -41.7000732421875, -43.37665557861328, -44.12235641479492, -44.83919906616211, -45.58688735961914, -46.33348846435547, -47.08747100830078, -47.86366653442383, -48.628150939941406, -49.43378448486328, -50.21290969848633, -51.0478401184082, -51.8458251953125, -51.456478118896484, -49.496952056884766, -52.27008056640625, -54.05083465576172, -53.340553283691406, -55.1651611328125, -54.28925323486328, -56.1627082824707, -57.1577033996582, -57.4947624206543, -58.560482025146484, 16.70970916748047, 16.96590232849121, 16.588821411132812, 17.302526473999023, 19.09515380859375, 16.693416595458984, 15.744462013244629, 16.35842514038086, 15.438992500305176, 16.029354095458984, 17.680248260498047, 15.462859153747559, 17.01536750793457, 14.911099433898926, 14.134074211120605, 14.561488151550293, 16.000106811523438, 14.017610549926758, 15.378988265991211, 13.485238075256348, 14.789929389953613, 12.968857765197754, 12.260245323181152, 12.589587211608887, 13.902763366699219, 16.11175537109375, 13.3578519821167, 11.639630317687988, 10.878172874450684, 10.739867210388184, 11.25317096710205, 12.947202682495117, 11.02008056640625, 10.254191398620605, 10.72831916809082, 10.028216361999512, 10.469204902648926, 9.838679313659668, 10.081477165222168, 9.609055519104004, 9.982455253601074, 9.4464111328125, 9.749494552612305, 11.156461715698242, 9.840692520141602, 11.266512870788574, 10.023856163024902, 9.963217735290527, 10.126771926879883, 10.223444938659668, 10.309179306030273, 11.596709251403809, 14.647881507873535, 20.116416931152344, 15.7312650680542, 13.200026512145996, 12.58830738067627, 13.588249206542969, 12.909220695495605, 13.86764144897461, 13.633208274841309, 14.93297004699707, 16.823040008544922, 15.00060749053955, 16.80690574645996, 18.335186004638672, 16.48262596130371, 14.720860481262207, 13.699496269226074, 15.019783973693848, 16.53319549560547, 14.939926147460938, 16.346616744995117, 14.816868782043457, 16.119924545288086, 14.651726722717285, 15.855266571044922, 14.445880889892578, 15.554400444030762, 14.200718879699707, 15.218750953674316, 13.917484283447266, 14.849333763122559, 13.597264289855957, 14.446856498718262, 13.24090576171875, 14.011672019958496, 12.848993301391602, 12.13128662109375, 12.574131965637207, 11.975226402282715, 12.226496696472168, 12.608736991882324, 12.938835144042969, 11.88167667388916, 10.950910568237305, 11.229426383972168, 10.345603942871094, 10.540698051452637, 10.757345199584961, 10.82253360748291, 9.876598358154297, 9.960009574890137, 9.961868286132812, 9.089656829833984, 8.151199340820312, 7.044088840484619, 7.431649208068848, 6.314099311828613, 6.723358631134033, 5.597305774688721, 4.487648963928223, 4.9612650871276855, 5.372627258300781, 4.250913619995117, 4.6903300285339355, 3.5701632499694824, 2.4009199142456055, 2.967733144760132, 1.801228642463684, 0.9750916361808777, 1.269266963005066, 1.7423484325408936, 0.6158199906349182, -0.165374755859375]} +{"type": "SampleBatch", "eps_id": [1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1730362651, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAMvvrD4uqHA+duhCPFH4F77gV68+hiTcPhRHEjyv59++Ab+zPitacD7Xnj45CDwRvk4mtj7BLSI9g/wtuxSIGj4ajrY+bWNwPhqDvjk+CBK+f/W4Pg5QIj1xGyO74MoZPmBduT5ia3A+Hv0GOsC3Er7ZxLs+mm0iPTQNGrvlJxk+zSy8PkFycD6e8yc6VU8TvliUvj7yHtw+mpESu7Rn375d+8I+KnhwPhefM7xd0xO+92LFPnU33D737GK854bhvnrKyT7oGyA/p6G5vLvMPL/+MdA+cSNSP007Gb3Vw4S/z5nYPspRID+DM269NatBv3sD3z6qK90+bhaWvdoi977gb+M+ZqtzPsfbqb2fG1u+q9/lPh+xND1xn7K9BLFSPU9T5j7yI3Y+E4SwvantiL5uyeg+L8k+PV14u73jiE27iEPpPn2mFr5Ambu9kMWEPt7B5z6/c0k9FvqwvfgAeb3MQug+PhIUvol3s719TlA+vMfmPj2XUz16Iqu9Eq3svSdP5z7ulRG+Qt6vvXw4GT502uU+7wOsvke9qb07/NM+vGniPhMhD77Sx5i9hHXFPVP74D6b4aq+1NSUvdu0uj5pkN0+NfsMvhXlhb1EtUw9gCfcPijiqb4H2YO9YIakPrLB2D4oFgu+Hl9tvTmOFDyjXdc+Kf+ovvegbL2E2pA+X/zTPhs6Br/Dc1W9tYwOP+Odzj6RMqi+GdYnvesnfj63QMs+7OQFv/yAE70bFgc/pOXFPuaip75rjdC8oD5lPleLwj7WGge+j9+rvIQlnb15McE+NkynvutxuLzkR1Y+6di9Puh3Br76KJa8CTm5vauAvD7r/6a+V/qkvOMdST6iKbk+0ecFvpnMhLwlDdK91dK3Pv+7pr5zmpW8TWY9Pid9tD4UQAW/SJluvE+T8T6sKK8+6iU3vyH7p7vLh0I/PNWnPswtBb8sAiU8M2LuPnuBoj5ph6a+eMmePOtTND7aLJ8+CkkFv7KjuzyjHfM+A9iZPmvWpr7ftwQ9iAFCPs6Blj4YfAW/HT0UPaP/+z7rKpE+dlanvvmOPD3WMFg+J9KNPhKTB76Q2k09jpSIvRR3jD4C1H09+2NIPRarr76IGY0+OwYJvptILD2SARG9v7qLPsm1qL4tYik9Jl6KPvRaiD5rPgq+uIU/PVw9FbwM+YY+5lmpvrHGPj08jZg++JWDPqGbC741L1c9xmumPJMwgj4EjG09P9lYPQC8gr6ayII+TyYNvl7uQz2QYVs9Q1+BPsGuZz2ZUUg97gVlvorzgT7xfIA+NP81PQgiAL9lhYQ+TA7kPo3+DD1ZZEe/CxWJPgTXfz65YJo8FX3zvv6jiz44El89RuwYPNA/Nb7CMow+LiwQvqrYvTuS6vI9rcGKPu4vXj0vygU8KF8wvuBPiz50YBC+oLOaO47s+z1F3ok+IG9dPT5R6ztdNyy+/WuKPgWNEL6oMno7Us0BPvD5iD6Nylw9GCzQO2CrKL4/h4k+suh+PpZySDumKOm+0BOMPmI9XD0tOMa76J8lvsSgjD7dpxC+DxwYvE0fBD5yLos+5CJdPTSp27vKkSq++buLPgFsEL6naSS84u39Pcnk+zpMZi09spfJut1RKztJbzU7r24cvmK9wrqF8ZY+v2OWuaOTLT3NhZA7/ZzZOr77EjqmeRy+WJyROx5qlz7biiO7HSyyvu21KTz1/RY/d+oavJqwHL5/fbU8Zs+ZPoAOTbx0Iys9n7XmPH1rcjyVXT+8ISpyPjci6TxMqYm+7L7ju9fSJz0AFb08ur0FPd7kyLsULh6+g27CPO9Aqj6KEBe8XBUlPbHp+DxrPUI9odsJvGT3Hr5aVwA9MPayPii6PLz/eCE9oPkcPen/iD0yzy+87ZZvPoB0Ij33k1q+C0jGu09l2z4G+BA9spr8vhsXJTuFfm4+tBrRPEBMQr5xLus7OfXaPkMEsjwU0/K+pNyAPL/NbT4ooEg8mgIzvhXppjymtto+pVcPPDte7b4X5uw8IX9tPucrCbreNyy++nIJPSzMFj1zu3671qP+PQ93DD1aDCK+d4a3um/i1D6fAP88kRsXPWGc4jtUOPs9+4UCPfMgIr4pgBk8isXVPiwb6zzfJxY9QSiRPB/gAj7EHPE8UBRtPuoYpjzODSO+xYULPWBK2j45Aow8FxLkvvNyLj1tiGw+Wg0GPMT3Fr4kX0E9wRvaPiF8qzuhAuC+3URkPSVWbD4bf2a77qASvgktdz1jVRI9MBfRu7vwFz5DGno9mSAjvo6yX7soyuA+ag1tPUroEj3N4a87r8UUPpX9bz0FKCO+UowHPLIa4T4k8GI9UxwSPbTOizz5Lhk+OtxlPf4TbD4ZUaQ8/fUMvhu/eD2h8A89V8ONPBgoJT4UoHs9KIhrPicwqDzK6AC+4zuHPcaD2T4HkJM8xPTSvpaimD1epB4/AR0gPPLvMr+RBLI9IFDZPlTaibs8dM6+I2fDPXifHj+YDkm8HX4yv1XI3D2Abtk+iMPWvFQf0b5VLe496XJrPmvXDL02PP69VJj3Pcje2T7JAhe9Fd3aviiDBD5XfWw+cAc6vZwmFr77PQk++IQVPYcKRr2UiQY+Xv0JPszbbT40Rzu9LWE0vjS/Dj5SAxs9YLVJvYtp0D2ehQ8+eksgvjxfQb1gzME+6VAMPtOmID1BXSK9OROSPYseDT6uhnA+coUcvagzb74J7hE+RTQlPUyoL70Zqj89gMESPh7ZHb760iu9q7qmPlGZDz7dKio9wCURvXsZpDwhcxA+KtlyPqeBD71cNpG+g04VPsCS3T6NvSa97E4Wv2wrHj7r8nM+ztZWvdp3nb5xDCM+ssozPbQIcL33gwK9k/IjPqGPdT7xpHK9ClWvvtnbKD45rDo9RlmHvTtYjb3Kyik+XWB3PvUsir3IesO+W70uPmx2Qj1i0Jm96ZXjvUW2Lz5wAha+n12evTW3ID44tiw+7FhLPeTvl70B6SK+gbotPvfNE74XdJ69C9bfPb/FKj4oMq2+DfqZvbjRwD442CM+n5QRvh2Nir33ono92e4gPvyWXD18C4i9PAyBvjQJIj4o238+Yl6SvZ6eEL8vJyc+I5hkPQCCqb2FVpe+yUsoPrNaDb5nnbW9IKT2vA14JT7936m+G9m2vXbubj6IrB4+9ogGv3JKrb2FPv4+QekTPrWcqL6G85i96782PqsqDT4vaQi+K6SRvf4XDL4+cAo+Ko2nvro+l726zAc+hLwDPvRvBb8j0JG9SF/NPnAf8j17fqa+HWKBvfzlsT2kzeQ92vMEv42me71Tyrc+6IfPPZSVpb6APl69M85CPb1Iwj2siAS/GFlavVsxpT4nFK09WMykvszqP73ZaV88FeWfPaorBL/UzD698RSVPmC/ij1yHaS+cPImvZGogbyaPHs9bNoDv10+KL2YBoc+MAtRPTmEo761oxK9MYIqvY3hNj3QkgO/tgwWvQ1MdT4Nxww9lvyivgdtAr1TBIS9PGblPPZSA7/htAe9XD1fPhpakTzsgqK+3bHrvNb0rb1Vsjo8MBkDv32c+bwlS0s+FiCXOgEUor6WFdm8hTDUvb6tqbv74wK/Ow/qvKbuOD4WYXy86ayhvmh4zLy8ufe9/+yxvM1g9r3TSeC8QcPXvtGixbwCS6G+jqoSvaHIDL7xP/m8PXYCv87tHb2lIxM+aF8mvTZENL9kKBK94xHZPtUOYL3LMAK/a9revGxF9j0G3IS9FQo0v8kmy7zk9s4+cKqhvRwAAr9D7Ii8BKXUPUF3tr1T+p++mtJvvIy8Rr6aQ8O9weIBv4y1l7xqacA9uQvYvY7EM7/0UIi8lu/CPgPP9L3SwQG/nd8TvPytqT3uyAS+z4ufvhJz8bv1xFm+ryoLvlFc7r0zaT68PfYBv+OMDb5kaJ++jWGyvN3pX7457RO+35IBvw011rzqW4k95EoevuMOn77wN8u8X1tvvqanJL4pYwG//4PxvMTuUD2AAS++LDozv4co6bxTIqs+E1g9vmUSZb85ZbK8bikeP3erT75mCzO/+VcavN//oj5M/l2+CfZkv4oYSLuJqRs/a09wvl9zi79FORU8rVdmP6lPg75l+2S/8wfePPAmHD9veIy+YyEzv/T7ID034qY+u6KTvhY4Zb+Crzs9YHkhP+/NnL5fcDO/eFtvPd6otD5k+6O+XrYBv6Uhhj2TGqI9pSupvjbjM7+eX4k9kJTIPrFdsL4UMgK/gGuZPXm39z3lkrW+sxOhvs9fnj2btxa+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 1, 0], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 0, 0, 0, 1, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 1], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAOBXrz6GJNw+FEcSPK/n374Bv7M+K1pwPteePjkIPBG+Tia2PsEtIj2D/C27FIgaPhqOtj5tY3A+GoO+OT4IEr5/9bg+DlAiPXEbI7vgyhk+YF25PmJrcD4e/QY6wLcSvtnEuz6abSI9NA0au+UnGT7NLLw+QXJwPp7zJzpVTxO+WJS+PvIe3D6akRK7tGffvl37wj4qeHA+F58zvF3TE773YsU+dTfcPvfsYrznhuG+esrJPugbID+nobm8u8w8v/4x0D5xI1I/TTsZvdXDhL/Pmdg+ylEgP4Mzbr01q0G/ewPfPqor3T5uFpa92iL3vuBv4z5mq3M+x9upvZ8bW76r3+U+H7E0PXGfsr0EsVI9T1PmPvIjdj4ThLC9qe2Ivm7J6D4vyT49XXi7veOITbuIQ+k+faYWvkCZu72QxYQ+3sHnPr9zST0W+rC9+AB5vcxC6D4+EhS+iXezvX1OUD68x+Y+PZdTPXoiq70Srey9J0/nPu6VEb5C3q+9fDgZPnTa5T7vA6y+R72pvTv80z68aeI+EyEPvtLHmL2EdcU9U/vgPpvhqr7U1JS927S6PmmQ3T41+wy+FeWFvUS1TD2AJ9w+KOKpvgfZg71ghqQ+ssHYPigWC74eX229OY4UPKNd1z4p/6i+96BsvYTakD5f/NM+GzoGv8NzVb21jA4/453OPpEyqL4Z1ie96yd+PrdAyz7s5AW//IATvRsWBz+k5cU+5qKnvmuN0LygPmU+V4vCPtYaB76P36u8hCWdvXkxwT42TKe+63G4vORHVj7p2L0+6HcGvvoolrwJObm9q4C8Puv/pr5X+qS84x1JPqIpuT7R5wW+mcyEvCUN0r3V0rc+/7umvnOalbxNZj0+J320PhRABb9ImW68T5PxPqworz7qJTe/Ifunu8uHQj881ac+zC0FvywCJTwzYu4+e4GiPmmHpr54yZ4861M0Ptosnz4KSQW/sqO7PKMd8z4D2Jk+a9amvt+3BD2IAUI+zoGWPhh8Bb8dPRQ9o//7PusqkT52Vqe++Y48PdYwWD4n0o0+EpMHvpDaTT2OlIi9FHeMPgLUfT37Y0g9FquvvogZjT47Bgm+m0gsPZIBEb2/uos+ybWovi1iKT0mXoo+9FqIPms+Cr64hT89XD0VvAz5hj7mWam+scY+PTyNmD74lYM+oZsLvjUvVz3Ga6Y8kzCCPgSMbT0/2Vg9ALyCvprIgj5PJg2+Xu5DPZBhWz1DX4E+wa5nPZlRSD3uBWW+ivOBPvF8gD40/zU9CCIAv2WFhD5MDuQ+jf4MPVlkR78LFYk+BNd/PrlgmjwVffO+/qOLPjgSXz1G7Bg80D81vsIyjD4uLBC+qti9O5Lq8j2twYo+7i9ePS/KBTwoXzC+4E+LPnRgEL6gs5o7juz7PUXeiT4gb109PlHrO103LL79a4o+BY0QvqgyejtSzQE+8PmIPo3KXD0YLNA7YKsovj+HiT6y6H4+lnJIO6Yo6b7QE4w+Yj1cPS04xrvonyW+xKCMPt2nEL4PHBi8TR8EPnIuiz7kIl09NKnbu8qRKr75u4s+AWwQvqdpJLzi7f09QUqKPmkdXj1tkfe7nfcvvklvNTuvbhy+Yr3CuoXxlj6/Y5a5o5MtPc2FkDv9nNk6vvsSOqZ5HL5YnJE7HmqXPtuKI7sdLLK+7bUpPPX9Fj936hq8mrAcvn99tTxmz5k+gA5NvHQjKz2fteY8fWtyPJVdP7whKnI+NyLpPEypib7svuO719InPQAVvTy6vQU93uTIuxQuHr6DbsI870CqPooQF7xcFSU9sen4PGs9Qj2h2wm8ZPcevlpXAD0w9rI+KLo8vP94IT2g+Rw96f+IPTLPL7ztlm8+gHQiPfeTWr4LSMa7T2XbPgb4ED2ymvy+GxclO4V+bj60GtE8QExCvnEu6zs59do+QwSyPBTT8r6k3IA8v81tPiigSDyaAjO+FemmPKa22j6lVw88O17tvhfm7Dwhf20+5ysJut43LL76cgk9LMwWPXO7frvWo/49D3cMPVoMIr53hre6b+LUPp8A/zyRGxc9YZziO1Q4+z37hQI98yAivimAGTyKxdU+LBvrPN8nFj1BKJE8H+ACPsQc8TxQFG0+6himPM4NI77FhQs9YEraPjkCjDwXEuS+83IuPW2IbD5aDQY8xPcWviRfQT3BG9o+IXyrO6EC4L7dRGQ9JVZsPht/ZrvuoBK+CS13PWNVEj0wF9G7u/AXPkMaej2ZICO+jrJfuyjK4D5qDW09SugSPc3hrzuvxRQ+lf1vPQUoI75SjAc8shrhPiTwYj1THBI9tM6LPPkuGT463GU9/hNsPhlRpDz99Qy+G794PaHwDz1Xw408GCglPhSgez0oiGs+JzCoPMroAL7jO4c9xoPZPgeQkzzE9NK+lqKYPV6kHj8BHSA88u8yv5EEsj0gUNk+VNqJuzx0zr4jZ8M9eJ8eP5gOSbwdfjK/VcjcPYBu2T6Iw9a8VB/RvlUt7j3pcms+a9cMvTY8/r1UmPc9yN7ZPskCF70V3dq+KIMEPld9bD5wBzq9nCYWvvs9CT74hBU9hwpGvZSJBj5e/Qk+zNttPjRHO70tYTS+NL8OPlIDGz1gtUm9i2nQPZ6FDz56SyC+PF9BvWDMwT7pUAw+06YgPUFdIr05E5I9ix4NPq6GcD5yhRy9qDNvvgnuET5FNCU9TKgvvRmqPz2AwRI+HtkdvvrSK72ruqY+UZkPPt0qKj3AJRG9exmkPCFzED4q2XI+p4EPvVw2kb6DThU+wJLdPo29Jr3sTha/bCsePuvycz7O1la92nedvnEMIz6yyjM9tAhwvfeDAr2T8iM+oY91PvGkcr0KVa++2dsoPjmsOj1GWYe9O1iNvcrKKT5dYHc+9SyKvch6w75bvS4+bHZCPWLQmb3pleO9RbYvPnACFr6fXZ69NbcgPji2LD7sWEs95O+XvQHpIr6Bui0+980Tvhd0nr0L1t89v8UqPigyrb4N+pm9uNHAPjjYIz6flBG+HY2Kvfeiej3Z7iA+/JZcPXwLiL08DIG+NAkiPijbfz5iXpK9np4Qvy8nJz4jmGQ9AIKpvYVWl77JSyg+s1oNvmedtb0gpPa8DXglPv3fqb4b2ba9du5uPoisHj72iAa/ckqtvYU+/j5B6RM+tZyovobzmL3rvzY+qyoNPi9pCL4rpJG9/hcMvj5wCj4qjae+uj6XvbrMBz6EvAM+9G8FvyPQkb1IX80+cB/yPXt+pr4dYoG9/OWxPaTN5D3a8wS/jaZ7vVPKtz7oh889lJWlvoA+Xr0zzkI9vUjCPayIBL8YWVq9WzGlPicUrT1YzKS+zOo/vdlpXzwV5Z89qisEv9TMPr3xFJU+YL+KPXIdpL5w8ia9kaiBvJo8ez1s2gO/XT4ovZgGhz4wC1E9OYSjvrWjEr0xgiq9jeE2PdCSA7+2DBa9DUx1Pg3HDD2W/KK+B20CvVMEhL08ZuU89lIDv+G0B71cPV8+GlqRPOyCor7dseu81vStvVWyOjwwGQO/fZz5vCVLSz4WIJc6ARSivpYV2byFMNS9vq2pu/vjAr87D+q8pu44PhZhfLzprKG+aHjMvLy5973/7LG8zWD2vdNJ4LxBw9e+0aLFvAJLob6OqhK9ocgMvvE/+bw9dgK/zu0dvaUjEz5oXya9NkQ0v2QoEr3jEdk+1Q5gvcswAr9r2t68bEX2PQbchL0VCjS/ySbLvOT2zj5wqqG9HAACv0PsiLwEpdQ9QXe2vVP6n76a0m+8jLxGvppDw73B4gG/jLWXvGppwD25C9i9jsQzv/RQiLyW78I+A8/0vdLBAb+d3xO8/K2pPe7IBL7Pi5++EnPxu/XEWb6vKgu+UVzuvTNpPrw99gG/44wNvmRon76NYbK83elfvjntE77fkgG/DTXWvOpbiT3kSh6+4w6fvvA3y7xfW2++pqckviljAb//g/G8xO5QPYABL74sOjO/hyjpvFMiqz4TWD2+ZRJlvzllsrxuKR4/d6tPvmYLM7/5Vxq83/+iPkz+Xb4J9mS/ihhIu4mpGz9rT3C+X3OLv0U5FTytV2Y/qU+DvmX7ZL/zB9488CYcP294jL5jITO/9PsgPTfipj67opO+Fjhlv4KvOz1geSE/782cvl9wM794W2893qi0PmT7o75etgG/pSGGPZMaoj2lK6m+NuMzv55fiT2QlMg+sV2wvhQyAr+Aa5k9ebf3PeWStb6zE6G+z1+ePZu3Fr6cy7i+CcACv3dYmD0C6Sw+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.3704218566417694, 0.862652063369751, 0.6408324241638184, 0.7067926526069641, 0.6424024105072021, 0.704544723033905, 0.643830418586731, 0.7023808360099792, 0.35486194491386414, 0.867941677570343, 0.3351086974143982, 0.12471930682659149, 0.06076493486762047, 0.9595835208892822, 0.9443088173866272, 0.9071905016899109, 0.8101104497909546, 0.4039580225944519, 0.8424755930900574, 0.6691440939903259, 0.5873192548751831, 0.7227833867073059, 0.518470823764801, 0.7663354277610779, 0.5499677658081055, 0.6909266710281372, 0.6000205874443054, 0.6496092677116394, 0.6416495442390442, 0.6098460555076599, 0.6762430667877197, 0.4280563294887543, 0.7881345748901367, 0.44517654180526733, 0.7843468189239502, 0.5491093993186951, 0.720304012298584, 0.5353097915649414, 0.7302203178405762, 0.5228539109230042, 0.7390195727348328, 0.4885759651660919, 0.23121054470539093, 0.9006240367889404, 0.7968468070030212, 0.43280747532844543, 0.8218129277229309, 0.3880397081375122, 0.8501409888267517, 0.6692323088645935, 0.37749165296554565, 0.831940233707428, 0.5919114947319031, 0.7414047122001648, 0.5435774922370911, 0.7814743518829346, 0.5154067277908325, 0.7675591707229614, 0.5614597201347351, 0.25875985622406006, 0.11049404740333557, 0.9392894506454468, 0.8906344771385193, 0.7393347024917603, 0.5849305391311646, 0.7357144951820374, 0.5903993248939514, 0.7325564622879028, 0.5951409339904785, 0.27020227909088135, 0.890124499797821, 0.7414395809173584, 0.5737789273262024, 0.747101902961731, 0.5647081136703491, 0.43349379301071167, 0.8456490635871887, 0.422675758600235, 0.14891332387924194, 0.9358048439025879, 0.866979718208313, 0.6477257013320923, 0.7104734182357788, 0.3343040645122528, 0.8837423920631409, 0.3000265955924988, 0.8944603800773621, 0.737734317779541, 0.39631572365760803, 0.8532589673995972, 0.40593230724334717, 0.8532748222351074, 0.40211984515190125, 0.8582062125205994, 0.614926278591156, 0.25683698058128357, 0.9029576182365417, 0.24387361109256744, 0.907710075378418, 0.7756183743476868, 0.44150394201278687, 0.8408247232437134, 0.43454691767692566, 0.8470814824104309, 0.5849999189376831, 0.23604242503643036, 0.9088169932365417, 0.22439728677272797, 0.9132001399993896, 0.7937383055686951, 0.5279405117034912, 0.804244339466095, 0.49281221628189087, 0.1818321794271469, 0.9264438152313232, 0.16466377675533295, 0.9328151345252991, 0.8593114614486694, 0.37187713384628296, 0.8782414197921753, 0.686177134513855, 0.6617769002914429, 0.728885293006897, 0.3895016610622406, 0.8505775928497314, 0.5772995948791504, 0.786056637763977, 0.47462278604507446, 0.8140077590942383, 0.49266523122787476, 0.17261965572834015, 0.9300323128700256, 0.8569923639297485, 0.3558591902256012, 0.8785298466682434, 0.29066774249076843, 0.8968467712402344, 0.7688615322113037, 0.5123347043991089, 0.8096652626991272, 0.567115306854248, 0.7342634797096252, 0.3765783905982971, 0.13958685100078583, 0.9334431290626526, 0.8842563033103943, 0.7608109712600708, 0.5036918520927429, 0.7659969925880432, 0.4404829740524292, 0.8249092102050781, 0.6225041151046753, 0.6757571697235107, 0.6640529632568359, 0.6374146938323975, 0.696735143661499, 0.6027605533599854, 0.7227198481559753, 0.5718508958816528, 0.7437239289283752, 0.5443801283836365, 0.7610464096069336, 0.5198507905006409, 0.7756580710411072, 0.49768534302711487, 0.7882869243621826, 0.4772900938987732, 0.7994855642318726, 0.4580875635147095, 0.19032010436058044, 0.9116104245185852, 0.8271104693412781, 0.6045768857002258, 0.7105083465576172, 0.6136308908462524, 0.7094883918762207, 0.3857044577598572, 0.8438200950622559, 0.6264723539352417, 0.7051078677177429, 0.37881338596343994, 0.15192367136478424, 0.9262439608573914, 0.8586941361427307, 0.3390149474143982, 0.8663967251777649, 0.6815566420555115, 0.35366547107696533, 0.8646475076675415, 0.33449050784111023, 0.12341172993183136, 0.9437460899353027, 0.8987603187561035, 0.23489360511302948, 0.9163520932197571, 0.8179803490638733, 0.4328829050064087, 0.8565827012062073, 0.6541871428489685, 0.6612712144851685], "advantages": [0.8562566637992859, 0.13267900049686432, 0.19159622490406036, 0.47047892212867737, -0.5777991414070129, -0.3338415324687958, -1.3717912435531616, -1.1614481210708618, -2.1901793479919434, -2.8175485134124756, -3.016653537750244, -3.589956521987915, -3.196915626525879, -0.936923086643219, -3.474667549133301, -5.0085225105285645, -5.810187339782715, -6.1565775871276855, -6.966286659240723, -7.394003868103027, -7.659208297729492, -8.722640037536621, -9.01205062866211, -10.068300247192383, -10.387497901916504, -10.76607894897461, -11.859007835388184, -12.2404203414917, -13.349700927734375, -13.737081527709961, -14.865225791931152, -15.260808944702148, -15.791305541992188, -16.885093688964844, -17.416196823120117, -18.52495002746582, -19.714616775512695, -20.119800567626953, -21.337295532226562, -21.749876022338867, -22.9976749420166, -23.41666603088379, -23.96413230895996, -24.559900283813477, -25.809267044067383, -27.0604190826416, -27.578529357910156, -28.857717514038086, -29.355125427246094, -30.662277221679688, -32.123741149902344, -33.61427307128906, -33.80275344848633, -34.068180084228516, -35.61397933959961, -35.84345245361328, -37.437870025634766, -39.15348815917969, -39.1553955078125, -40.92556381225586, -42.536808013916016, -43.56732177734375, -43.932865142822266, -44.057289123535156, -44.17548751831055, -45.95310592651367, -46.07291030883789, -47.88583755493164, -48.00776290893555, -49.85635757446289, -51.50632095336914, -51.687564849853516, -51.880245208740234, -53.752532958984375, -53.95876693725586, 16.800418853759766, 17.167081832885742, 16.44361686706543, 16.810150146484375, 18.162490844726562, 16.310836791992188, 15.591723442077637, 15.698098182678223, 15.32346248626709, 15.739047050476074, 14.946310997009277, 15.394089698791504, 14.573724746704102, 14.539593696594238, 15.436141967773438, 14.322662353515625, 15.242505073547363, 14.12369441986084, 15.080483436584473, 13.946696281433105, 13.799705505371094, 14.582615852355957, 13.400273323059082, 14.21284294128418, 13.019174575805664, 12.789155960083008, 13.546785354614258, 12.623613357543945, 13.403199195861816, 12.484156608581543, 12.581799507141113, 13.71827220916748, 12.192377090454102, 13.367371559143066, 11.826988220214844, 11.387402534484863, 11.639986991882324, 11.143689155578613, 11.68420696258545, 13.833220481872559, 11.818547248840332, 14.190725326538086, 12.068867683410645, 11.512876510620117, 12.141493797302246, 11.625069618225098, 12.483038902282715, 11.46791934967041, 12.392973899841309, 14.363967895507812, 11.919318199157715, 10.838932991027832, 11.819147109985352, 13.793970108032227, 11.315160751342773, 10.1919527053833, 10.893885612487793, 10.449422836303711, 11.507760047912598, 10.366683959960938, 11.408164024353027, 10.329667091369629, 11.325005531311035, 13.12551498413086, 10.775116920471191, 12.52396011352539, 14.353045463562012, 11.462501525878906, 9.224157333374023, 8.7572021484375, 9.283324241638184, 10.888875007629395, 12.7832670211792, 14.483699798583984, 11.475472450256348, 8.645523071289062, 10.7061128616333, 12.696712493896484, 9.534913063049316, 11.67943286895752, 8.439910888671875, 10.71435546875, 7.443734169006348, 9.808943748474121, 6.577042102813721, 8.977127075195312, 5.871778964996338, 8.237143516540527, 5.350921154022217, 7.606163024902344, 5.018596649169922, 7.093194484710693, 4.855752468109131, 6.694051742553711, 4.824299335479736, 4.579370975494385, 5.330660343170166, 6.647513389587402, 8.090636253356934, 6.073055267333984, 7.258737087249756, 5.546675205230713, 4.932263374328613, 5.372824668884277, 6.054940700531006, 4.895090579986572, 4.693336009979248, 5.598761558532715, 5.041888236999512, 4.79701566696167, 5.10745906829834, 4.651683330535889, 4.441667079925537, 4.534855842590332, 3.645660877227783, 3.6229147911071777, 3.9138808250427246, 2.6262741088867188, 1.9431759119033813, 1.8512492179870605, 1.186320185661316, 1.0952794551849365, 0.5830588936805725, 0.45987290143966675, 0.6946868896484375]} +{"type": "SampleBatch", "eps_id": [1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1906370738, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAJzLuL4JwAK/d1iYPQLpLD5+Br6+Xi6ivhFDnz3Cssu93ETBvjyQ+70iMJs9Rbe7vtyGwr7mTKO+tiuMPXAbUb31ysW+P7v/vWUUij3Ui6S+SxLHvsRKpL761Hk9D6oGvHhbyr6XTgS/myh5PaErmz5Lps++Yy6lvjL+iD00K/c8BfTSvi/FBL+SOoo9OravPpZD2L4o8za/JkmYPasHKD/+lN++XUgFv6grsz0hp8Y+z+nkvtNZp74REMM90DX+PaVC6L68agi+nyXIPdpXDL7fn+m+xMCovoKIwj3WOT0+4//svgc3C74uGso9AqKcvUdk7r4eCmw9OPjGPetFrL42ze2+vqKAPhEwuT1a1Ri/mTrrvhExYT0DvKA9sPWNvnqq6r73jhC+rWCVPWkqID2LHOy+b2hYPbP6lj3IN2u+C5LrvrGyEr4Sko09/qyuPZcJ7b6vGlA9aRCRPfU9Pb5nhOy+VL0UvpJ+iT03dQQ+LQHuvnASSD3wyo49dM8QviGB7b6su3g+FQCJPbaX0r5gBOu+SRpAPUpOcD3NWMm9bonqvnmWGL5/QGg9oWFZPg4Q7L4uXjk9d6R5Pa/jfb1rmeu+RDN1Po2QdD0tb6u+tCXpvnlRMj2bIlk9WhLEvJWz6L4JkXM+qSxXPSBPmb4NROa+pCUsPR+lPj0PHRk84NXlvqggcj4baT89MF6Jvghq477rsCY9hG4pPQK6Hj1Z/+K+lNhwPjKbLD16X3a+yJbgvk3OIT165Rg9BEWFPTov4L5DsG8+KToePf7GXL6gyd2+u1wdPaSQDD1iS7Y96mTdvv2fbj5W2xM9HEZFvggC275QPhk9KhMEPbu74z31n9q+taBtPigvDT1aPy++oT/Yvox02j4sVP48Ps/nviPh076lq2w+Via0PAYPGr5Dg9G+jTMSPReAmzw/qxg+sSXRvloSbD5p7bM8StgMvlnJzr6vzg89aGSdPNLfJT5Qbc6+QXhrPp7utzyGIP+9gxLMvhB42T6dhaM8PffRvhK5x76hnB4/cKpAPB9IMr/iYMG+Hz3ZPq4jDruN08y+nwi9vke0aj6snya8/Erdvcevur4SVNk+1QdKvI3Pzr4PV7a+gAhrPtoxp7yW1+u9X/2zvpUiDj3mD7q8LB8vPmiis75J7iO+6QqevL+z6T4SRrW+aJcQPRCEJrzPiCE+iOm0vqgZbD5UpuW7NXANvh2Nsr5qmBE9yhUgvBv/Gz7vL7K+6kIjvh1V3Lv5Q+I+4tGzvu17tb4kkwo72oM7PxVzt757MSO+21SJPH2E4T7bFLm+WnURPUF/0Tx5yRw+w7e4vv/Iaz5Aleo8PowGvidctr4zSg49Kw7VPJBELj4WAba+iPxqPi7w8Dyi2Om9haezvngs2T4CO948s33LvpdPr77PKGo+Bh2dPP9Jxb0l+Ky+udbYPoxUjTzqC8S+7qGovotOHj/TMB08fYsrv95Mor7BpNg+Aox5u4myv76n952+sYlpPrUSOby/1qm9zKGbvtvA2D5OP1S8iCDCvgVMl77WXx4/iz6ovF4HLb9E9pC+1f7YPsp9C73qjMe+P5+Mvn+RHj9gayu9d2Yxv0pw+DrO5co8yy3IuVC9Iz3Vrhw7NixhPof+2jmpcoG+tnPeO8jkyjyTAZi7JMkjPfmu7jsBPGE+05l7u8Iggr6waj88L4zUPp8uErzYTAy/PrmjPHlrYT4X4qK8XDGEvnDKxzxZSs88Uy/NvNOo5jzE78s8S4QtvlmSyLykK6A+hSywPEQX1TwvUZW8rEtNPItvtDxk5Sy+oUOTvOBLmT65xZg82FTZPC1rRLxqZJE6dR6dPNd1LL4gDkS8AnmUPnyGgTwxI9w8uBDKuzZC07uW7YU8UzIsvl1KzrunjZE+1cBUPOeY3TyX2x+6YAgqvP2dXTx0g2M+rERWugJEm77yNZM8rsPdPAmG4bvIaDG8YaWXPDSgYz61nui7Y4Gcvu8QvDzoZt88NHlYvAGtebzAiMA87etjPot3Xbx7xp++awDlPEaI4jyR3KG8COvBvESI6TzhZ2Q+bb2lvKwhpb7iCQc96zbnPACV2ry0ohS9y1kJPYAWZT4Hh+C8ea6svoOtGz2ai+08jeQLvduIWr2hDR49+JcpvnJDEL2U12k+WnwQPRWq9Ty5HPu87wyavUDxEj3p8mY+17cDvXU8wb4VayU97QL9PMuiIr1MpcK9y/InPceSJ773ayq9bEo9PuSKGj2hQgM9Skcbvbob973yKh09jydpPq4pJb38oNm+89EvPZPUBz3B+0e9YdcUvmaJMj0GGSW+B+RTvWSxBj4zVCU9mQ22voQdSb2intA+TTMIPfufI750vCe9HDXMPYc49jzWBRM9X5EfvVaKUr4LGvw8CnUivjxpML1np5g9yxviPELNtL4PTiq9Vue0PoNAqDwlOCG+P1wNvVjoQz34dI48QUwcPTNxCb0e1oK+drWUPGNabz5CYB69YwEPv10Buzw0gCA9QyNMvT+Fjr7kbME8rOoevufwYr3iYuy6qf+nPLbusr65FmO9F7mLPgd7XTxFNAu/q7tMveQuDD9sMi07USqyvt3fH71lVnU+/nONu7/iCr9aPwy9jgsFPxKAeLyIobG+l1jDvAuhXT6RF7W8aq0Kv6nin7xhYgA/POwGvXhQsb5bcBu895hPPghLI72prxq+7wOyu1hRvr0Bqy+9mDSxvsTq7rvMyUo+VgVMve2GCr/eQ1q7zRD6PnlZeL2wILG+kvPSO9xZRz5QWIq9nosavp9EKTwPisS924aQvQpAsb5f0gk8ysFMPvG0nr3W0Rq+EFhLPB9xuL1L5qS9+GaxvlPVLTwveFM+fhezvRQoG77ngHE8MJWpvUtMub1ZlrG+y15WPPujWz5Igce9UZAbvuBTjjxQn5e9QbrNvUvtLz2mMoI8Te24vuL3y73+DBy+3QoOPPcdgr3XNdK9rP2xvopy8jvRb20+GHPgvTZJHL4XNEU8MnxvvXaz5r36IbK+cAsyPCCzcz6f8/S9Cp8cvqgDgDza6FG9bDf7vXjNKz1ePG88FIytvpx/+b0pinI+ZioAPDaDIL8AzO+94H4qPbWUmrtT6qm+iBfuvaUcHb5ICTq8IpomvVxg9L3EX7K+S11HvMJdfj67UgG+fcQcvjnv67uk+0S9YXUEvrqRLD3SuQW8m6Kvvn6YA77uihy+5CF2vA3XWL39uQa+7uotPWO9g7ytXbO+YNsFvmgWHL4LI7280YWAvYv6CL6RMTA9MGvHvO6qub4EGQi+yGQbvotqAb18LZ+9oTQLvjlcsb6GyAe93b1RPs1MEr5Wchq+BwLuvPX2yL2RYxW+kxY3PckV/rzutsy+N3kUvo2RGb4AzB+94MvvvXyLF77FY7C+hGMpvd3yJj63mR6+DP0Jv2gIHL0SgeI+uKMpvjbPr76Xle+8FkINPgKsML7UZRe+qPvYvJu/J74qszO+9GavvqLS87wHkfY9SLc6vnaTFr70GOC8puI5vju6Pb4T/K6+0db9vGS30T0TukS+klMJv9MP7bw0MsU+hbZPvt6Lrr6F9a28ofmqPeCxVr4VJQm/80egvPkgvT6aqmG+bj+uvgmFR7w9mZA95qJovpqDFL5HYjC8FlFnvkqba75yFK6+vmd6vGjJgT3ekXK+RyIUvq6jZbwZtm++UIh1vhl+Tz1qLJm8XfoHv7l+dL5KqBO+DDPwvBhMer67cne+CkFSPZwfDL3W1Au/m2V2voW7Er6Z3ji9QG2HvuBUeb6q7Ky+rolOvW/wXjzPH4C+KjgIv1FsTb0odpQ+spKFvnIwrL5VqzW9NFGUvE4Eib5G4Ae/Bic3veNAhT6tc46+t4mrvvbUIb1RKz299OGRvtORB7+CnSW9CWNvPi9Ol77+XTm/3XYSvYpwAz9YuJ6+Vi9rv6rO0LyBbUs/oyCovoQlOb8fOh286PP8PoqIr744KQe/CDWVOf0dSz6W8LS+/xw5vwdSiztMdfs+JVi8vm8tB7/tl2Y8lJVMPl3Awb6Yjqq+vAeUPPIVtb2eKcW+eUsHvxmLhTwk8lY+CZPKvgNOOb9I76c83vcBP4780b7RbQe/Rx37PDfaYj5ZZ9e+wC+rvpO0Dz1eLXu909PaviyrB7+Mrgo9hAh4PhJB4L7OsKu+RYYePUItIr0gsOO+2O4Hv+1HGz3ks4c+FCDpvjNArL5P/jA9Zwp9vAGS7L6PWxG+arovPaCflr4eBu6+beGsvuCgFz2x5T88lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0, 1], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 0, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 1, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 0], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAH4Gvr5eLqK+EUOfPcKyy73cRMG+PJD7vSIwmz1Ft7u+3IbCvuZMo762K4w9cBtRvfXKxb4/u/+9ZRSKPdSLpL5LEse+xEqkvvrUeT0Pqga8eFvKvpdOBL+bKHk9oSubPkumz75jLqW+Mv6IPTQr9zwF9NK+L8UEv5I6ij06tq8+lkPYvijzNr8mSZg9qwcoP/6U375dSAW/qCuzPSGnxj7P6eS+01mnvhEQwz3QNf49pULovrxqCL6fJcg92lcMvt+f6b7EwKi+gojCPdY5PT7j/+y+BzcLvi4ayj0Copy9R2Tuvh4KbD04+MY960WsvjbN7b6+ooA+ETC5PVrVGL+ZOuu+ETFhPQO8oD2w9Y2+eqrqvveOEL6tYJU9aSogPYsc7L5vaFg9s/qWPcg3a74Lkuu+sbISvhKSjT3+rK49lwntvq8aUD1pEJE99T09vmeE7L5UvRS+kn6JPTd1BD4tAe6+cBJIPfDKjj10zxC+IYHtvqy7eD4VAIk9tpfSvmAE675JGkA9Sk5wPc1Yyb1uieq+eZYYvn9AaD2hYVk+DhDsvi5eOT13pHk9r+N9vWuZ675EM3U+jZB0PS1vq760Jem+eVEyPZsiWT1aEsS8lbPovgmRcz6pLFc9IE+Zvg1E5r6kJSw9H6U+PQ8dGTzg1eW+qCByPhtpPz0wXom+CGrjvuuwJj2Ebik9AroePVn/4r6U2HA+MpssPXpfdr7IluC+Tc4hPXrlGD0ERYU9Oi/gvkOwbz4pOh49/sZcvqDJ3b67XB09pJAMPWJLtj3qZN2+/Z9uPlbbEz0cRkW+CALbvlA+GT0qEwQ9u7vjPfWf2r61oG0+KC8NPVo/L76hP9i+jHTaPixU/jw+z+e+I+HTvqWrbD5WJrQ8Bg8avkOD0b6NMxI9F4CbPD+rGD6xJdG+WhJsPmntszxK2Ay+WcnOvq/ODz1oZJ080t8lPlBtzr5BeGs+nu63PIYg/72DEsy+EHjZPp2Fozw999G+ErnHvqGcHj9wqkA8H0gyv+Jgwb4fPdk+riMOu43TzL6fCL2+R7RqPqyfJrz8St29x6+6vhJU2T7VB0q8jc/Ovg9Xtr6ACGs+2jGnvJbX671f/bO+lSIOPeYPurwsHy8+aKKzvknuI77pCp68v7PpPhJGtb5olxA9EIQmvM+IIT6I6bS+qBlsPlSm5bs1cA2+HY2yvmqYET3KFSC8G/8bPu8vsr7qQiO+HVXcu/lD4j7i0bO+7Xu1viSTCjvagzs/FXO3vnsxI77bVIk8fYThPtsUub5adRE9QX/RPHnJHD7Dt7i+/8hrPkCV6jw+jAa+J1y2vjNKDj0rDtU8kEQuPhYBtr6I/Go+LvDwPKLY6b2Fp7O+eCzZPgI73jyzfcu+l0+vvs8oaj4GHZ08/0nFvSX4rL651tg+jFSNPOoLxL7uoai+i04eP9MwHTx9iyu/3kyivsGk2D4CjHm7ibK/vqf3nb6xiWk+tRI5vL/Wqb3MoZu+28DYPk4/VLyIIMK+BUyXvtZfHj+LPqi8Xgctv0T2kL7V/tg+yn0LveqMx74/n4y+f5EeP2BrK713ZjG/gkeGvj2L2T4BMGS9CdfTvtWuHDs2LGE+h/7aOalygb62c947yOTKPJMBmLskySM9+a7uOwE8YT7TmXu7wiCCvrBqPzwvjNQ+ny4SvNhMDL8+uaM8eWthPhfiorxcMYS+cMrHPFlKzzxTL82806jmPMTvyzxLhC2+WZLIvKQroD6FLLA8RBfVPC9RlbysS008i2+0PGTlLL6hQ5O84EuZPrnFmDzYVNk8LWtEvGpkkTp1Hp0813UsviAORLwCeZQ+fIaBPDEj3Dy4EMq7NkLTu5bthTxTMiy+XUrOu6eNkT7VwFQ855jdPJfbH7pgCCq8/Z1dPHSDYz6sRFa6AkSbvvI1kzyuw908CYbhu8hoMbxhpZc8NKBjPrWe6LtjgZy+7xC8POhm3zw0eVi8Aa15vMCIwDzt62M+i3ddvHvGn75rAOU8RojiPJHcobwI68G8RIjpPOFnZD5tvaW8rCGlvuIJBz3rNuc8AJXavLSiFL3LWQk9gBZlPgeH4Lx5rqy+g60bPZqL7TyN5Au924havaENHj34lym+ckMQvZTXaT5afBA9Far1PLkc+7zvDJq9QPESPenyZj7XtwO9dTzBvhVrJT3tAv08y6IivUylwr3L8ic9x5InvvdrKr1sSj0+5IoaPaFCAz1KRxu9uhv3vfIqHT2PJ2k+riklvfyg2b7z0S89k9QHPcH7R71h1xS+ZokyPQYZJb4H5FO9ZLEGPjNUJT2ZDba+hB1JvaKe0D5NMwg9+58jvnS8J70cNcw9hzj2PNYFEz1fkR+9VopSvgsa/DwKdSK+PGkwvWenmD3LG+I8Qs20vg9OKr1W57Q+g0CoPCU4Ib4/XA29WOhDPfh0jjxBTBw9M3EJvR7Wgr52tZQ8Y1pvPkJgHr1jAQ+/XQG7PDSAID1DI0y9P4WOvuRswTys6h6+5/BiveJi7Lqp/6c8tu6yvrkWY70XuYs+B3tdPEU0C7+ru0y95C4MP2wyLTtRKrK+3d8fvWVWdT7+c427v+IKv1o/DL2OCwU/EoB4vIihsb6XWMO8C6FdPpEXtbxqrQq/qeKfvGFiAD887Aa9eFCxvltwG7z3mE8+CEsjvamvGr7vA7K7WFG+vQGrL72YNLG+xOruu8zJSj5WBUy97YYKv95DWrvNEPo+eVl4vbAgsb6S89I73FlHPlBYir2eixq+n0QpPA+KxL3bhpC9CkCxvl/SCTzKwUw+8bSevdbRGr4QWEs8H3G4vUvmpL34ZrG+U9UtPC94Uz5+F7O9FCgbvueAcTwwlam9S0y5vVmWsb7LXlY8+6NbPkiBx71RkBu+4FOOPFCfl71Bus29S+0vPaYygjxN7bi+4vfLvf4MHL7dCg489x2Cvdc10r2s/bG+inLyO9FvbT4Yc+C9Nkkcvhc0RTwyfG+9drPmvfohsr5wCzI8ILNzPp/z9L0Knxy+qAOAPNroUb1sN/u9eM0rPV48bzwUjK2+nH/5vSmKcj5mKgA8NoMgvwDM773gfio9tZSau1Pqqb6IF+69pRwdvkgJOrwimia9XGD0vcRfsr5LXUe8wl1+PrtSAb59xBy+Oe/ru6T7RL1hdQS+upEsPdK5Bbyboq++fpgDvu6KHL7kIXa8DddYvf25Br7u6i09Y72DvK1ds75g2wW+aBYcvgsjvbzRhYC9i/oIvpExMD0wa8e87qq5vgQZCL7IZBu+i2oBvXwtn72hNAu+OVyxvobIB73dvVE+zUwSvlZyGr4HAu689fbIvZFjFb6TFjc9yRX+vO62zL43eRS+jZEZvgDMH73gy++9fIsXvsVjsL6EYym93fImPreZHr4M/Qm/aAgcvRKB4j64oym+Ns+vvpeV77wWQg0+AqwwvtRlF76o+9i8m78nviqzM770Zq++otLzvAeR9j1Itzq+dpMWvvQY4Lym4jm+O7o9vhP8rr7R1v28ZLfRPRO6RL6SUwm/0w/tvDQyxT6Ftk++3ouuvoX1rbyh+ao94LFWvhUlCb/zR6C8+SC9PpqqYb5uP66+CYVHvD2ZkD3momi+moMUvkdiMLwWUWe+SptrvnIUrr6+Z3q8aMmBPd6Rcr5HIhS+rqNlvBm2b75QiHW+GX5PPWosmbxd+ge/uX50vkqoE74MM/C8GEx6vrtyd74KQVI9nB8MvdbUC7+bZXa+hbsSvpneOL1AbYe+4FR5vqrsrL6uiU69b/BePM8fgL4qOAi/UWxNvSh2lD6ykoW+cjCsvlWrNb00UZS8TgSJvkbgB78GJze940CFPq1zjr63iau+9tQhvVErPb304ZG+05EHv4KdJb0JY28+L06Xvv5dOb/ddhK9inADP1i4nr5WL2u/qs7QvIFtSz+jIKi+hCU5vx86Hbzo8/w+ioivvjgpB78INZU5/R1LPpbwtL7/HDm/B1KLO0x1+z4lWLy+by0Hv+2XZjyUlUw+XcDBvpiOqr68B5Q88hW1vZ4pxb55Swe/GYuFPCTyVj4Jk8q+A045v0jvpzze9wE/jvzRvtFtB79HHfs8N9piPlln177AL6u+k7QPPV4te73T09q+LKsHv4yuCj2ECHg+EkHgvs6wq75Fhh49Qi0ivSCw477Y7ge/7UcbPeSzhz4UIOm+M0Csvk/+MD1nCn28AZLsvo9bEb5qui89oJ+Wvh4G7r5t4ay+4KAXPbHlPzxEe/G+9YQSvoGWGD3Xwom+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.7213780879974365, 0.4139295220375061, 0.821854829788208, 0.47617658972740173, 0.7916843295097351, 0.46624302864074707, 0.8372516632080078, 0.39363035559654236, 0.1346515417098999, 0.9414415955543518, 0.8930035829544067, 0.761457622051239, 0.49698659777641296, 0.8107317090034485, 0.590855062007904, 0.3267541229724884, 0.839712917804718, 0.628971517086029, 0.6975117921829224, 0.5732042193412781, 0.7396360635757446, 0.516476035118103, 0.775055468082428, 0.5403838753700256, 0.7092104554176331, 0.4202454090118408, 0.8226398229598999, 0.6275216937065125, 0.6324013471603394, 0.6606491208076477, 0.6001468300819397, 0.6882561445236206, 0.5700432062149048, 0.7115375399589539, 0.5418425798416138, 0.7315053343772888, 0.5151776075363159, 0.7489808201789856, 0.48962071537971497, 0.7646146416664124, 0.5352729558944702, 0.7173808813095093, 0.45746511220932007, 0.7826154828071594, 0.4412807524204254, 0.791641116142273, 0.5755326151847839, 0.30852392315864563, 0.8614036440849304, 0.7151077389717102, 0.5467969179153442, 0.7358790636062622, 0.480135053396225, 0.2292848378419876, 0.8930206298828125, 0.7712062001228333, 0.49097132682800293, 0.23212605714797974, 0.10669825226068497, 0.9389421939849854, 0.8982293605804443, 0.7953478097915649, 0.4254602789878845, 0.8066340088844299, 0.597457230091095, 0.6792062520980835, 0.604806125164032, 0.3208724558353424, 0.8617354035377502, 0.7043945789337158, 0.5751790404319763, 0.2738915681838989, 0.8897159099578857, 0.22978118062019348, 0.9097772836685181, 0.6228432655334473, 0.7435401082038879, 0.61360764503479, 0.2477831095457077, 0.9095168709754944, 0.7801698446273804, 0.4525766372680664, 0.8327217102050781, 0.46640169620513916, 0.8284577131271362, 0.4737626910209656, 0.8268309235572815, 0.4748761057853699, 0.8278769254684448, 0.5301427841186523, 0.8026209473609924, 0.5149714350700378, 0.8121863007545471, 0.4923042953014374, 0.8245675563812256, 0.46186426281929016, 0.8391351103782654, 0.42364269495010376, 0.8551170229911804, 0.6217204332351685, 0.7291638255119324, 0.3495219051837921, 0.8799537420272827, 0.6966665983200073, 0.6533008217811584, 0.27202799916267395, 0.9022729992866516, 0.7715863585472107, 0.4595305919647217, 0.8235061764717102, 0.5077021718025208, 0.8155178427696228, 0.5410616993904114, 0.7819459438323975, 0.43156698346138, 0.15438759326934814, 0.9313734769821167, 0.8684849143028259, 0.6835302114486694, 0.34543290734291077, 0.8698524832725525, 0.36346012353897095, 0.8670570850372314, 0.3677535951137543, 0.8692216873168945, 0.6417256593704224, 0.718813419342041, 0.3588016629219055, 0.8764611482620239, 0.6657156944274902, 0.6944708824157715, 0.6801754236221313, 0.6809449791908264, 0.6964210867881775, 0.6643675565719604, 0.7144343852996826, 0.3558897376060486, 0.8658710718154907, 0.6425511240959167, 0.7324427366256714, 0.6258729696273804, 0.7475165128707886, 0.39534711837768555, 0.1475498229265213, 0.9318910241127014, 0.8624523282051086, 0.637840747833252, 0.7293006181716919, 0.36049243807792664, 0.869653582572937, 0.3399391770362854, 0.8777683973312378, 0.31278350949287415, 0.8874017000198364, 0.7198340892791748, 0.6339855790138245, 0.26065030694007874, 0.9031516313552856, 0.7724509835243225, 0.44871971011161804, 0.8334833979606628, 0.536676824092865, 0.798795223236084, 0.5082334280014038, 0.8131476044654846, 0.5221230387687683, 0.7985019683837891, 0.5302758812904358, 0.7966148853302002, 0.4706451892852783, 0.825904905796051, 0.45574015378952026, 0.16687451303005219, 0.9280266165733337, 0.1487257480621338, 0.9340620040893555, 0.8719624280929565, 0.6773000359535217, 0.6666180491447449, 0.7052404880523682, 0.6365298628807068, 0.7281424403190613, 0.39164525270462036, 0.14720593392848969, 0.9347227811813354, 0.8628158569335938, 0.35975733399391174, 0.8720171451568604, 0.6704529523849487, 0.6713008284568787, 0.3087306618690491, 0.8904967904090881, 0.7295355796813965, 0.5921083092689514, 0.7583100199699402, 0.5478333234786987, 0.786098062992096, 0.5016189813613892, 0.7788039445877075, 0.5348271131515503], "advantages": [-1.2552050352096558, -1.0517654418945312, -0.4892127811908722, -1.290427327156067, -0.7073538899421692, -1.5826246738433838, -2.528796434402466, -2.309816837310791, -3.239999294281006, -3.818681478500366, -4.268354892730713, -4.434645175933838, -4.364510536193848, -5.139334678649902, -5.211651802062988, -4.85575532913208, -3.6680212020874023, -4.993865489959717, -5.984230995178223, -5.776572227478027, -6.801794052124023, -6.704524517059326, -7.753292083740234, -7.781363010406494, -7.046987056732178, -8.465328216552734, -9.592223167419434, -9.8331937789917, -9.31308364868164, -10.794843673706055, -10.369719505310059, -11.892854690551758, -11.556217193603516, -13.119613647460938, -12.866257667541504, -14.468029975891113, -14.29408073425293, -15.931794166564941, -15.834827423095703, -17.505586624145508, -17.48463249206543, -16.117403030395508, -18.6453800201416, -20.416658401489258, -20.494916915893555, -22.29063606262207, -22.42340850830078, -21.14980125427246, -18.51313018798828, -21.95074462890625, -24.68416976928711, -23.370380401611328, -26.152067184448242, -28.225223541259766, -29.922889709472656, -30.89512825012207, -30.91232681274414, -32.97352600097656, -34.50631332397461, -35.56045150756836, -37.430458068847656, -38.79111099243164, -38.971351623535156, -40.95563507080078, -41.212249755859375, -39.90666961669922, -43.00902557373047, -41.7252082824707, -38.9774055480957, -43.00685501098633, -46.236942291259766, -44.87086486816406, -41.99107360839844, -46.076778411865234, -43.07352066040039, 14.24731159210205, 14.364058494567871, 13.964797019958496, 14.079599380493164, 15.526520729064941, 13.997466087341309, 13.594863891601562, 14.129584312438965, 13.175332069396973, 13.689531326293945, 12.755794525146484, 13.250836372375488, 12.33254337310791, 12.810294151306152, 11.90279483795166, 12.02852725982666, 11.621094703674316, 11.759895324707031, 11.347976684570312, 11.51004409790039, 11.089213371276855, 11.288022994995117, 10.853058815002441, 11.106826782226562, 10.651590347290039, 11.22917652130127, 10.262110710144043, 10.641066551208496, 10.127187728881836, 10.698798179626465, 9.804163932800293, 10.397441864013672, 9.800763130187988, 10.372212409973145, 12.024816513061523, 9.812178611755371, 9.186883926391602, 9.624637603759766, 11.11975383758545, 9.173702239990234, 8.791208267211914, 10.102252960205078, 9.235032081604004, 9.602256774902344, 11.004692077636719, 13.010529518127441, 10.338736534118652, 12.168316841125488, 9.745804786682129, 11.381994247436523, 9.197344779968262, 8.293270111083984, 9.029199600219727, 10.332113265991211, 8.525392532348633, 7.811553955078125, 8.328146934509277, 7.667278289794922, 8.115945816040039, 7.490826606750488, 7.877608776092529, 7.272388458251953, 7.672658443450928, 7.33319091796875, 7.645315647125244, 7.077125549316406, 7.340902805328369, 6.7694220542907715, 7.184791088104248, 8.917820930480957, 7.601263523101807, 7.1260223388671875, 7.2899675369262695, 6.844161510467529, 7.474212646484375, 6.907777309417725, 7.642219543457031, 7.01216459274292, 7.878201007843018, 7.169477462768555, 7.030662536621094, 6.97844123840332, 8.073103904724121, 7.214142322540283, 6.834805011749268, 6.7157368659973145, 6.278583526611328, 6.5770697593688965, 6.0395188331604, 6.468071937561035, 5.8187785148620605, 5.413116455078125, 5.285552024841309, 4.815357208251953, 4.74419641494751, 5.347853183746338, 4.5259599685668945, 5.2427897453308105, 6.725831031799316, 5.618044376373291, 7.248331546783447, 6.096678733825684, 4.939546585083008, 3.666160821914673, 4.604979515075684, 3.271275281906128, 4.3152594566345215, 2.9351181983947754, 1.6809160709381104, 0.9655347466468811, 0.7439358830451965, 1.3089163303375244, 0.07055867463350296, 0.632910430431366, 1.5766246318817139, 0.25799182057380676, -1.0514847040176392, -0.48950424790382385, 0.3930617570877075, -0.9520876407623291, -0.087607242166996, -1.4689116477966309, -0.6378072500228882, 0.25155287981033325, -0.8568115234375]} +{"type": "SampleBatch", "eps_id": [1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 1552982767, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAER78b71hBK+gZYYPdfCib5b8vK+ueJSPc+LAj2ovQy/ZGvyvtaGE76kBKs8dSZ9vg/l876y0K2+m4OCPMzlVD3+Xve+KhAUvq4HizzXSnG+Cdr4vlgOrr6w2Eg8P2N/PTNV/L5FfRS+B0ddPMzfZ75V0f2+nT6uvuQTEzz9U5A9vKYAvy0bCb+SKyo8/GC7PrdkA7/gYq6+3AuRPEXYnD0lIwW/ZjsVvguYnTzcgle+KuIFv7YkST1pOXY8f3j+vsyhBb+RuxW+3bmmO1tsTL50YQa/TM6uvp2Ujzqp2ME99SAIv6vSFb4X2kM7F25KvrvgCL+TwEc9wQh9utKV+r7PoAi/VdoVvm8wMLxbxkm+n2AJv2luSD3UwXC8zHf8vnwgCb9RexW+FyvJvPoBUr7S3wm/+bFKPQHF6ry7YgG/9Z4Jv660FL7HyR69Ajljvk1dCr8B9q2+TPcwvQfWbj2kGgy/x4ITvnUwLL1KoX2+dNcMv4VWrb7MekC9Ts4APTOTDr+dcAi/UOc9vXotnj7GTRG/x6esvluYJL0dVwM7xQcTvysgCL9UbiS9jECQPrzAFb+x7Dm/w1kNvWC+Dz+reBm/ldoHv4+0vrzCK4Q+PTAcv9C3Ob8eaZS8CxwLPx3nH7/uswe/o4Xtu0nxej7qnSK/XHOrvm7WGbtSY0y91FQkv98PD77kPVu76MevvvILJb91KmM9aU8nvDwWI79BwyS/t90OvtsHvLy5+bG+H3olvx8qq76b+/S8rfp+vU4wJ783GQ6+mC7/vKV6ur4x5ie/1beqvnttHb058qa9Opspv98WDb4CGyS9dqvFvtJPKr9hGG09kLtDvefsML/0Ayq/5tELvkpZfL0y5NO+7LYqv79Wqb4uII+9EpkQvm5oLL/3WAa/3eiUvYG4Az5KGC+/jAM4vwukj73Mh8s+cMYyv6jTBb97t369aymrPaJzNb9NiTe/yd53vatGtj5XHzm/+mAFv8K0Wr1WzTc9Pco7v8EfN7+xB1e98PajPtVzP7/r/QS/s8s8vfYMPDzAHEK/McQ2v//aO70QG5Q+hMRFv/KKaL+WKCS9MzARPyJrSr8idDa/hmXrvN4yhj5MEU6/iGYEv/hzwLzbySG9MLdQv+66pL6u7Ma8CtOtvuVcUr/KOQS/YIz+vPuHX73kAVW/OQ82v6u+A709n2k+CaZYv5T+A78tHOK8r5aYvdlJW7+446O+MVHuvKtkwL5o7Vy/nCD/vQbxFb1Sby2/sJBdvxNqo77OcE29uQfLvgczX7+mdwO/6+xtvbHv9b0k1GG/xKGivk/Dd71mbNy+enRjv/8IA7/tg42930YhvmETZr9xtzS/ZveTvd6g5j2msGm/OIUCv5Vaj70xz06+6Uxsv8kyNL9QoJe9J/KKPYfnb78c/wG/6NiUvYgufb4cgXK/7qkzv335nr37Y7E8/Rh2vzBPZb9uFp698uaTPg2ver9rGTO/ZUGSvZoy3rwKRH6/TsVkv85dk71RA3g+rGuBv4aTMr8ncom9hwmUvdQ0g79fZQC/G2iMvVJOxb6GfYS/7xQyv+4wnL1opOu9aUWGv7W7Y79r56C9rUgcPlKrEj3bJSU8reQjvDRjz7y2fhM9pUVSPlMwLLw1bKS+FVEkPdHzKTyrtYq8BDQCvZ8qJT3bq1I+8+qPvOvWqL4rBTY9EggyPEjyxbwryi69DOk2PSlHUz4h8My8GI6vvgXQRz1biz08zY4CvVNZbr2jwkg903k7viVTB71upmU+IsM5PTa9TDzT5+m8xBOhvTPJOj3wkzq+r8r2vEXOUT4V3Cs9OHlaPAs51bwn88a9u/MsPdXUVT6JI+W8mbzLvv8OPj3NGmc80yoTvb3b6b3PNj89ucFWPoiFHL1nAta+B2VQPQdveDxbwz693+QMvgajUT13+1c+3ghKvSWi475X6mI9/miHPLt0br2x4yu+/URkPVv6Nb4GNXy9xOPYPRW2VT3Jab6+E4hzvT3QwT7BPjc9+TU0vnqFVL00sYo9CtQoPaybvb5F+U69aOuvPrF9Cj3wjxC/mtMyvUezHj9Ndrg8xFlCv+AKAL3eBmY/kE/wO2xLEL9OvFm8h6AYP1oVgbvGl7y+xvuyur9FmT67PTm8Z1wxvl9xlztAs8w7I/9xvN+dvL5viZs7/8uZPgxbtbyEfzG+yzIwPE3FFjxfwdG8Yy2xPL02MzwZZY++OTbOvNTQMb7W4a47um+DPI6p6rxoSK88p2W5OzjGjL4bKOe8CvoxvmyXJjko2J88BNEBvVjkvL7C8g86nt6fPggKIL0yZhC/aqDeO+znGj83P0696PG8vuXLmjzPDqE+ZnpsvcR9EL/BVc48QfccP5Bbjb3PQ72+hWUZPZExqD61f5y9b0ozvrtOND2MCkQ9paujvRS5njx2Ojg9EiZsvnvgor2UlzS+IlYlPd92mz2+Gaq958aUPBeOKz3MqFC+T1upvXDLNb6/3Bo9aorQPeOgsL38eos8NDQjPST7Nr5a7q+9Ce42vsCQFD3vUQE+jT+3vZ+fv7466R49+UHcPgCUxr2vBji++yZCPf+VGT5u8M29SNVwPGxwTj1cbgK+TFbNvSN0Ob4xAUQ9yhw5PlfB1L1p9sC+TNBSPSjw+T43MeS9qeY6vsPNej0wOVk+FavrvUpIPzw/F4Y9NFF4vakw6709vVI+jpuDPar1qb6xwuK9MPkgPJEFbD0VhaG8q1vivR75UD4TaGo901yWvsj/2b0MHwY8OVlSPaAZhzzyqdm98CJAvhSzUz2YXKY+bFnhvf4S3DtBUW49A55IPQAT4b3Vx00+alRyPXYwZr7O19i93FXKPiPqXz3TGf+++qfIvT0fTD41GTc95WBBvsR9wL01n8k+0KAnPZ8n776NXLC9tJsWPwldAT2lIT+/o0OYvS0ZyT4ZZ4g8RGjjviUtiL2XJ0o+kIf+O9THFb4VF4C9BycPO4Ornjv6bBg+LQCAvWX4ST54HAA8nbYRvgDYb71u1Mg+RPeiOytx3b4Dtk+9cshJPr31cLtekw2+f5E/vQUQAjugFtO7CfEcPuBnP713rUW+pUpdu8VH4z5QOE+9P0QLO+dFtDvxwxk+wAtPvVrmST5qVws80ygQvtfkPr1uUvw6vWu6O2pIHj54vD69OLFJPmfcDzy/lAu+0JkuvQGvyD7ZY8Y7Ijjavs99Dr2feUk+fNwhu3rIBr4zv/y8dqjIPhsxp7vFpNm+S4m8vHRMFj9J4168yCg4v7qwOLzmysg+TU7lvE+s3L57u2C7hzJKPuz1Fb27zha+NVIIOm5CyT52BiK93A3nvn5TCTwnT0s+cP5GvdxuL76VYko8hfyHO04HVb0WXtg9tb5LPNHGTD60X0y9xNlPvgGjhjxSmbc7fABdvYCklj0Djoc8S0pOPuf5Vr1JTnG+qI+oPM9W6TvYR2q96PIjPVW6qTyeoD++bQBnvX7VoD5GEYs8VkjDvqtETb1ZqBY/dCcZPGQUPr7MDh29dZmPPkGouDsqNiQ89xQGvXIVxbzIOb87sAs9voANCL0/KIQ+FHkMOw2nMzyu0OW8G7c3vV3YGjvwblM+7CntvAtIsb5svdQ7kP5APGnyEr0csYC9r3XcO50yO743GBi9o41fPmxOSTtAOMG+2DUGvUmE/z4Hq5K7YnASv8GnurwM/0c/UmOCvK3RwL51oeq7e4P2Pg8XwLwAuzm+PdAhO48EPz6Uzt28PpxgPHooyzshyNe9k4/bvG7bOb6iG4Y7FNBBPklM+bxQM1489hIBPKEj0b10E/e8NAg6vj45vzsUrEU+qWsKvVX9Wjzt3R481knIvVtTCb1LQjq+OaT9O/WtSj7wORi9LuhWPJ6tPzzHCL293CYXvRgQVT7GbiE88TXDvlUbBr3D21E8zv0RO2kZr722DgW9ZeNUPiq2BzqpRMG+hQ3ovLY2UTxba+a771Ktve/15bwGoDq+B/EOvNvDUj4M6QG9IedUPE7+lruffre9iNgAvTZuOr4vttG7uXdOPqHCD73eelc8CiUbu46Zvr3Rrg69dE06vnWQirtCpEs+S5YdvR8GWTz8vIO5bdvCvYCAHL25PDq+BC0Nu5MySj6jZiu9MJZZPCpG6zrdaMS9IVAqvWU7Or66CgG59RRKPik2Ob3bMVk8hJl6O5ZUw70mIDi9vlJVPiot+zo8Eca+Sg8nvTzZVzxpu767Vp6/vQH7Jb1eYVU+zAz8uxSyxr756BS9lwJbPLAwfbzbWsi9lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 0, 0, 1, 1, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 1, 0], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAFvy8r654lI9z4sCPai9DL9ka/K+1oYTvqQEqzx1Jn2+D+XzvrLQrb6bg4I8zOVUPf5e974qEBS+rgeLPNdKcb4J2vi+WA6uvrDYSDw/Y389M1X8vkV9FL4HR108zN9nvlXR/b6dPq6+5BMTPP1TkD28pgC/LRsJv5IrKjz8YLs+t2QDv+Birr7cC5E8RdicPSUjBb9mOxW+C5idPNyCV74q4gW/tiRJPWk5djx/eP6+zKEFv5G7Fb7duaY7W2xMvnRhBr9Mzq6+nZSPOqnYwT31IAi/q9IVvhfaQzsXbkq+u+AIv5PARz3BCH260pX6vs+gCL9V2hW+bzAwvFvGSb6fYAm/aW5IPdTBcLzMd/y+fCAJv1F7Fb4XK8m8+gFSvtLfCb/5sUo9AcXqvLtiAb/1ngm/rrQUvsfJHr0COWO+TV0KvwH2rb5M9zC9B9ZuPaQaDL/HghO+dTAsvUqhfb501wy/hVatvsx6QL1OzgA9M5MOv51wCL9Q5z29ei2ePsZNEb/Hp6y+W5gkvR1XAzvFBxO/KyAIv1RuJL2MQJA+vMAVv7HsOb/DWQ29YL4PP6t4Gb+V2ge/j7S+vMIrhD49MBy/0Lc5vx5plLwLHAs/Hecfv+6zB7+jhe27SfF6PuqdIr9cc6u+btYZu1JjTL3UVCS/3w8PvuQ9W7vox6++8gslv3UqYz1pTye8PBYjv0HDJL+33Q6+2we8vLn5sb4feiW/Hyqrvpv79Lyt+n69TjAnvzcZDr6YLv+8pXq6vjHmJ7/Vt6q+e20dvTnypr06mym/3xYNvgIbJL12q8W+0k8qv2EYbT2Qu0O95+wwv/QDKr/m0Qu+Sll8vTLk077stiq/v1apvi4gj70SmRC+bmgsv/dYBr/d6JS9gbgDPkoYL7+MAzi/C6SPvcyHyz5wxjK/qNMFv3u3fr1rKas9onM1v02JN7/J3ne9q0a2PlcfOb/6YAW/wrRavVbNNz09yju/wR83v7EHV73w9qM+1XM/v+v9BL+zyzy99gw8PMAcQr8xxDa//9o7vRAblD6ExEW/8opov5YoJL0zMBE/ImtKvyJ0Nr+GZeu83jKGPkwRTr+IZgS/+HPAvNvJIb0wt1C/7rqkvq7sxrwK062+5VxSv8o5BL9gjP68+4dfveQBVb85Dza/q74DvT2faT4Jpli/lP4Dvy0c4ryvlpi92Ulbv7jjo74xUe68q2TAvmjtXL+cIP+9BvEVvVJvLb+wkF2/E2qjvs5wTb25B8u+BzNfv6Z3A7/r7G29se/1vSTUYb/EoaK+T8N3vWZs3L56dGO//wgDv+2Djb3fRiG+YRNmv3G3NL9m95O93qDmPaawab84hQK/lVqPvTHPTr7pTGy/yTI0v1Cgl70n8oo9h+dvvxz/Ab/o2JS9iC59vhyBcr/uqTO/ffmevftjsTz9GHa/ME9lv24Wnr3y5pM+Da96v2sZM79lQZK9mjLevApEfr9OxWS/zl2TvVEDeD6sa4G/hpMyvydyib2HCZS91DSDv19lAL8baIy9Uk7FvoZ9hL/vFDK/7jCcvWik671pRYa/tbtjv2vnoL2tSBw+aYyIv5GEMb8Sp5q9J7UnvrZ+Ez2lRVI+UzAsvDVspL4VUSQ90fMpPKu1irwENAK9nyolPdurUj7z6o+869aovisFNj0SCDI8SPLFvCvKLr0M6TY9KUdTPiHwzLwYjq++BdBHPVuLPTzNjgK9U1luvaPCSD3TeTu+JVMHvW6mZT4iwzk9Nr1MPNPn6bzEE6G9M8k6PfCTOr6vyva8Rc5RPhXcKz04eVo8CznVvCfzxr278yw91dRVPokj5byZvMu+/w4+Pc0aZzzTKhO9vdvpvc82Pz25wVY+iIUcvWcC1r4HZVA9B294PFvDPr3f5Ay+BqNRPXf7Vz7eCEq9JaLjvlfqYj3+aIc8u3RuvbHjK779RGQ9W/o1vgY1fL3E49g9FbZVPclpvr4TiHO9PdDBPsE+Nz35NTS+eoVUvTSxij0K1Cg9rJu9vkX5Tr1o668+sX0KPfCPEL+a0zK9R7MeP012uDzEWUK/4AoAvd4GZj+QT/A7bEsQv068WbyHoBg/WhWBu8aXvL7G+7K6v0WZPrs9ObxnXDG+X3GXO0CzzDsj/3G83528vm+Jmzv/y5k+DFu1vIR/Mb7LMjA8TcUWPF/B0bxjLbE8vTYzPBllj745Ns681NAxvtbhrju6b4M8jqnqvGhIrzynZbk7OMaMvhso57wK+jG+bJcmOSjYnzwE0QG9WOS8vsLyDzqe3p8+CAogvTJmEL9qoN477OcaPzc/Tr3o8by+5cuaPM8OoT5memy9xH0Qv8FVzjxB9xw/kFuNvc9Dvb6FZRk9kTGoPrV/nL1vSjO+u040PYwKRD2lq6O9FLmePHY6OD0SJmy+e+CivZSXNL4iViU933abPb4Zqr3nxpQ8F44rPcyoUL5PW6m9cMs1vr/cGj1qitA946Cwvfx6izw0NCM9JPs2vlrur70J7ja+wJAUPe9RAT6NP7e9n5+/vjrpHj35Qdw+AJTGva8GOL77JkI9/5UZPm7wzb1I1XA8bHBOPVxuAr5MVs29I3Q5vjEBRD3KHDk+V8HUvWn2wL5M0FI9KPD5Pjcx5L2p5jq+w816PTA5WT4Vq+u9Skg/PD8Xhj00UXi9qTDrvT29Uj6Om4M9qvWpvrHC4r0w+SA8kQVsPRWFobyrW+K9HvlQPhNoaj3TXJa+yP/ZvQwfBjw5WVI9oBmHPPKp2b3wIkC+FLNTPZhcpj5sWeG9/hLcO0FRbj0Dnkg9ABPhvdXHTT5qVHI9djBmvs7X2L3cVco+I+pfPdMZ/776p8i9PR9MPjUZNz3lYEG+xH3AvTWfyT7QoCc9nyfvvo1csL20mxY/CV0BPaUhP7+jQ5i9LRnJPhlniDxEaOO+JS2IvZcnSj6Qh/471McVvhUXgL0HJw87g6ueO/psGD4tAIC9ZfhJPngcADydthG+ANhvvW7UyD5E96I7K3HdvgO2T71yyEk+vfVwu16TDb5/kT+9BRACO6AW07sJ8Rw+4Gc/vXetRb6lSl27xUfjPlA4T70/RAs750W0O/HDGT7AC0+9WuZJPmpXCzzTKBC+1+Q+vW5S/Dq9a7o7akgePni8Pr04sUk+Z9wPPL+UC77QmS69Aa/IPtljxjsiONq+z30OvZ95ST583CG7esgGvjO//Lx2qMg+GzGnu8Wk2b5Liby8dEwWP0njXrzIKDi/urA4vObKyD5NTuW8T6zcvnu7YLuHMko+7PUVvbvOFr41Ugg6bkLJPnYGIr3cDee+flMJPCdPSz5w/ka93G4vvpViSjyF/Ic7TgdVvRZe2D21vks80cZMPrRfTL3E2U++AaOGPFKZtzt8AF29gKSWPQOOhzxLSk4+5/lWvUlOcb6oj6g8z1bpO9hHar3o8iM9VbqpPJ6gP75tAGe9ftWgPkYRizxWSMO+q0RNvVmoFj90Jxk8ZBQ+vswOHb11mY8+Qai4Oyo2JDz3FAa9chXFvMg5vzuwCz2+gA0IvT8ohD4UeQw7DaczPK7Q5bwbtze9XdgaO/BuUz7sKe28C0ixvmy91DuQ/kA8afISvRyxgL2vddw7nTI7vjcYGL2jjV8+bE5JO0A4wb7YNQa9SYT/PgerkrticBK/wae6vAz/Rz9SY4K8rdHAvnWh6rt7g/Y+DxfAvAC7Ob490CE7jwQ/PpTO3bw+nGA8eijLOyHI172Tj9u8bts5vqIbhjsU0EE+SUz5vFAzXjz2EgE8oSPRvXQT97w0CDq+Pjm/OxSsRT6pawq9Vf1aPO3dHjzWSci9W1MJvUtCOr45pP079a1KPvA5GL0u6FY8nq0/PMcIvb3cJhe9GBBVPsZuITzxNcO+VRsGvcPbUTzO/RE7aRmvvbYOBb1l41Q+KrYHOqlEwb6FDei8tjZRPFtr5rvvUq297/XlvAagOr4H8Q6828NSPgzpAb0h51Q8Tv6Wu59+t72I2AC9Nm46vi+20bu5d04+ocIPvd56VzwKJRu7jpm+vdGuDr10TTq+dZCKu0KkSz5Llh29HwZZPPy8g7lt28K9gIAcvbk8Or4ELQ27kzJKPqNmK70wllk8KkbrOt1oxL0hUCq9ZTs6vroKAbn1FEo+KTY5vdsxWTyEmXo7llTDvSYgOL2+UlU+Ki37OjwRxr5KDye9PNlXPGm7vrtWnr+9AfslvV5hVT7MDPy7FLLGvvnoFL2XAls8sDB9vNtayL2k0BO98uI5vp2fjrwff0I+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.24131493270397186, 0.895858645439148, 0.7556158900260925, 0.5627695322036743, 0.7457023859024048, 0.5738638639450073, 0.7374980449676514, 0.4179460108280182, 0.8396899104118347, 0.6064826846122742, 0.2953885793685913, 0.8750197887420654, 0.7103742361068726, 0.6078583002090454, 0.29195210337638855, 0.8786695599555969, 0.27571606636047363, 0.8873761296272278, 0.25129878520965576, 0.8985724449157715, 0.7796880006790161, 0.4905412793159485, 0.8019941449165344, 0.5516629815101624, 0.746748685836792, 0.5776807069778442, 0.2728274464607239, 0.8870747089385986, 0.2769257128238678, 0.8864772319793701, 0.7282561659812927, 0.4269391894340515, 0.18631026148796082, 0.9157050251960754, 0.8317717909812927, 0.3750867247581482, 0.8453786373138428, 0.3430943787097931, 0.13919605314731598, 0.935564398765564, 0.8849738836288452, 0.7536598443984985, 0.49243804812431335, 0.7689984440803528, 0.5307973623275757, 0.7431386709213257, 0.5617672801017761, 0.7192195653915405, 0.5866039395332336, 0.302573025226593, 0.8675600290298462, 0.6909656524658203, 0.3980485200881958, 0.8212164640426636, 0.6238670945167542, 0.655376672744751, 0.3653869032859802, 0.16502514481544495, 0.9216912388801575, 0.8561554551124573, 0.29327332973480225, 0.8727355003356934, 0.7451703548431396, 0.4868038594722748, 0.7732892632484436, 0.43962687253952026, 0.7995861172676086, 0.6083648800849915, 0.6436713337898254, 0.642200767993927, 0.6060551404953003, 0.32846659421920776, 0.8494446873664856, 0.710797905921936, 0.5141618847846985, 0.46378377079963684, 0.8353226184844971, 0.4367847442626953, 0.8474537134170532, 0.4021541178226471, 0.8611968159675598, 0.6393392086029053, 0.7144891023635864, 0.6649620532989502, 0.6920929551124573, 0.3124138116836548, 0.889151394367218, 0.27347052097320557, 0.9007779359817505, 0.23188798129558563, 0.9121896028518677, 0.8089476823806763, 0.5409591794013977, 0.772519588470459, 0.5850352644920349, 0.25397947430610657, 0.09775908291339874, 0.950168251991272, 0.9098732471466064, 0.771769642829895, 0.5680984854698181, 0.7832019329071045, 0.4519575536251068, 0.8285578489303589, 0.4541161358356476, 0.8291754722595215, 0.5492079854011536, 0.20560041069984436, 0.923535943031311, 0.17992544174194336, 0.9317284822463989, 0.8505476117134094, 0.6034125089645386, 0.7351111769676208, 0.6447386741638184, 0.7040464878082275, 0.6819301247596741, 0.6704595685005188, 0.2843828797340393, 0.9047502279281616, 0.7631457448005676, 0.5617387890815735, 0.20330935716629028, 0.9253365397453308, 0.8364806771278381, 0.590406596660614, 0.7320359349250793, 0.642178475856781, 0.6925275325775146, 0.31479865312576294, 0.8909397125244141, 0.7377358675003052, 0.4197559058666229, 0.832724392414093, 0.4501296281814575, 0.17669673264026642, 0.924437940120697, 0.8319843411445618, 0.5569643378257751, 0.7775733470916748, 0.44905632734298706, 0.8366489410400391, 0.5699752569198608, 0.23228183388710022, 0.9086724519729614, 0.7788719534873962, 0.55019611120224, 0.7831694483757019, 0.45721670985221863, 0.8337205648422241, 0.43971604108810425, 0.15513934195041656, 0.9351429343223572, 0.8691209554672241, 0.338882714509964, 0.8880844116210938, 0.7198560833930969, 0.6208048462867737, 0.7629486322402954, 0.5596981644630432, 0.8015178442001343, 0.5096896886825562, 0.20435230433940887, 0.9162795543670654, 0.7844460010528564, 0.5726713538169861, 0.7668732404708862, 0.40077295899391174, 0.8630709648132324, 0.6452500224113464, 0.29033249616622925, 0.10666076093912125, 0.9473624229431152, 0.900137186050415, 0.7340761423110962, 0.6292524337768555, 0.740675151348114, 0.6198363900184631, 0.7485766410827637, 0.608159601688385, 0.7577989101409912, 0.4061185419559479, 0.8529402017593384, 0.3982008397579193, 0.8579545617103577, 0.619327187538147, 0.7435746788978577, 0.6222450137138367, 0.7422121167182922, 0.6229832172393799, 0.7424502372741699, 0.6216181516647339, 0.7442542314529419, 0.6181512475013733, 0.7476000189781189, 0.3874887228012085, 0.8613584637641907, 0.37104445695877075, 0.8689646124839783, 0.6546879410743713], "advantages": [7.580617904663086, 8.718170166015625, 7.821845054626465, 6.710877418518066, 7.57504415512085, 6.450265884399414, 7.2691426277160645, 6.134671688079834, 4.556761264801025, 5.388761520385742, 6.112790584564209, 7.11945104598999, 6.005244255065918, 4.8401079177856445, 5.466280937194824, 6.50697660446167, 5.222599983215332, 6.333145618438721, 4.92026424407959, 6.132260322570801, 4.581396102905273, 3.320007801055908, 3.9024581909179688, 2.618546962738037, 1.3531938791275024, 1.6328734159469604, 0.42050260305404663, -1.108751893043518, -0.7711300253868103, -2.2222580909729004, -1.945410966873169, -1.7926582098007202, -1.2718160152435303, 0.16502735018730164, -1.7258199453353882, -3.1616945266723633, -2.4941656589508057, -4.030231952667236, -3.2669949531555176, -1.4177271127700806, -3.6995935440063477, -5.53087043762207, -6.868790626525879, -7.898977279663086, -8.268068313598633, -9.268308639526367, -9.646649360656738, -10.629170417785645, -11.01728343963623, -11.992298126220703, -12.800217628479004, -13.534740447998047, -13.965821266174316, -13.630119323730469, -15.145596504211426, -16.13077735900879, -16.555500030517578, -16.08648109436035, -14.561285018920898, -16.86104393005371, -18.68360710144043, -17.8935546875, -19.820213317871094, -21.19304847717285, -21.2159423828125, -22.66560935974121, -22.569801330566406, -24.100130081176758, -25.109458923339844, -25.794538497924805, -26.832387924194336, -27.45136260986328, -27.0532283782959, -28.74446678161621, -29.91606330871582, 20.753456115722656, 20.834613800048828, 20.694181442260742, 20.807846069335938, 20.65848159790039, 20.82159996032715, 20.657909393310547, 21.5726261138916, 20.457366943359375, 21.328603744506836, 20.27153205871582, 20.618633270263672, 20.35637664794922, 20.83203887939453, 20.537174224853516, 21.197174072265625, 20.864004135131836, 21.806888580322266, 23.83989715576172, 21.514575958251953, 23.422138214111328, 25.847166061401367, 27.924537658691406, 24.49230194091797, 21.513973236083984, 19.95418357849121, 21.1712589263916, 19.751596450805664, 19.605371475219727, 19.815196990966797, 19.73043441772461, 19.922773361206055, 20.94406509399414, 22.653141021728516, 20.381067276000977, 21.9349365234375, 19.80426597595215, 18.8150634765625, 18.747034072875977, 18.672504425048828, 18.58218002319336, 18.494504928588867, 18.377090454101562, 18.2767391204834, 18.862350463867188, 17.799306869506836, 17.587078094482422, 17.441600799560547, 17.999296188354492, 16.841815948486328, 16.490354537963867, 16.861783981323242, 16.066755294799805, 16.38523292541504, 15.610090255737305, 15.543232917785645, 14.969192504882812, 15.117175102233887, 16.25647735595703, 14.667774200439453, 15.825532913208008, 18.732955932617188, 15.693540573120117, 14.018409729003906, 13.383036613464355, 13.4642972946167, 14.775519371032715, 13.10832405090332, 12.505882263183594, 12.744063377380371, 11.837752342224121, 11.855700492858887, 11.318857192993164, 11.307025909423828, 12.451415061950684, 10.928651809692383, 12.139447212219238, 15.426178932189941, 12.20697021484375, 10.523116111755371, 12.132685661315918, 10.362631797790527, 9.812178611755371, 10.042710304260254, 9.477104187011719, 9.794739723205566, 9.208287239074707, 9.648468017578125, 11.064179420471191, 8.870158195495605, 8.055410385131836, 8.38062572479248, 7.614809513092041, 8.114533424377441, 7.436225891113281, 7.744305610656738, 8.997642517089844, 10.93917179107666, 7.903730392456055, 6.137599468231201, 5.561269283294678, 5.646867275238037, 5.077147483825684, 5.13960599899292, 4.572854518890381, 4.613318920135498, 4.045536518096924, 4.490211009979248, 3.6949615478515625, 4.19376277923584, 3.3772928714752197, 3.421656608581543, 2.8677380084991455, 2.8865480422973633, 2.3476696014404297, 2.3416287899017334, 1.814570426940918, 1.7843564748764038, 1.2657872438430786, 1.2121046781539917, 0.698505699634552, 1.2680789232254028, 0.3474351763725281, 0.999905526638031, 0.0529327392578125]} +{"type": "SampleBatch", "eps_id": [896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 896829513, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAKTQE73y4jm+nZ+OvB9/Qj6ZryK9/rNiPBECX7yFkN29a40hvftuOb6tOoG8/X44PhhjML3QlGk8b2tHvDuH8L0dOC+9gpRWPnjnbbxk8NO+gw0evRXTbzzIxbq89OQAvonaHL3jHlc+S2XPvEPw2b7dpAu9JRN7PGmRCr2AcxC+fWMKvaqZN77FHxa9NCIQPp8TGb2gxoU86pcKvfotJ74ovRe9EZI2vsD3F72TzvI9NFgmvVQQjjxoQQ69tQw+voXsJL1ZhTW+oHUdvS54xD0RcjO9zF6+vseZFb30osA+oudRva5sNL7Fju28KAOUPbpWYL1V5b2+erfhvN8Ztj7cuH690pgzvsJxp7yE6149gYuGvcIDpDwOh568N4J6vpG5hb2+AzO+5JvGvLKJKz2t4oy9oxmpPFi/v7zeRoS+OgqMvTtRMr52E+q8IQHcPDIsk71w2Ly+Cq3lvDvrnj7AR6K9HX0xvmfSsrxuaBM8PGGpvax7vL4KWbG8s+SWPl51uL2wHBC/2w+BvA6bFD8zhM+94zW8vjXRh7uO1pA+wJLevUuOML5eTcY6Wd41vK6i5b0DMby+CDSpOhtqkD7YsPS9zpgwvr4m4zvwYCe8MsH7vQnYuDzJdNw71vWZvpjU+r0DzDC+lhg7OhKMwbt/8wC+X068vuUgHDra8ZI+wXsIvkEbEL/2ms87gHIUPw0DFL4yW7y+V+iSPPIQlD7Sixu+kSwxvuxJwjxRYBE79BYfvgIysjz3psI8vtaQvuikHr5+3jG+xE2UPGvYjDyZMyK+rve8vuUelzx3jaE+oMIpvnNnMr5J0co8GU7rPA1ULb4eSL2+DIbPPKaBqD5M5jS+FCMzvgy5Aj08YTY9e3s4vkF6oTzVXgY9/qBzviIUOL6Mfls+oMLlPLYjBb9SsDO++1+aPAyNkDz271++hk0zvsbWWj4mcVk8FXkBvxHtLr6SOdE+x91OO1RSS7+bjia+xplaPvqIULwYJAC/Xi8ivuTalzwcR7q84v1Yvi7OIb6PezS+F//cvLGQlj1BaiW+xqidPITz0LyNA2m+WgUlvjjgWz7DO/a8NTYHv5afIL6vIqQ8bWImvc39er6KNiC+o6syvrt2Or3sRA09VckjvojjvL5voze98PufPo5XK75ZWTG+fAoevYROxDuV4y6+d0i8vtmMHb33j5I+mWs2vsU4ML6nGQa9HPmVvNvxOb4DYsA8lZkHvR9spL67djm+NkIvvlPoIb3dDSC9Dvg8vv86u77NGyW9zah2Pkt1RL7NaQ+/N2ARvahEBT9m7k++gq26vsZ1zbz5IV4++mVXvmAxLb4366m8KSKrvbrcWr7yV7q+B5y3vL5dTz7iUGK+uo8svk9ulrzr/Ma9ZsRlvkcQ3DyVWaa8qonKvo83Zb73/iu+einnvBH9370tqGi+DLC5vsMU+bw0djI+nhVwvoMjK774hty8IegCvtmBc77oRLm+6XjxvBv7Hz4B63q+g3kOvxjg17xhAuE+cSiDvnDcuL5V34+8+ewNPu7ahr4Oqim+HFRyvORmI75GjYi+5J+4vv9Ok7xzfQM+jT6Mvm8vKb5Viny8dvgtviVlp7zjW4Y85cE2vWOcoLw6taS8lTlZPg9dOL3c7Ka+svOBvNLWkDxQElO9gRNEvT0cfrxMoFo+Of5WvUx0tr5iJji8MP6cPI0vdL0BOaW9xt4xvI5BM75vy3q98T1FPm87a7ytG6s86wNrvWgq871JY2S8Tocxvu+9dL2gCR8+NpmOvCs0vL7YBGi9R2zbPt7Syrx81C++R+lEvfjV8j3d9Oa8aSnFPKQyO705akG+ZQPjvOx1Lr7Hq0q9A1u2PU7t/rxzRNA8dWBDveQVYL75wvq8ioBiPrpNVb1Iswa/c4XWvKri2zwwNIC9EC2AvqIf0rxpWSu+PHWKvTPcszwaiu286QbrPAOPib0lIZW+w9bovBxiKb4vfZW9IvWnvFr4Ab13Bbi+K1SWveI+fj7daR+9TT8nvrIojL0Rboi9F8ssvcT9tr434469UppQPmkSSr0MOyW+H4uGvUeP4b1WSle9e+MNPf0Ni71hedi+3XNUvapII75hX5y9NO0bvuuDYb3vMxY9EZyivfCh777hgl693AQhvr7Htb08Ok6+jGRrvVe1s76DB769QQt+PbMShL02Yh6+KX27vV9RhL6NaIq9QIQqPQcTxr2KGhS/B7SIvUapG75bxd29aLiivv/tjr1s3bC+3snqvU51er0xFJ29UOsJvwpL7b0Y60U+XSWzvclmO79bYOW9PQLlPlQh0b3B52y/Qg7TvQHiMz/4CPe9SKE6v0JGtr0+QcI+q3IKvtBiCL/ru6a9U+t1PdpbFb64Uqy+XUakvfysg75xQBy+o8QPvhXPrr2i9BK/iCAfvsEfq75iUsa9D22evtj4Jb7bIAe/8/7SvRzSR71FyDC+a6o4v33+1L1Au1Y+Oo4/vk5gBr+kZ8y9BkLpvUBOSr6A7Te/6xHRvfszFT4XBVm+oKQFvxUay705nTW+GrZjvuS7pr7PXdK9Pcz/vnVhar556QS/jNTmvQnDdr5+A3W+omg2v2Sz8L0KlWc8nc2BvqMUBL8tH/C9TFugvh4Wh74CjzW/SPP8vc1mdL1HWY6+lQFnv/Nk/71RPEQ+yJaXvvWoNL9/i/e9B/EMvr3Qnr7MUgK/vS79vWmH7r5AB6S+7MYzv+ghCL5LuVu+KzirvhUvZb/khgy+SckIPQNjtL72SIu/ztcLvlxFjz6Rh7++Gvujv7YcBr5myQY/5aXMvrPNir+OqPa9LkpIPpjA177ciKO/l6Xuvfpe5T7I1eS+1Ua8vxJM3L1tqDM/sOXzvgkio79Ijb+9KSrBPlN5AL937Lu/RRmwvaqZIz+q/Qe/GM+ivzTslb02JKQ+1IAOv1Kku7+Wyoi9gN0WP0gCFr8bjqK/S05hvat/jT7Yghy/G3yJv4KqSr2YxQe9sAIiv4Jdor+pYU29GF55Pk+BKL+qPru/nW45vUQdBT+y/i+/BDGiv+PVDr07gFo+iXs2v4Enib/6tfq8WoS4vf/3O791EqK/cbwEvRtnRT6dc0K/BAqJv0Xj6bzwLeG95e5Hv/71ob/x5vu8b8ExPmBpTr9E7oi/E3bfvG+4A76M41O/4Nqhv1iJ9LxYDB8+8Vxav4vTiL+6Ftu8ryUWvgvWX79rwKG/whzzvInMDD5hTma/2Ku6v6KV3LyIYtc+5MVtv++lob9NqZe88/30PSs9dL/Xlrq/3g+EvPwU0D7Ys3u/t5Whv5XzArwqmN49PBWBvzyYiL86rL67NPs+vprQg79Aj6G/TXMcvF+u1T3JC4e/eoS6v9eF9LtCuck+wcaKv0eHob+A79o5Sa7KPcgBjr87gbq/KRUdO1+XyD6vvJG/GXzTvyumJzxrai8/ffeVvwyHur8xF8Q81aLKPoKymb+vmKG/kncCPcPH4j3i7Zy/6aC6v82JCz0Gm9M+bKmgv4G3ob8qZS09p7IGPmnlo79u0oi/xis4PRgIF77woaa/YN1fv6YWLD3G79m+CN+ovxj7iL/3Nwk9B9P1vWCcq78KBKK/csX+PBxsOz7l2a6/TxmJvyJhDj1TJsy915exv9pgYL+kNgY9hSzDvkDWs7/qOIm/ovjNPBt/oL3UlLa/Lj6iv6ohwTzqeGM+g9O5v7lPib/uhuU8/QyBvYySvL9EVqK/+DPbPPoYdD620b+/NVy7vxohAT2xFAo//5DDv21xor+vUC09XnqDPrTQxr8dfru/BlpCPQ8BED+rkMq/MZuiv9hucD1i/5E+NtHNv2m+ib905YM9uiQ9PHWS0L9m1KK/gV6EPULNpT4m1NO/qfqJvyCikT1821U9mpbWvxlJYr+ZxZM9wm1evuTZ2L/JnTC/7t+KPYD7+L4Hntq/f8tiv4rpbT29LzG+oOLcv3sRMb/CvF89fcjkvuun3r+8NGO/zSE7PZPADL6R7eC/j6iKvzLfLz0+WS0+f7Pjv4yHY79gvT09+Frgvfn55b8/0oq/+sM0PXMeSj69wOi/v9+jv1/vRD1t7wA/xgfsvzz9ir+/MW49CPBnPmbP7r9zQGS/6l+APZuuQb25F/G/gzWLvy3gfD0nbIc+eeDzv6m0ZL+LRYk9LJUEvPUp9r9icYu/sPCIPRYknD7o8/i/qzFlv3RulT2cpgs9pD77v26HM7/205Y9Y1Zwvj0K/b+q3QG/6DaNPdjWAL+yVv6/HFygvk0zcT1tAUa/lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 1, 1, 0], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 0, 0, 0, 1, 0, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 1, 1], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAJmvIr3+s2I8EQJfvIWQ3b1rjSG9+245vq06gbz9fjg+GGMwvdCUaTxva0e8O4fwvR04L72ClFY+eOdtvGTw076DDR69FdNvPMjFurz05AC+idocveMeVz5LZc+8Q/DZvt2kC70lE3s8aZEKvYBzEL59Ywq9qpk3vsUfFr00IhA+nxMZvaDGhTzqlwq9+i0nvii9F70Rkja+wPcXvZPO8j00WCa9VBCOPGhBDr21DD6+hewkvVmFNb6gdR29LnjEPRFyM73MXr6+x5kVvfSiwD6i51G9rmw0vsWO7bwoA5Q9ulZgvVXlvb56t+G83xm2Pty4fr3SmDO+wnGnvITrXj2Bi4a9wgOkPA6Hnrw3gnq+kbmFvb4DM77km8a8sokrPa3ijL2jGak8WL+/vN5GhL46Coy9O1EyvnYT6rwhAdw8MiyTvXDYvL4KreW8O+uePsBHor0dfTG+Z9KyvG5oEzw8Yam9rHu8vgpZsbyz5JY+XnW4vbAcEL/bD4G8DpsUPzOEz73jNby+NdGHu47WkD7Akt69S44wvl5NxjpZ3jW8rqLlvQMxvL4INKk6G2qQPtiw9L3OmDC+vibjO/BgJ7wywfu9Cdi4PMl03DvW9Zm+mNT6vQPMML6WGDs6EozBu3/zAL5fTry+5SAcOtrxkj7Bewi+QRsQv/aazzuAchQ/DQMUvjJbvL5X6JI88hCUPtKLG76RLDG+7EnCPFFgETv0Fh++AjKyPPemwjy+1pC+6KQevn7eMb7ETZQ8a9iMPJkzIr6u97y+5R6XPHeNoT6gwim+c2cyvknRyjwZTus8DVQtvh5Ivb4Mhs88poGoPkzmNL4UIzO+DLkCPTxhNj17ezi+QXqhPNVeBj3+oHO+IhQ4vox+Wz6gwuU8tiMFv1KwM777X5o8DI2QPPbvX76GTTO+xtZaPiZxWTwVeQG/Ee0uvpI50T7H3U47VFJLv5uOJr7GmVo++ohQvBgkAL9eLyK+5NqXPBxHurzi/Vi+Ls4hvo97NL4X/9y8sZCWPUFqJb7GqJ08hPPQvI0Dab5aBSW+OOBbPsM79rw1Nge/lp8gvq8ipDxtYia9zf16voo2IL6jqzK+u3Y6vexEDT1VySO+iOO8vm+jN73w+58+jlcrvllZMb58Ch69hE7EO5XjLr53SLy+2YwdvfePkj6Zaza+xTgwvqcZBr0c+ZW82/E5vgNiwDyVmQe9H2ykvrt2Ob42Qi++U+ghvd0NIL0O+Dy+/zq7vs0bJb3NqHY+S3VEvs1pD783YBG9qEQFP2buT76Crbq+xnXNvPkhXj76ZVe+YDEtvjfrqbwpIqu9utxavvJXur4HnLe8vl1PPuJQYr66jyy+T26WvOv8xr1mxGW+RxDcPJVZpryqicq+jzdlvvf+K756Kee8Ef3fvS2oaL4MsLm+wxT5vDR2Mj6eFXC+gyMrvviG3Lwh6AK+2YFzvuhEub7pePG8G/sfPgHrer6DeQ6/GODXvGEC4T5xKIO+cNy4vlXfj7z57A0+7tqGvg6qKb4cVHK85GYjvkaNiL7kn7i+/06TvHN9Az6NPoy+by8pvlWKfLx2+C2+qu+Nvme89jwAG5q83E7vvjq1pLyVOVk+D104vdzspr6y84G80taQPFASU72BE0S9PRx+vEygWj45/la9THS2vmImOLww/pw8jS90vQE5pb3G3jG8jkEzvm/Ler3xPUU+bztrvK0bqzzrA2u9aCrzvUljZLxOhzG+7710vaAJHz42mY68KzS8vtgEaL1HbNs+3tLKvHzUL75H6US9+NXyPd305rxpKcU8pDI7vTlqQb5lA+O87HUuvserSr0DW7Y9Tu3+vHNE0Dx1YEO95BVgvvnC+ryKgGI+uk1VvUizBr9zhda8quLbPDA0gL0QLYC+oh/SvGlZK748dYq9M9yzPBqK7bzpBus8A4+JvSUhlb7D1ui8HGIpvi99lb0i9ae8WvgBvXcFuL4rVJa94j5+Pt1pH71NPye+siiMvRFuiL0Xyyy9xP22vjfjjr1SmlA+aRJKvQw7Jb4fi4a9R4/hvVZKV7174w09/Q2LvWF52L7dc1S9qkgjvmFfnL007Ru+64Nhve8zFj0RnKK98KHvvuGCXr3cBCG+vse1vTw6Tr6MZGu9V7WzvoMHvr1BC349sxKEvTZiHr4pfbu9X1GEvo1oir1AhCo9BxPGvYoaFL8HtIi9RqkbvlvF3b1ouKK+/+2OvWzdsL7eyeq9TnV6vTEUnb1Q6wm/CkvtvRjrRT5dJbO9yWY7v1tg5b09AuU+VCHRvcHnbL9CDtO9AeIzP/gI971IoTq/Qka2vT5Bwj6rcgq+0GIIv+u7pr1T63U92lsVvrhSrL5dRqS9/KyDvnFAHL6jxA++Fc+uvaL0Er+IIB++wR+rvmJSxr0PbZ6+2PglvtsgB7/z/tK9HNJHvUXIML5rqji/ff7UvUC7Vj46jj++TmAGv6RnzL0GQum9QE5KvoDtN7/rEdG9+zMVPhcFWb6gpAW/FRrLvTmdNb4atmO+5Lumvs9d0r09zP++dWFqvnnpBL+M1Oa9CcN2vn4Ddb6iaDa/ZLPwvQqVZzydzYG+oxQEvy0f8L1MW6C+HhaHvgKPNb9I8/y9zWZ0vUdZjr6VAWe/82T/vVE8RD7Ilpe+9ag0v3+L970H8Qy+vdCevsxSAr+9Lv29aYfuvkAHpL7sxjO/6CEIvku5W74rOKu+FS9lv+SGDL5JyQg9A2O0vvZIi7/O1wu+XEWPPpGHv74a+6O/thwGvmbJBj/lpcy+s82Kv46o9r0uSkg+mMDXvtyIo7+Xpe69+l7lPsjV5L7VRry/EkzcvW2oMz+w5fO+CSKjv0iNv70pKsE+U3kAv3fsu79FGbC9qpkjP6r9B78Yz6K/NOyVvTYkpD7UgA6/UqS7v5bKiL2A3RY/SAIWvxuOor9LTmG9q3+NPtiCHL8bfIm/gqpKvZjFB72wAiK/gl2iv6lhTb0YXnk+T4Eov6o+u7+dbjm9RB0FP7L+L78EMaK/49UOvTuAWj6Jeza/gSeJv/q1+rxahLi9//c7v3USor9xvAS9G2dFPp1zQr8ECom/RePpvPAt4b3l7ke//vWhv/Hm+7xvwTE+YGlOv0TuiL8Tdt+8b7gDvozjU7/g2qG/WIn0vFgMHz7xXFq/i9OIv7oW27yvJRa+C9Zfv2vAob/CHPO8icwMPmFOZr/Yq7q/opXcvIhi1z7kxW2/76Whv02pl7zz/fQ9Kz10v9eWur/eD4S8/BTQPtize7+3laG/lfMCvCqY3j08FYG/PJiIvzqsvrs0+z6+mtCDv0CPob9Ncxy8X67VPckLh796hLq/14X0u0K5yT7Bxoq/R4ehv4Dv2jlJrso9yAGOvzuBur8pFR07X5fIPq+8kb8ZfNO/K6YnPGtqLz9995W/DIe6vzEXxDzVoso+grKZv6+Yob+SdwI9w8fiPeLtnL/poLq/zYkLPQab0z5sqaC/gbehvyplLT2nsgY+aeWjv27SiL/GKzg9GAgXvvChpr9g3V+/phYsPcbv2b4I36i/GPuIv/c3CT0H0/W9YJyrvwoEor9yxf48HGw7PuXZrr9PGYm/ImEOPVMmzL3Xl7G/2mBgv6Q2Bj2FLMO+QNazv+o4ib+i+M08G3+gvdSUtr8uPqK/qiHBPOp4Yz6D07m/uU+Jv+6G5Tz9DIG9jJK8v0RWor/4M9s8+hh0PrbRv781XLu/GiEBPbEUCj//kMO/bXGiv69QLT1eeoM+tNDGvx1+u78GWkI9DwEQP6uQyr8xm6K/2G5wPWL/kT420c2/ab6Jv3Tlgz26JD08dZLQv2bUor+BXoQ9Qs2lPibU07+p+om/IKKRPXzbVT2alta/GUliv5nFkz3CbV6+5NnYv8mdML/u34o9gPv4vgee2r9/y2K/iultPb0vMb6g4ty/exExv8K8Xz19yOS+66fev7w0Y7/NITs9k8AMvpHt4L+PqIq/Mt8vPT5ZLT5/s+O/jIdjv2C9PT34WuC9+fnlvz/Sir/6wzQ9cx5KPr3A6L+/36O/X+9EPW3vAD/GB+y/PP2Kv78xbj0I8Gc+Zs/uv3NAZL/qX4A9m65BvbkX8b+DNYu/LeB8PSdshz554PO/qbRkv4tFiT0slQS89Sn2v2Jxi7+w8Ig9FiScPujz+L+rMWW/dG6VPZymCz2kPvu/boczv/bTlj1jVnC+PQr9v6rdAb/oNo092NYAv7JW/r8cXKC+TTNxPW0BRr/0I/+/SFMCv6/WMT0G+uy+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.7106685638427734, 0.6668575406074524, 0.7000078558921814, 0.32227280735969543, 0.8854774832725525, 0.291730672121048, 0.8958092927932739, 0.7450428009033203, 0.5966638326644897, 0.7699317336082458, 0.5571064352989197, 0.7931985855102539, 0.48584210872650146, 0.8177021145820618, 0.5065338611602783, 0.8101471066474915, 0.48160436749458313, 0.8255948424339294, 0.4523012936115265, 0.838689386844635, 0.5816727876663208, 0.7651889324188232, 0.5962967872619629, 0.2425028383731842, 0.9109862446784973, 0.77065509557724, 0.5777051448822021, 0.7786453366279602, 0.43751171231269836, 0.8378877639770508, 0.5662569403648376, 0.21439701318740845, 0.9209125638008118, 0.8097880482673645, 0.5025780200958252, 0.8031399250030518, 0.4809136688709259, 0.8341689109802246, 0.4441937804222107, 0.8504447937011719, 0.5996227860450745, 0.2603808343410492, 0.8962920308113098, 0.261137455701828, 0.10159697383642197, 0.9471202492713928, 0.9088575839996338, 0.7872995138168335, 0.5310128927230835, 0.19528867304325104, 0.9223884344100952, 0.8343828320503235, 0.5757730007171631, 0.7634307742118835, 0.6077878475189209, 0.7426372766494751, 0.3664761483669281, 0.8727752566337585, 0.6759410500526428, 0.318977028131485, 0.8840462565422058, 0.6751586198806763, 0.7008159160614014, 0.6605274081230164, 0.28753361105918884, 0.8947510719299316, 0.7407499551773071, 0.605197548866272, 0.7575953602790833, 0.4200104773044586, 0.8475610017776489, 0.5779494643211365, 0.7699623703956604, 0.5632368326187134, 0.22127433121204376, 0.422776460647583, 0.8589866757392883, 0.35968849062919617, 0.880551278591156, 0.7051267623901367, 0.6363570690155029, 0.7504859566688538, 0.4244270622730255, 0.839689314365387, 0.5360040068626404, 0.8068986535072327, 0.48021307587623596, 0.16773734986782074, 0.9297579526901245, 0.8628906011581421, 0.3338986337184906, 0.8841833472251892, 0.7301704287528992, 0.590936541557312, 0.7732299566268921, 0.5246951580047607, 0.19215992093086243, 0.9209063649177551, 0.15553231537342072, 0.9304401874542236, 0.8749923706054688, 0.2799084186553955, 0.10450006276369095, 0.9438902139663696, 0.913950502872467, 0.8364750742912292, 0.6517844796180725, 0.37332117557525635, 0.831131100654602, 0.5788437724113464, 0.25698432326316833, 0.10453684628009796, 0.9427295327186584, 0.9109273552894592, 0.8311253190040588, 0.34625235199928284, 0.8574703931808472, 0.28901779651641846, 0.12165766209363937, 0.936893880367279, 0.8978785276412964, 0.18978732824325562, 0.9121976494789124, 0.844907820224762, 0.2898705303668976, 0.13065455853939056, 0.9321234822273254, 0.8911815285682678, 0.8097609281539917, 0.6649650931358337, 0.538287878036499, 0.7086115479469299, 0.51210618019104, 0.6975406408309937, 0.5411106944084167, 0.6807903051376343, 0.5588160753250122, 0.6732602119445801, 0.4336960017681122, 0.7775503993034363, 0.5796176791191101, 0.6648098826408386, 0.4240368902683258, 0.784044623374939, 0.4207342863082886, 0.7867113351821899, 0.4180883765220642, 0.7887042164802551, 0.4156350791454315, 0.7902842164039612, 0.5870823264122009, 0.6664329767227173, 0.5750594735145569, 0.6797168254852295, 0.44329673051834106, 0.7674208283424377, 0.5479292869567871, 0.7015162706375122, 0.525517463684082, 0.28095874190330505, 0.8653427958488464, 0.7485175132751465, 0.449240505695343, 0.7724223732948303, 0.5910126566886902, 0.3788011968135834, 0.7832255363464355, 0.5994870662689209, 0.635342538356781, 0.4285106658935547, 0.7436121702194214, 0.5554605722427368, 0.6636852025985718, 0.5333572626113892, 0.3222014307975769, 0.830599308013916, 0.3011248707771301, 0.8411584496498108, 0.7221211194992065, 0.4397510886192322, 0.7416161894798279, 0.590021550655365, 0.4335390627384186, 0.6965445280075073, 0.45166945457458496, 0.6808656454086304, 0.5349255800247192, 0.6335983872413635, 0.5182574391365051, 0.35444557666778564, 0.7886951565742493, 0.6626697778701782, 0.4773338735103607, 0.6776759624481201, 0.4576045274734497, 0.6930818557739258, 0.56284099817276, 0.4418734610080719, 0.34181904792785645, 0.7381816506385803], "advantages": [-14.484781265258789, -15.078315734863281, -15.304299354553223, -15.869322776794434, -15.188665390014648, -16.362459182739258, -15.527762413024902, -16.737367630004883, -16.96587562561035, -17.3212833404541, -17.5750675201416, -17.840639114379883, -18.11956214904785, -17.680397033691406, -18.96535301208496, -18.61017608642578, -19.816015243530273, -19.9058837890625, -20.327102661132812, -20.334800720214844, -20.787870407104492, -20.59699249267578, -21.577484130859375, -21.455488204956055, -21.033933639526367, -22.681228637695312, -23.63094711303711, -23.64626121520996, -24.598915100097656, -24.57905387878418, -25.287006378173828, -25.391054153442383, -25.24933624267578, -26.800457000732422, -27.81361961364746, -27.914308547973633, -28.73761558532715, -28.953563690185547, -30.030853271484375, -30.27353858947754, -31.412626266479492, -31.670278549194336, -30.77577018737793, -32.50199508666992, -31.50899314880371, -28.935720443725586, -31.675037384033203, -33.52141571044922, -34.59233474731445, -34.51129150390625, -33.061500549316406, -34.97571563720703, -36.23207473754883, -37.01788330078125, -37.580421447753906, -38.42612075805664, -38.92631912231445, -38.43696594238281, -39.8529052734375, -40.92127990722656, -41.733375549316406, -42.64072036743164, -42.90554428100586, -44.06563949584961, -44.26015090942383, -43.642452239990234, -45.19123840332031, -46.599796295166016, -46.51935577392578, -48.01278305053711, -49.343505859375, -49.771995544433594, -49.62903594970703, -51.206783294677734, -50.992889404296875, 8.386072158813477, 9.242635726928711, 8.318897247314453, 9.362587928771973, 8.40546703338623, 8.517806053161621, 8.303901672363281, 8.395092964172363, 9.29047966003418, 8.003610610961914, 8.021709442138672, 8.034357070922852, 8.190627098083496, 10.05794906616211, 8.954859733581543, 8.935110092163086, 9.549631118774414, 9.4678955078125, 10.095630645751953, 9.704157829284668, 10.212653160095215, 10.010623931884766, 11.269890785217285, 10.831707000732422, 12.433157920837402, 11.796772956848145, 11.732799530029297, 12.439088821411133, 14.711362838745117, 13.701303482055664, 13.11536693572998, 12.889432907104492, 12.967811584472656, 13.286262512207031, 12.04516887664795, 11.949445724487305, 13.052851676940918, 15.059213638305664, 14.138010025024414, 13.177241325378418, 12.223518371582031, 13.554078102111816, 12.447104454040527, 14.100028991699219, 16.052907943725586, 15.255127906799316, 14.207146644592285, 16.08226776123047, 15.220738410949707, 13.967734336853027, 15.876638412475586, 17.559951782226562, 17.040679931640625, 16.304439544677734, 15.002102851867676, 13.089179039001465, 15.41972541809082, 13.686527252197266, 11.47396183013916, 13.960506439208984, 11.804299354553223, 14.220174789428711, 12.197535514831543, 14.398192405700684, 15.626882553100586, 14.757031440734863, 13.119203567504883, 14.638959884643555, 15.24930477142334, 14.611503601074219, 14.963932037353516, 14.400782585144043, 14.551420211791992, 14.040635108947754, 14.042128562927246, 13.567328453063965, 12.871500015258789, 12.814275741577148, 12.241110801696777, 12.012792587280273, 11.757238388061523, 11.36470890045166, 10.993191719055176, 10.500683784484863, 10.238116264343262, 9.84238052368164, 9.327629089355469, 8.593448638916016, 8.625150680541992, 7.771075248718262, 7.172928810119629, 7.142759323120117, 6.691532135009766, 6.794996738433838, 6.0212016105651855, 5.939462184906006, 5.643169403076172, 6.004397392272949, 5.040994167327881, 5.55625057220459, 6.211331844329834, 4.9173583984375, 5.666738510131836, 4.396872043609619, 3.1750826835632324, 4.2400221824646, 2.9801526069641113, 2.3265228271484375, 2.5976719856262207, 2.6909706592559814, 2.7249648571014404, 3.0128118991851807, 3.991132974624634, 2.8632235527038574, 3.8758416175842285, 4.41290807723999, 3.267967700958252, 2.0713136196136475, 3.038623809814453, 1.8079873323440552, 2.6710569858551025, 1.430898904800415, 0.41751307249069214, -0.09992370754480362, -0.17169952392578125]} +{"type": "SampleBatch", "eps_id": [90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 90188607, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAPQj/79IUwK/r9YxPQb67L7LOADAMG00vx/sCz35jyC+vR8BwKJ/Zr+bJ/48GvoVPsdGAsAeqjS/VRMLPXSLC74HLgPAZrxmv+nS/zw97io+X1UEwBLnNL8dlg09/wvtve08BcAcFQO/wRoEPUFuy7625AXAEYGivngcxzy6/i2/t0wGwIUd+70/gi88E/B2v+R0BsANwaK+VJIMvIcnK78O3QbAGyr7vRrTs7wGxXa/PgUHwAKKor7r4Ci98KAtv0RtB8BwEQO/lHBgvXIxzL4JFQjAUM80vyqOgL19vv29ePwIwGGEZr9WoYW9LOYXPogjCsBwG4y/xht/vb/o1j41igvAjPakvxi5XL1vMTE/gzANwIDWvb9wBSS9yqd3P38WD8Dly6S/AoupvAqUKT9gvBDA6MiLvzYN9LvoALo+OSISwGuXZb9S8j65G8+LPRpIE8BhxYu/UxabOsjGuD7qrRTAWphlv5KkCTy2dIw9zNMVwNjJi7+lHSA8k1G6Pqg5F8BeqWW/DK6LPBw0mD2fXxjAJcYzvy3blzwbDFq+vEUZwFLkAb/q72k8qdr/vkdyDL2+2Am9F/uHvLPZuLsNNA+9OZUlPrPniLxLZpu+6vQBvW3lB70Jorq820qEvLSsBL1eKCY+X0e9vNjBob6Ww+68xjUFvYUK8bz03/q8qBf0vLapaL4AD/a88ViBPsSoDL24ugG93KrMvDc6Sr37QA+9htxnvqvB1LwB+XA+gM0hvTZ+/bxtM668KwOGvXJWJL16bCg+AOy4vK3Aur4g3Ra9zlz4vMeu9LyxWKK971gZvSRmZr7R1QC9sbtQPoHHK722L/G87EXgvJ3syb3xMC69FQwqPlVt8Lwiscy+YJYgvWqN6rzZ9hi9a5juvdXuIr1ijmS+EIIivQAcKD6pNzW96+DVvi8PFb0EXuM+IXBXveVxY75xXOG8+IIPPjOiab3xatW+NWbKvAYa2T7l44W9aa5ivj3thLykQP09HPWOvUYf1b5BVWG89YXSPtoBoL0ydhy/PjK1uwYSMz+GCrm9YPvUvqucCjwGZs8+ZRTKvWtIYr5srIc8oarrPYgh073SKdW+34aaPPtr0z4eL+S9/8xivowu3jxWSQE+jkHtvfw727wf3vI8uYghvi1a7r11oWO+rwXZPP6YEz4ddfe9gMvhvEij8Dypbg++Ipb4vSojKz5MsNk8ecDYvq+98b3/Vui8/FOUPH+u+r0V5/K99n0qPgdGgDxGktG+PhXsvdlK7LzB1/Q7e9jkvbJD7b1rKCo+uJyrO3Pazb5Idea9AsjtvJbCN7thoNy9pKXnvX18Zb4Ee6K7o3w8PpTT8L3sm9a+T2WnukpL8z5j/wC+E2VlvqTIBjwPeDo+45UFvt2o1r4udEI8Omn0PgEsDr4hsGW+PnCvPCX4QD4BxBK+taDxvERQzjzoe8e9pl4TvgI5KT7VWr48Q5rDvjr8D75KV7g+JIZ/PGDUKr+TnAi+magoPspyEzv3T72+Cz0FvoNu+bx8mKi7OmKcva7cBb5u6Ga+b6Pau0LZWz7uegq+p07XvtzeG7soWQE/rRcTvn7GZr5BMv07U+tYPj61F74Brfi8JwNEPFiQoL1lVBi+GBBnvnVSKjwNRF8+cPMcvjdL+7xlxHE8UiGSvUSUHb5iLCg+52JaPOL7t744Nxq+DeS3PvRFyTvS1SW/LdwSvsPjJz7SQ9+7adS0vpWAD765hf68Dl1jvP9RgL16IxC+5Wlnvg3ld7zoB2c+UMQUvjR/174B9y2893MDPwBjHb6VCWe+qI+2uaGyXj7qASK+1tH5vNMdgzt4Ppq9zKEivkEXZ75PhCM7FOBfPvxAJ76klPq84gnhO84Mlr1b4Se+QzpnvskFsTuY5GI+PYEsvoIE/Lz3HSE89CCOvYgiLb41c2e+X2AKPKjNZz6OwzG+ULDXvrONVDx+jgU/NWQ6vhPDZ77EwL886LtuPtQGP77bOgG9T/PlPDVHVb0+rD++8BInPldr3TyE6au+01Q8vqN0BL1MaKY8aRAOvV7+PL4CYyY+kLmgPIxLpL54qjm+TcsGvQNNWDwJ6bS8AVc6vvLoJT5+EFE8VAOfvowFN76VUAi9mpfWO8KdY7wHtDe+p/ppvtB8zTt9y48+AGI8vs0QCb3CxUI8UV4hvHIRPb6iXCU+jYs/PMX2mL7Lwjm+WXIKveFLuzvI/B27AXQ6vtobJT5ut7k7aymWvqYmN74KHQu9u7JPuVezmjq22De+iQclPjPyNrkHSJW+w4s0vnQXC70VzMS7BUyLOs09Nb4zfWq+yBnEu61slT5i7jm+qmMKvXvLGrlCOzK7hp86vrA1JT4z1FO5G0WXvqZRN75sXgq9xD7Iu6p1ObvDAji+i05qvosZyrsUapM+abI8vlGmCb3LjNa5I6Xbu5piPb4CZiU+D2sOuglamb7DEzq+CpgJvasX1ru4feW74sM6vnGAJT6qrtq7wX2avoN0N74H0gi9LTdQvBkCN7ylIzi+LtJpvi3gU7xiEY4+z9A8vplRB72L5/G77MCdvAR+Pb5+hmm+Vob+u/3Lij6qKUK+dW4Gvce6GbtC5cS8vdVCvjs7Jj6eOzm7wImivqKCP77vIAa9JVUWvIxD0rxRLkC+/mcmPj++HrxaeKS+Udo8vu4stz6EAIS8p/Adv5qGNb7IxyY+ahXpvKKlqL6vMDK+gWQCvX2GD73QpTu9ltcyvu2zJz4/RxO969yyvvN8L77bevy8eeUvveyxi72JHjC+VbVmvvN7Nb3arFc+w7s0vlpG8rzsOiS9OgDEvdJWNb5demW++BEsvXB7PD6/7Tm+OVLWvtv9HL0vKe0+ZYBCvg9OZL5+F+68DYAiPlERR7741dW+fBfUvHFW4j7/nk++Z4Bjvuepi7w1uhA+zStUvo+b27y9A2m89W8gvlq4VL44jyw+Yi2OvB9a6L7ZRFG+EevXvKmH2Lxspiq+Cc9RvrBlYr591fO80szwPTBWVr5Z4NS+55HgvK8rzT4L2l6+lY9hvlPqnrwt1Ms96lxjvqPQy7zpm468sf5LvlvfY74XlC4+ij+vvPSi/r6EYWC+mEnHvK9dAL0bjFi+D+FgvjI+YL6SsBG9u72RPS5dZb6jer+8L9wLvYEbbr6612W+JVowPp/oHr3/KAm/zlBivmz7trzIzEq9O9qCvurFYr5W612+f7xfvT/LrDwkNme+qnDSviUCXr1clJc+DKFvviZUXL5zwUW9hpNYvCEJdL4lstG+qtZGvTIXhz5qbHy+uulavls5Mb2NKjO9oGaAvi0H0b6vzjS9IZ9wPtmUhL4CTBq/wY4hvYtvAz/ZwIq+K2vQvhX/7rxnlFU+9OuOvi+eWL7b0sy8ic6+vX8Wkb7/BdC+lBbcvLseRD6TP5W+89tXvoG1vLzgR+C9LGiXvsGoz77Gps683wg0PmSPm76lrxm/jtixvBO+6z4itaG+8lDPvg/RTLwq2iQ+mNqlvoSlVr5eEBi81N0KvhcAqL4LT2u8UIBEvFfH3L69Jai+8FVWvljmqLwFvxG+cEqqvoLqzr4cOMC8azoTPtltrr74sFW+pKmovLn2H77mkLC+F5jOvsJBwrwxBQU+qbK0vjoLVb4/+ay8CEAuvg3Utr40RM6+iNrIvOUe7T0j9Lq+Cv4Yv03itbxOH80+xxLBvrbszb5efWi82ujOPRwxxb7S1lO+WmJHvFbNSL5rT8e+WLvNvgfSg7zS5b09xGzLvtXBGL/ZQWm86bTCPgCJ0b6Fgs2+NErZu/dJqj02pdW+PR5TviPMoruYsli+rMHXvspszb7wvRa8Ic2iPXPd275F6VK+MmP5u7lDXb5h+d2+ru4vvJd/Q7zl2gK/iBXevmCgUr4Xf7W8dZZjvrww4L7NDM2+FenZvGzCgT2XSuS+R+pRvpuHz7xyTHO++WPmvoyrzL4edfa86ndAPeJ76r5rG1G+PsLuvIiTgr4zk+y+ozzMvodFDL02+uc85KjwvpPwF7+q8wm9j7iePsG89r5ovcu+6RzhvCzJ4Tvnz/q+srcXv+f737zf4JQ+v3AAv7FWy77JV7C8l2kqvEt5Ar8zihe/CgyyvO8CjT4tgQW/oQXLvmLshLzT+cS86YgHv9VmF7/m3Ii8KOaGPheQCr/qx8q+8GM7vBgHDb01lwy/mUwXvy+sRrzNXYI+3J0Pv8Kbyr7neea7cHYrvYqkEb+8Ohe/gvQAvI6Pfj7VqhS/1X/Kvo77Pbtatj69lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 0, 1, 0, 1, 1, 1, 1, 0, 1, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 1, 1, 0, 0, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [0, 0, 0, 1, 0, 1, 1, 1, 1, 0, 1, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 1, 1, 0, 0, 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 0, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAAMs4AMAwbTS/H+wLPfmPIL69HwHAon9mv5sn/jwa+hU+x0YCwB6qNL9VEws9dIsLvgcuA8BmvGa/6dL/PD3uKj5fVQTAEuc0vx2WDT3/C+297TwFwBwVA7/BGgQ9QW7LvrbkBcARgaK+eBzHPLr+Lb+3TAbAhR37vT+CLzwT8Ha/5HQGwA3Bor5Ukgy8hycrvw7dBsAbKvu9GtOzvAbFdr8+BQfAAoqivuvgKL3woC2/RG0HwHARA7+UcGC9cjHMvgkVCMBQzzS/Ko6AvX2+/b14/AjAYYRmv1ahhb0s5hc+iCMKwHAbjL/GG3+9v+jWPjWKC8CM9qS/GLlcvW8xMT+DMA3AgNa9v3AFJL3Kp3c/fxYPwOXLpL8Ci6m8CpQpP2C8EMDoyIu/Ng30u+gAuj45IhLAa5dlv1LyPrkbz4s9GkgTwGHFi79TFps6yMa4PuqtFMBamGW/kqQJPLZ0jD3M0xXA2MmLv6UdIDyTUbo+qDkXwF6pZb8Mros8HDSYPZ9fGMAlxjO/LduXPBsMWr68RRnAUuQBv+rvaTyp2v++/+sZwNXkM7+MYYw7rW1Pvg00D705lSU+s+eIvEtmm77q9AG9beUHvQmiurzbSoS8tKwEvV4oJj5fR7282MGhvpbD7rzGNQW9hQrxvPTf+ryoF/S8tqlovgAP9rzxWIE+xKgMvbi6Ab3cqsy8NzpKvftAD72G3Ge+q8HUvAH5cD6AzSG9Nn79vG0zrrwrA4a9clYkvXpsKD4A7Li8rcC6viDdFr3OXPi8x670vLFYor3vWBm9JGZmvtHVAL2xu1A+gccrvbYv8bzsReC8nezJvfEwLr0VDCo+VW3wvCKxzL5gliC9ao3qvNn2GL1rmO691e4ivWKOZL4QgiK9ABwoPqk3Nb3r4NW+Lw8VvQRe4z4hcFe95XFjvnFc4bz4gg8+M6JpvfFq1b41Zsq8BhrZPuXjhb1prmK+Pe2EvKRA/T0c9Y69Rh/VvkFVYbz1hdI+2gGgvTJ2HL8+MrW7BhIzP4YKub1g+9S+q5wKPAZmzz5lFMq9a0hivmyshzyhqus9iCHTvdIp1b7fhpo8+2vTPh4v5L3/zGK+jC7ePFZJAT6OQe29/DvbvB/e8jy5iCG+LVruvXWhY76vBdk8/pgTPh11972Ay+G8SKPwPKluD74ilvi9KiMrPkyw2Tx5wNi+r73xvf9W6Lz8U5Q8f676vRXn8r32fSo+B0aAPEaS0b4+Fey92UrsvMHX9Dt72OS9skPtvWsoKj64nKs7c9rNvkh15r0CyO28lsI3u2Gg3L2kpee9fXxlvgR7orujfDw+lNPwveyb1r5PZae6SkvzPmP/AL4TZWW+pMgGPA94Oj7jlQW+3ajWvi50Qjw6afQ+ASwOviGwZb4+cK88JfhAPgHEEr61oPG8RFDOPOh7x72mXhO+AjkpPtVavjxDmsO+OvwPvkpXuD4khn88YNQqv5OcCL6ZqCg+ynITO/dPvb4LPQW+g275vHyYqLs6Ypy9rtwFvm7oZr5vo9q7QtlbPu56Cr6nTte+3N4buyhZAT+tFxO+fsZmvkEy/TtT61g+PrUXvgGt+LwnA0Q8WJCgvWVUGL4YEGe+dVIqPA1EXz5w8xy+N0v7vGXEcTxSIZK9RJQdvmIsKD7nYlo84vu3vjg3Gr4N5Lc+9EXJO9LVJb8t3BK+w+MnPtJD37tp1LS+lYAPvrmF/rwOXWO8/1GAvXojEL7laWe+DeV3vOgHZz5QxBS+NH/XvgH3Lbz3cwM/AGMdvpUJZ76oj7a5obJePuoBIr7W0fm80x2DO3g+mr3MoSK+QRdnvk+EIzsU4F8+/EAnvqSU+rziCeE7zgyWvVvhJ75DOme+yQWxO5jkYj49gSy+ggT8vPcdITz0II69iCItvjVzZ75fYAo8qM1nPo7DMb5QsNe+s41UPH6OBT81ZDq+E8NnvsTAvzzou24+1AY/vts6Ab1P8+U8NUdVvT6sP77wEic+V2vdPITpq77TVDy+o3QEvUxopjxpEA69Xv48vgJjJj6QuaA8jEukvniqOb5Nywa9A01YPAnptLwBVzq+8uglPn4QUTxUA5++jAU3vpVQCL2al9Y7wp1jvAe0N76n+mm+0HzNO33Ljz4AYjy+zRAJvcLFQjxRXiG8chE9vqJcJT6Niz88xfaYvsvCOb5Zcgq94Uu7O8j8HbsBdDq+2hslPm63uTtrKZa+piY3vgodC727sk+5V7OaOrbYN76JByU+M/I2uQdIlb7DizS+dBcLvRXMxLsFTIs6zT01vjN9ar7IGcS7rWyVPmLuOb6qYwq9e8sauUI7MruGnzq+sDUlPjPUU7kbRZe+plE3vmxeCr3EPsi7qnU5u8MCOL6LTmq+ixnKuxRqkz5psjy+UaYJvcuM1rkjpdu7mmI9vgJmJT4Paw66CVqZvsMTOr4KmAm9qxfWu7h95bviwzq+cYAlPqqu2rvBfZq+g3Q3vgfSCL0tN1C8GQI3vKUjOL4u0mm+LeBTvGIRjj7P0Dy+mVEHvYvn8bvswJ28BH49vn6Gab5Whv67/cuKPqopQr51bga9x7oZu0LlxLy91UK+OzsmPp47ObvAiaK+ooI/vu8gBr0lVRa8jEPSvFEuQL7+ZyY+P74evFp4pL5R2jy+7iy3PoQAhLyn8B2/moY1vsjHJj5qFem8oqWovq8wMr6BZAK9fYYPvdClO72W1zK+7bMnPj9HE73r3LK+83wvvtt6/Lx55S+97LGLvYkeML5VtWa+83s1vdqsVz7DuzS+WkbyvOw6JL06AMS90lY1vl16Zb74ESy9cHs8Pr/tOb45Uta+2/0cvS8p7T5lgEK+D05kvn4X7rwNgCI+URFHvvjV1b58F9S8cVbiPv+eT75ngGO+56mLvDW6ED7NK1S+j5vbvL0Dabz1byC+WrhUvjiPLD5iLY68H1rovtlEUb4R69e8qYfYvGymKr4Jz1G+sGVivn3V87zSzPA9MFZWvlng1L7nkeC8ryvNPgvaXr6Vj2G+U+qevC3Uyz3qXGO+o9DLvOmbjryx/ku+W99jvheULj6KP6+89KL+voRhYL6YSce8r10AvRuMWL4P4WC+Mj5gvpKwEb27vZE9Ll1lvqN6v7wv3Au9gRtuvrrXZb4lWjA+n+gevf8oCb/OUGK+bPu2vMjMSr072oK+6sVivlbrXb5/vF+9P8usPCQ2Z76qcNK+JQJevVyUlz4MoW++JlRcvnPBRb2Gk1i8IQl0viWy0b6q1ka9MheHPmpsfL666Vq+WzkxvY0qM72gZoC+LQfRvq/ONL0hn3A+2ZSEvgJMGr/BjiG9i28DP9nAir4ra9C+Ff/uvGeUVT70646+L55YvtvSzLyJzr69fxaRvv8F0L6UFty8ux5EPpM/lb7z21e+gbW8vOBH4L0saJe+wajPvsamzrzfCDQ+ZI+bvqWvGb+O2LG8E77rPiK1ob7yUM++D9FMvCraJD6Y2qW+hKVWvl4QGLzU3Qq+FwCovgtPa7xQgES8V8fcvr0lqL7wVVa+WOaovAW/Eb5wSqq+gurOvhw4wLxrOhM+2W2uvviwVb6kqai8ufYfvuaQsL4XmM6+wkHCvDEFBT6psrS+OgtVvj/5rLwIQC6+DdS2vjREzr6I2si85R7tPSP0ur4K/hi/TeK1vE4fzT7HEsG+tuzNvl59aLza6M49HDHFvtLWU75aYke8Vs1IvmtPx75Yu82+B9KDvNLlvT3EbMu+1cEYv9lBabzptMI+AInRvoWCzb40Stm790mqPTal1b49HlO+I8yiu5iyWL6swde+ymzNvvC9FrwhzaI9c93bvkXpUr4yY/m7uUNdvmH53b6u7i+8l39DvOXaAr+IFd6+YKBSvhd/tbx1lmO+vDDgvs0Mzb4V6dm8bMKBPZdK5L5H6lG+m4fPvHJMc775Y+a+jKvMvh519rzqd0A94nvqvmsbUb4+wu68iJOCvjOT7L6jPMy+h0UMvTb65zzkqPC+k/AXv6rzCb2PuJ4+wbz2vmi9y77pHOG8LMnhO+fP+r6ytxe/5/vfvN/glD6/cAC/sVbLvslXsLyXaSq8S3kCvzOKF78KDLK87wKNPi2BBb+hBcu+YuyEvNP5xLzpiAe/1WYXv+bciLwo5oY+F5AKv+rHyr7wYzu8GAcNvTWXDL+ZTBe/L6xGvM1dgj7cnQ+/wpvKvud55rtwdiu9iqQRv7w6F7+C9AC8jo9+PtWqFL/Vf8q+jvs9u1q2Pr07sRa/vyZNvrUCe7sHKa6+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.6544507145881653, 0.5457584857940674, 0.5860438346862793, 0.5376443266868591, 0.5921902060508728, 0.47045043110847473, 0.3688003420829773, 0.28719666600227356, 0.7795383930206299, 0.27813324332237244, 0.789827823638916, 0.7364574670791626, 0.6673202514648438, 0.5781663656234741, 0.46807172894477844, 0.34549811482429504, 0.23202326893806458, 0.851290225982666, 0.7621870636940002, 0.6445932984352112, 0.4761773943901062, 0.6428894400596619, 0.4743407666683197, 0.6430926322937012, 0.5292347073554993, 0.4302200973033905, 0.651005208492279, 0.47257623076438904, 0.8321378231048584, 0.44006696343421936, 0.846711277961731, 0.5994668006896973, 0.7510499954223633, 0.6216145753860474, 0.7359360456466675, 0.3600485026836395, 0.8741902112960815, 0.6773024797439575, 0.6830852627754211, 0.2998620867729187, 0.8930201530456543, 0.7390009164810181, 0.39635995030403137, 0.8568435311317444, 0.41082262992858887, 0.8541404008865356, 0.4139454662799835, 0.14439328014850616, 0.9384337067604065, 0.8699994087219238, 0.3571036756038666, 0.8817934989929199, 0.6820552945137024, 0.6750481128692627, 0.7081896662712097, 0.3528076708316803, 0.8665525317192078, 0.3597326874732971, 0.8664984703063965, 0.35669490694999695, 0.8697679042816162, 0.6561554074287415, 0.2855561375617981, 0.8997054696083069, 0.2662478983402252, 0.9064598679542542, 0.7613208889961243, 0.4218817949295044, 0.15980525314807892, 0.929334819316864, 0.8506399393081665, 0.6045637130737305, 0.24572443962097168, 0.9101377129554749, 0.7710452079772949, 0.5697740912437439, 0.7813941240310669, 0.4495353698730469, 0.16941535472869873, 0.9273754358291626, 0.8454079627990723, 0.5952923893928528, 0.2426290065050125, 0.9098319411277771, 0.7684805989265442, 0.5773200392723083, 0.7737385630607605, 0.5667344927787781, 0.7807573676109314, 0.5527257919311523, 0.21055147051811218, 0.919940710067749, 0.8117216229438782, 0.5185427665710449, 0.7889009714126587, 0.5330743789672852, 0.7833717465400696, 0.5405272841453552, 0.7816952466964722, 0.4587852954864502, 0.8338839411735535, 0.559673547744751, 0.7698712348937988, 0.5600287318229675, 0.7722243070602417, 0.5542570948600769, 0.7783169746398926, 0.45781058073043823, 0.831761360168457, 0.5481908321380615, 0.781792938709259, 0.46421143412590027, 0.8288088440895081, 0.5413583517074585, 0.7857681512832642, 0.5282618403434753, 0.7957147359848022, 0.4918915927410126, 0.8146284818649292, 0.49434101581573486, 0.8144001960754395, 0.508465051651001, 0.8045610189437866, 0.4914408028125763, 0.18430902063846588, 0.924901008605957, 0.8416562676429749, 0.4020020663738251, 0.8612001538276672, 0.6501213312149048, 0.7012129426002502, 0.6855759024620056, 0.3327900469303131, 0.8782522678375244, 0.34702253341674805, 0.8754173517227173, 0.6493537425994873, 0.27910852432250977, 0.8969979286193848, 0.7512675523757935, 0.411971777677536, 0.8512263298034668, 0.5800846815109253, 0.22813384234905243, 0.9121242761611938, 0.8007039427757263, 0.5000345706939697, 0.1792009472846985, 0.9271676540374756, 0.8509571552276611, 0.6207671761512756, 0.7240164279937744, 0.6591100096702576, 0.6916972398757935, 0.6911197304725647, 0.3402249217033386, 0.8746440410614014, 0.6500211954116821, 0.719810962677002, 0.6309230327606201, 0.7336190938949585, 0.38760659098625183, 0.8593938946723938, 0.616445004940033, 0.26291510462760925, 0.9001639485359192, 0.7577506899833679, 0.5744442939758301, 0.7690213322639465, 0.5550418496131897, 0.7803094983100891, 0.46576184034347534, 0.8227605223655701, 0.5346624255180359, 0.7869624495506287, 0.4784732162952423, 0.8164790868759155, 0.5286208391189575, 0.7853836417198181, 0.5218362212181091, 0.21094933152198792, 0.9143416285514832, 0.8076117038726807, 0.47136515378952026, 0.8195831179618835, 0.44397884607315063, 0.8325489163398743, 0.5871087312698364, 0.7374420762062073, 0.6010329723358154, 0.7265204191207886, 0.6098694801330566, 0.7187575697898865, 0.6143644452095032, 0.713981568813324, 0.6150655746459961, 0.7119860053062439, 0.6123565435409546, 0.7125591039657593, 0.39351317286491394], "advantages": [8.454148292541504, 9.196200370788574, 10.13184928894043, 8.526554107666016, 9.43067455291748, 7.7558722496032715, 6.444916725158691, 5.641612529754639, 5.24130916595459, 5.332033157348633, 4.707914352416992, 4.779219150543213, 5.189318656921387, 5.9860100746154785, 6.874715805053711, 7.2869672775268555, 6.798346996307373, 5.203686237335205, 4.268044471740723, 2.4109091758728027, 0.08171756565570831, 0.7529674172401428, -1.572979211807251, -0.9105083346366882, -3.2298479080200195, -5.45485258102417, -7.253082275390625, 21.1826171875, 21.755958557128906, 21.26007080078125, 21.923534393310547, 21.412616729736328, 21.80182456970215, 21.351381301879883, 21.706893920898438, 21.314987182617188, 22.19284439086914, 21.621185302734375, 21.960540771484375, 21.711021423339844, 22.8201904296875, 22.207000732421875, 22.533287048339844, 23.517616271972656, 22.411645889282227, 23.30562973022461, 22.292142868041992, 23.101261138916016, 24.29068946838379, 22.551244735717773, 21.62848663330078, 22.25595474243164, 21.334819793701172, 21.271648406982422, 21.246126174926758, 21.1569766998291, 22.092788696289062, 21.29338264465332, 22.29145622253418, 21.47045135498047, 22.546674728393555, 21.70326042175293, 21.687564849853516, 22.19637107849121, 21.277427673339844, 21.755176544189453, 20.80590057373047, 20.664134979248047, 21.587360382080078, 23.910627365112305, 22.0589542388916, 21.132137298583984, 21.04785919189453, 21.46733856201172, 20.51736068725586, 20.371789932250977, 20.247907638549805, 20.07720375061035, 21.049482345581055, 23.469865798950195, 21.53590202331543, 20.53139877319336, 20.386066436767578, 20.72189712524414, 19.805334091186523, 19.683643341064453, 19.514415740966797, 19.377185821533203, 19.197734832763672, 19.039470672607422, 18.851181030273438, 19.17801284790039, 18.146800994873047, 17.88332748413086, 18.7255859375, 17.691123962402344, 18.546552658081055, 17.494089126586914, 18.369646072387695, 17.29958724975586, 17.087799072265625, 16.749343872070312, 17.597673416137695, 16.51219940185547, 17.38596534729004, 16.283580780029297, 17.192625045776367, 16.072607040405273, 15.82264232635498, 15.496251106262207, 16.41285514831543, 15.27736759185791, 15.007856369018555, 14.68942642211914, 15.616179466247559, 14.464133262634277, 15.441168785095215, 14.268624305725098, 13.948468208312988, 13.697457313537598, 13.374672889709473, 13.111531257629395, 14.099346160888672, 12.90808391571045, 13.957744598388672, 16.537668228149414, 14.386635780334473, 13.119277000427246, 14.465866088867188, 13.170357704162598, 12.564947128295898, 12.840646743774414, 12.194572448730469, 11.955927848815918, 11.475937843322754, 11.258289337158203, 10.763253211975098, 11.073697090148926, 12.55833911895752, 11.265630722045898, 10.49610710144043, 10.074651718139648, 9.83311653137207, 10.370545387268066, 11.998294830322266, 10.678997993469238, 9.742652893066406, 10.581010818481445, 12.435568809509277, 11.064817428588867, 9.959051132202148, 8.884705543518066, 9.550018310546875, 8.42819595336914, 9.193204879760742, 8.024436950683594, 7.050354480743408, 7.310135841369629, 8.14702320098877, 6.96460485458374, 7.870704650878906, 6.6485395431518555, 5.455810070037842, 5.9837141036987305, 6.917320728302002, 8.14946174621582, 7.1181960105896, 5.809777736663818, 6.901096820831299, 5.584085464477539, 6.702444553375244, 5.387164115905762, 3.7831804752349854, 4.797779560089111, 5.898042678833008, 4.623800277709961, 2.9812726974487305, 4.0402607917785645, 5.092409133911133, 3.858175277709961, 4.8802289962768555, 6.062924861907959, 5.0544867515563965, 3.867476224899292, 4.799589157104492, 3.6534841060638428, 4.531859874725342, 3.4237136840820312, 1.8823411464691162, 2.8144659996032715, 1.3457458019256592, 2.200148582458496, 0.7978718280792236, 1.568223237991333, 0.22619734704494476, 0.9080145955085754, -0.3798799514770508, 0.21072059869766235, -1.0289427042007446, -0.5305595397949219]} +{"type": "SampleBatch", "eps_id": [1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 1996087797, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265, 563481265], "obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAADuxFr+/Jk2+tQJ7uwcprr7Ttxe/OnPKvhg3Lrz6aEe9Gb4ZvwcjF78DKz68a2Z2PuvDHL+NC0m/1qPeu3aVBz9FySC/0REXv8boeDuUcXA+v84jv8BDyr5XKws8OCNovYvUJb9NHRe/UzLxO6BndD4/2ii/dmHKvs3ORjy3qVO9WOAqvwcjTb7y3zU8ulauvuvmK78Cjcq+j5iMO7qgNb1z7S2/CD4XvyoSXzv0rn8+z/Mwv26byr4glgk8qbArvXv6Mr8OSRe/1LP3Oy2+gT4QATa/VrnKvv3iTjy1FBe9CQg4v67UTb7azEI8ca+mvoAPOb9J58q+Qj6wOzXJ7rzvFju/2hROvuwjnTsN6KO+uB48v1z6yr7Io9K6v3zUvFcmPr8ubxe/X1ELuzlQiD6vLUG/s2BJv/ekUTtj6Q4/vTRFvyRwF7+KVms8kXyIPho8SL+BFMu+QFihPE96sLz8Q0q/6I8Xv6/QnTzJ+I0+/EtNv4Bdy74EP8s8dagXvJlUT798TE++xbrJPGWIlr7xXVC/IbrLviCPmTwFqM87fGdSv9zhF7/tmJo8kRicPsLte7x/Ckw84En1vCDFCr0R2Xe81vtUPuLW+rx19au+cLEzvDE3Wjzg7hi9/gZZvS1UL7wwmTm+DkYdvfYmaT5luGq8BGnAvhmfCr1UDQI/ce6yvGCLOL53AsK8j8dRPmJ10LxK/7++5XGgvOne+j7m8ga9j+k3viFVILynx0M+bKkVvUfJv76uXcO7pCf2PvxYNL2Trje+rmxvOy6vPj7LCkO9x42APA7A8Ttirte9ssFBvdrDVz50u6w725PKvtN+ML3UJH48UyItu16Az72GOS+9QLpXPq73mLu5J8q+bPcdvVfafzzx3E28TjrUve6vHL2doDe+ytFvvNWBPT6fYCu9NRyDPF0tM7xxyOW9+hAqvTRaWD5G8Ve8PQ7RvhXCGL1M7oU8ed6uvIxg9b04axe9ab42vsx/wryHCyo+0AkmvaYYv77BSqe8S/7mPh2dRL0eGTa+j786vJLCGz58LlO9z/mOPLLnCLxmmxO+d8BRvWLMWT6qIzi8OQDhvvVTQL0ARJE85xGkvKTxGb4T4D69NUBaPnKzvLyhBua+TGotvShMljyVJwO9sN4nvorpK72giDS+j5UQvdKj8j3eWjq9xymePO3gBr3gkj2++MU4vVyJM75lCxa9LqDGPeEiR71yZL2+eBkOvbp6wT5kcGW9WYQQvxpJ3ry90Cg/nNeJvXBeQr/IfGS8E4ZxP/bwqL3MXhC/blOhO8l7JT9eCsC9m9+8vq49kjzF6LU+fybPvd80Mr6wc8w8BtyLPVVH1r24yKk8AqTXPM2bXb4CbtW986FcPusutDyDLAC/u5rMvawjpDzTTUQ8X/lNvqLIy734dTO+bWQCPEwowz1P9tK9i8+hPBOeITzui0e+MSfSvbu4M75uhsM7J6nOPYtX2b0p9p88CtQCPG9xQr7Kiti9/OwzvoY2iTtTqtc9O73fvVnlvb7TOc475HHMPkzu7r00FDS+KvVpPDFx3j1OIva9mhucPDLGhjyR1je+fVr1vcgSWz5fuFI8phfvvi2X7L3KrZg86c1mO9xaLr6/0+u9g9ZaPjYp9DgSduy+2RLjvUFAmDx8bRW8cy0tvvdP4r3991o+MNhMvIvo7b66jdm9q8aaPIyNsrzpKzS+ncfYvbIhNL5gYc+8/tPgPSn8370ISKA85WS9vEBcQ74AL9+9YGwzvtum3Lxhk8E9TFzmvdcgpjxtKs28yn5TvqeH5b3CqTK+RQHvvPsHoD0prey9hgO9vtUz4ryOE7k+Ksz7vQbVMb5g+qa8JKZ2PZV0Ab51ILI8sBydvPOOdL6VAgG+0EAxvs09xLz7kEM9Ho4Evskotzwza7y8UTqBvuUYBL7pkDC+i8XlvIruBj3qoAe+mhy9PNdf4Lwic4m+4icHvkLBL77cLQa95ll9PL+rCr4eiLu+kukEvWFjmD4SLBK+9swuvoAP2bx7Q6e7DasVvjIQyzyZ5dm8vK+cvhcpFb4RBi6+rQQGvTIBs7wYpBi+EvfRPO7OB73UO6a+tx0YvmUPLb7dZyK9zpwuvciTG74sbdo84eUlvanysb79Bxu+b+Mrvp9eQr05Fou9D3gevoF7ub7f7ke9UHRWPmbjJb47eyq+2cY2vTc0yb1DTCm+N824vi3TPr2WWjg+orAwvngsDr+cEzC9xlLqPlkQPL7TJbi+uZUKvUNjGz4GbkO+/gIovsRO/LzUDBu+PspGvvust77Pjgq9/osGPhUjTr5HESe+lZb/vKbjL754elG+WjO3voHdDb0gJuM9cc5YvpJrDb9/xwS93fHIProeZL7Vtba+mUHJvIfMtz2ubWu+zDYNv2WNuryayL8+vrl2vihdtr73XHq8yDGZPSYFfr6EEw2/H9phvPWpuT4yp4S+A/o+v+UN1rtPoyY/zEqMvpQADb/widQ7aF+2Pqnukb5MJra+6fxePHpHhj1Ek5W+RhMNv/x4dDwkm7k+4Debvrxbtr5WobU8priYPY3dnr6+NQ2/EtnBPKqSvz6Kg6S+jLG2vrkm/zwEWbY97iqovqANJr6a3gY9gE5GvgfUqb76AwU9kALuPNmY877mfqm+YPUmvhcPoDy0QTK+UCqrvqSFt76ziYM8dGT/PfLVrr58eie+I/iXPKHHJr6wgrC+bMa3vquRejxW3Ao+ni+0vhb5J76SgJM8td0bvqHdtb5tgfw8lyB1PMjZ4L7UjLW+iXQovhpyyjsONBG+Ejy3vtEsuL76B1s7SH4cPgzrur4MDQ6/2avRO6Gj5D6mmcC+Aj+4visqezy5pB8+/kjEvmHqKL4UIJc8TRIHvmr5xb40fri+jYOBPGGJKj4Fqsm+nGopvrvMnDynCvi9uVvLvprR8DzX9Ig8tr3QvqoOy77v8Cm+i1EMPBLa4L22wcy+YeXtPDqv0DtKrcy+lnXMvugoKr45NdW65DHXvTIpzr7yWe08JVR0uzjqy74+3c2+5BQqvnSWP7wzqNq9p5DPvkDbuL6mkmK835M6Ph5D077+Uw6/Nd4mvH3k8D6O9Ni+Ta64vgEmS7r4zjI+H6bcvu9JDr92FjI79yHvPilX4r7ksbi+HZFFPLZvMz7MCOa+g8Mpvor8fjyUs+i9ZLvnvnd57jwWwVk8kn7NvhRv576eLyq+xHmsO+EK1r3BIum+qwi5vs72Tzu9YUI+IdbsvtB6Dr/qYuQ7ZI73Ph+J8r6VG7m+iVCIPK+pRT7gPPa+iqgqvs7wpzyEPsG9wvH3vgJhub4qe5g8sKJRPuam+74HOyu+1QW6PCMAqL1AXf2+Oa65vi+VrDwz9F4++IkAv/reK75gQdA87b6LvfZlAb8OBbq+YRPFPHPvbT4sQgO/TJgsvjYl6zzKnFe9GB8Ev4xnur5XheI8RO9+Pkr8Bb9day2+u6cFPeHYDr1E2ga/ANi6vlvMAj0aLYk+lrgIvyRdLr4Xvxg9CeZtvMaXCb8FWbu+lI4XPXNSlD5idwu/SXMvvtpJLz3i4hE89VcMv/WZvTyWBDA9MDqKvp85DL9vtDC+zOYZPSI8Ez3OGw2/J4W8vqTYHD1VNK4+av4Ov2nQMb4NuDg9o0F1PQXiD78NIL2+xJ89PRieuz4uxhG/QCYzvpSkWz0JrLU9fasSv+Gfnjzn6GI9sBA/vhySEr9Cx1o+459TPYYu7L4TehG/IiGSPOLVLT32chy+sWIRv0MiNr7OUSE9QaccPtNLEr+Xkog8EdotPbMRAr75NRK/LFQ3vkByIz3zCDc+oiATv37mfTzOFjI93QbPvVIME79EAlc+2s4pPYttwr4c+RG/OQhqPBazCj3CIJi9Y+YRvx/rVT5LnQQ9k1G2vpLUEL+1hlo8C+POPDyiWr0XwxC/III6vjwkxjzvGH0+0rERv3/zTjwbo+48/soavUShEb+oTFQ+BnLoPNJspL6FkRC/E3xBPFLUszx89aC8C4IQv4GRUz42nLA8lFOcvjxzD7+nhs0+5St9PGlUF78XZQ2/cKgYP7HgbTsTAGG/e1cKv2pjzT7sh2S8KsgVv69JCL/XsRg/OiDSvD7TYb/jOwW/h6vNPrZTMb2W/Ri/Xy0Dv5kuVD6zSGK9bkSjvsgdAr/4m1M8IGh8vfG2NL3aDAK/meBVPrECgL0FELa+F/sAvwuXcDxWk469I4GqvdfnAL/XrTe+UfyRvXIoPz7z0gG/obyIPNxWir36rAK+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "actions": [0, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 0, 0, 1, 0], "rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "prev_actions": [1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 0, 0, 1, 0, 0, 0, 1, 0, 1, 0, 1, 1, 1, 0, 1, 0, 0, 1, 1, 0, 0, 0, 1, 1, 1, 0, 1, 0, 0, 1, 0, 0, 0, 0, 1, 1, 1, 1, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 0, 0, 1, 0, 0, 1, 1, 0, 1, 0, 0, 1, 0, 1, 0, 0, 1, 1, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 0, 0, 1, 1, 0, 1, 1, 0, 1, 0, 1, 0, 0, 0, 1, 0, 1, 1, 1, 0, 0, 0, 1, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 1, 0, 0, 1, 0, 1, 1, 1, 0, 0, 1, 0, 1, 1, 0, 1, 0, 0, 1, 1, 0, 1, 1, 1, 0, 1, 0, 0, 0, 1, 0, 0, 1], "prev_rewards": [1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0], "dones": [false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, true, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false, false], "new_obs": "BCJNGGhAFg0AAAAAAABQFg0AgIAFlQsNAAAAAAAAjCByYXkuY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEV9udW1weV9mcm9tYnVmZmVylJOUKJaADAAAAAAAANO3F786c8q+GDcuvPpoR70Zvhm/ByMXvwMrPrxrZnY+68Mcv40LSb/Wo967dpUHP0XJIL/RERe/xuh4O5RxcD6/ziO/wEPKvlcrCzw4I2i9i9Qlv00dF79TMvE7oGd0Pj/aKL92Ycq+zc5GPLepU71Y4Cq/ByNNvvLfNTy6Vq6+6+YrvwKNyr6PmIw7uqA1vXPtLb8IPhe/KhJfO/Sufz7P8zC/bpvKviCWCTypsCu9e/oyvw5JF7/Us/c7Lb6BPhABNr9Wucq+/eJOPLUUF70JCDi/rtRNvtrMQjxxr6a+gA85v0nnyr5CPrA7NcnuvO8WO7/aFE6+7COdOw3oo764Hjy/XPrKvsij0rq/fNS8VyY+vy5vF79fUQu7OVCIPq8tQb+zYEm/96RRO2PpDj+9NEW/JHAXv4pWazyRfIg+GjxIv4EUy75AWKE8T3qwvPxDSr/ojxe/r9CdPMn4jT78S02/gF3LvgQ/yzx1qBe8mVRPv3xMT77Fusk8ZYiWvvFdUL8husu+II+ZPAWozzt8Z1K/3OEXv+2YmjyRGJw+H3FVv5sAzL5OjMw8VB6VPBHZd7zW+1Q+4tb6vHX1q75wsTO8MTdaPODuGL3+Blm9LVQvvDCZOb4ORh299iZpPmW4arwEacC+GZ8KvVQNAj9x7rK8YIs4vncCwryPx1E+YnXQvEr/v77lcaC86d76PubyBr2P6Te+IVUgvKfHQz5sqRW9R8m/vq5dw7ukJ/Y+/Fg0vZOuN76ubG87Lq8+PssKQ73HjYA8DsDxO2Ku172ywUG92sNXPnS7rDvbk8q+034wvdQkfjxTIi27XoDPvYY5L71Aulc+rveYu7knyr5s9x29V9p/PPHcTbxOOtS97q8cvZ2gN77K0W+81YE9Pp9gK701HIM8XS0zvHHI5b36ECq9NFpYPkbxV7w9DtG+FcIYvUzuhTx53q68jGD1vThrF71pvja+zH/CvIcLKj7QCSa9phi/vsFKp7xL/uY+HZ1EvR4ZNr6Pvzq8ksIbPnwuU73P+Y48sucIvGabE753wFG9YsxZPqojOLw5AOG+9VNAvQBEkTznEaS8pPEZvhPgPr01QFo+crO8vKEG5r5Mai29KEyWPJUnA72w3ie+iukrvaCINL6PlRC90qPyPd5aOr3HKZ487eAGveCSPb74xTi9XIkzvmULFr0uoMY94SJHvXJkvb54GQ69unrBPmRwZb1ZhBC/GknevL3QKD+c14m9cF5Cv8h8ZLwThnE/9vCovcxeEL9uU6E7yXslP14KwL2b37y+rj2SPMXotT5/Js+93zQyvrBzzDwG3Is9VUfWvbjIqTwCpNc8zZtdvgJu1b3zoVw+6y60PIMsAL+7msy9rCOkPNNNRDxf+U2+osjLvfh1M75tZAI8TCjDPU/20r2Lz6E8E54hPO6LR74xJ9K9u7gzvm6Gwzsnqc49i1fZvSn2nzwK1AI8b3FCvsqK2L387DO+hjaJO1Oq1z07vd+9WeW9vtM5zjvkccw+TO7uvTQUNL4q9Wk8MXHePU4i9r2aG5w8MsaGPJHWN759WvW9yBJbPl+4UjymF+++LZfsvcqtmDzpzWY73Fouvr/T672D1lo+Nin0OBJ27L7ZEuO9QUCYPHxtFbxzLS2+90/ivf33Wj4w2Ey8i+jtvrqN2b2rxpo8jI2yvOkrNL6dx9i9siE0vmBhz7z+0+A9KfzfvQhIoDzlZL28QFxDvgAv371gbDO+26bcvGGTwT1MXOa91yCmPG0qzbzKflO+p4flvcKpMr5FAe+8+wegPSmt7L2GA72+1TPivI4TuT4qzPu9BtUxvmD6prwkpnY9lXQBvnUgsjywHJ288450vpUCAb7QQDG+zT3EvPuQQz0ejgS+ySi3PDNrvLxROoG+5RgEvumQML6LxeW8iu4GPeqgB76aHL0811/gvCJzib7iJwe+QsEvvtwtBr3mWX08v6sKvh6Iu76S6QS9YWOYPhIsEr72zC6+gA/ZvHtDp7sNqxW+MhDLPJnl2by8r5y+FykVvhEGLr6tBAa9MgGzvBikGL4S99E87s4HvdQ7pr63HRi+ZQ8tvt1nIr3OnC69yJMbvixt2jzh5SW9qfKxvv0HG75v4yu+n15CvTkWi70PeB6+gXu5vt/uR71QdFY+ZuMlvjt7Kr7Zxja9NzTJvUNMKb43zbi+LdM+vZZaOD6isDC+eCwOv5wTML3GUuo+WRA8vtMluL65lQq9Q2MbPgZuQ77+Aii+xE78vNQMG74+yka++6y3vs+OCr3+iwY+FSNOvkcRJ76Vlv+8puMvvnh6Ub5aM7e+gd0NvSAm4z1xzli+kmsNv3/HBL3d8cg+uh5kvtW1tr6ZQcm8h8y3Pa5ta77MNg2/ZY26vJrIvz6+uXa+KF22vvdcerzIMZk9JgV+voQTDb8f2mG89am5PjKnhL4D+j6/5Q3Wu0+jJj/MSoy+lAANv/CJ1DtoX7Y+qe6Rvkwmtr7p/F48ekeGPUSTlb5GEw2//Hh0PCSbuT7gN5u+vFu2vlahtTymuJg9jd2evr41Db8S2cE8qpK/PoqDpL6Msba+uSb/PARZtj3uKqi+oA0mvpreBj2ATka+B9SpvvoDBT2QAu482ZjzvuZ+qb5g9Sa+Fw+gPLRBMr5QKqu+pIW3vrOJgzx0ZP898tWuvnx6J74j+Jc8occmvrCCsL5sxre+q5F6PFbcCj6eL7S+FvknvpKAkzy13Ru+od21vm2B/DyXIHU8yNngvtSMtb6JdCi+GnLKOw40Eb4SPLe+0Sy4vvoHWztIfhw+DOu6vgwNDr/Zq9E7oaPkPqaZwL4CP7i+Kyp7PLmkHz7+SMS+YeoovhQglzxNEge+avnFvjR+uL6Ng4E8YYkqPgWqyb6caim+u8ycPKcK+L25W8u+mtHwPNf0iDy2vdC+qg7Lvu/wKb6LUQw8EtrgvbbBzL5h5e08Oq/QO0qtzL6Wdcy+6Cgqvjk11brkMde9MinOvvJZ7TwlVHS7OOrLvj7dzb7kFCq+dJY/vDOo2r2nkM++QNu4vqaSYrzfkzo+HkPTvv5TDr813ia8feTwPo702L5Nrri+ASZLuvjOMj4fpty+70kOv3YWMjv3Ie8+KVfivuSxuL4dkUU8tm8zPswI5r6Dwym+ivx+PJSz6L1ku+e+d3nuPBbBWTySfs2+FG/nvp4vKr7Eeaw74QrWvcEi6b6rCLm+zvZPO71hQj4h1uy+0HoOv+pi5Dtkjvc+H4nyvpUbub6JUIg8r6lFPuA89r6KqCq+zvCnPIQ+wb3C8fe+AmG5vip7mDywolE+5qb7vgc7K77VBbo8IwCovUBd/b45rrm+L5WsPDP0Xj74iQC/+t4rvmBB0Dztvou99mUBvw4Fur5hE8U8c+9tPixCA79MmCy+NiXrPMqcV70YHwS/jGe6vleF4jxE734+SvwFv11rLb67pwU94dgOvUTaBr8A2Lq+W8wCPRotiT6WuAi/JF0uvhe/GD0J5m28xpcJvwVZu76Ujhc9c1KUPmJ3C79Jcy++2kkvPeLiETz1Vwy/9Zm9PJYEMD0wOoq+nzkMv2+0ML7M5hk9IjwTPc4bDb8nhby+pNgcPVU0rj5q/g6/adAxvg24OD2jQXU9BeIPvw0gvb7Enz09GJ67Pi7GEb9AJjO+lKRbPQmstT19qxK/4Z+ePOfoYj2wED++HJISv0LHWj7jn1M9hi7svhN6Eb8iIZI84tUtPfZyHL6xYhG/QyI2vs5RIT1Bpxw+00sSv5eSiDwR2i09sxECvvk1Er8sVDe+QHIjPfMINz6iIBO/fuZ9PM4WMj3dBs+9UgwTv0QCVz7azik9i23Cvhz5Eb85CGo8FrMKPcIgmL1j5hG/H+tVPkudBD2TUba+ktQQv7WGWjwL4848PKJavRfDEL8ggjq+PCTGPO8YfT7SsRG/f/NOPBuj7jz+yhq9RKERv6hMVD4Gcug80mykvoWREL8TfEE8UtSzPHz1oLwLghC/gZFTPjacsDyUU5y+PHMPv6eGzT7lK308aVQXvxdlDb9wqBg/seBtOxMAYb97Vwq/amPNPuyHZLwqyBW/r0kIv9exGD86INK8PtNhv+M7Bb+Hq80+tlMxvZb9GL9fLQO/mS5UPrNIYr1uRKO+yB0Cv/ibUzwgaHy98bY0vdoMAr+Z4FU+sQKAvQUQtr4X+wC/C5dwPFaTjr0jgaq91+cAv9etN75R/JG9cig/PvPSAb+hvIg83FaKvfqsAr4TvQG/nag1vvqQj707cxI+lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUSwBLAYeUUpQoSwOMATyUTk5OSv////9K/////0sAdJRiS8hLBIaUjAFDlHSUUpQuAAAAAA==", "action_prob": [0.8386780619621277, 0.6170680522918701, 0.29762154817581177, 0.8812835812568665, 0.7161768674850464, 0.5861281752586365, 0.7250239253044128, 0.4307222366333008, 0.8073381185531616, 0.5680949091911316, 0.7324593663215637, 0.5545117855072021, 0.7401416301727295, 0.46197816729545593, 0.7817394137382507, 0.46400633454322815, 0.7808087468147278, 0.5391877889633179, 0.2582356929779053, 0.8891968131065369, 0.7556220889091492, 0.49789300560951233, 0.7666703462600708, 0.5249384045600891, 0.7187854051589966, 0.46403372287750244, 0.783170759677887, 0.42215362191200256, 0.8561401963233948, 0.6264865398406982, 0.27430078387260437, 0.8981290459632874, 0.28031423687934875, 0.8985252976417542, 0.2752423882484436, 0.9019171595573425, 0.7403168082237244, 0.3794480562210083, 0.863484263420105, 0.3660590350627899, 0.8700645565986633, 0.6569063067436218, 0.7097680568695068, 0.33334463834762573, 0.8819758296012878, 0.6965839862823486, 0.3344842791557312, 0.8818414807319641, 0.6680472493171692, 0.292938768863678, 0.8935829997062683, 0.26617470383644104, 0.9025678038597107, 0.7668690085411072, 0.5634430646896362, 0.7894557118415833, 0.47738903760910034, 0.17747515439987183, 0.07229751348495483, 0.9583975076675415, 0.9343551397323608, 0.8567352890968323, 0.6097709536552429, 0.2616086006164551, 0.8977722525596619, 0.7410420775413513, 0.623099684715271, 0.735676109790802, 0.6295695900917053, 0.7314804196357727, 0.3656616806983948, 0.8761318922042847, 0.6603246331214905, 0.2985423505306244, 0.8877400159835815, 0.28967994451522827, 0.8925988078117371, 0.2708573341369629, 0.9000024795532227, 0.756344735622406, 0.5857874751091003, 0.774013340473175, 0.5562936663627625, 0.7918604016304016, 0.4769485294818878, 0.825166642665863, 0.5114687085151672, 0.8129246234893799, 0.48256686329841614, 0.827178955078125, 0.4487704932689667, 0.842247724533081, 0.5899760127067566, 0.7585522532463074, 0.3902861773967743, 0.8634312152862549, 0.35294678807258606, 0.8768473267555237, 0.311359703540802, 0.8904280662536621, 0.7327364087104797, 0.6089584231376648, 0.7617281079292297, 0.4334581196308136, 0.8386838436126709, 0.5480337142944336, 0.7936604022979736, 0.5161818265914917, 0.8099989295005798, 0.5172784924507141, 0.7994233965873718, 0.5283619165420532, 0.7958510518074036, 0.5302416086196899, 0.20287469029426575, 0.9215015172958374, 0.8156279921531677, 0.4774072468280792, 0.8304324746131897, 0.4410741329193115, 0.8472754955291748, 0.6032063961029053, 0.2706635594367981, 0.8906767964363098, 0.7236868143081665, 0.6357442736625671, 0.7082706689834595, 0.6526896953582764, 0.3077017366886139, 0.879630982875824, 0.6955613493919373, 0.33899566531181335, 0.8788006901741028, 0.6858945488929749, 0.6545942425727844, 0.7015239596366882, 0.364946573972702, 0.8549193739891052, 0.3647949993610382, 0.8572750091552734, 0.3558928370475769, 0.8630107641220093, 0.6615264415740967, 0.31762996315956116, 0.8809774518013, 0.3071174621582031, 0.8852274417877197, 0.7119345664978027, 0.3860858678817749, 0.842163622379303, 0.6167616844177246, 0.281674325466156, 0.892139196395874, 0.7381556630134583, 0.5700241327285767, 0.7520197629928589, 0.5463149547576904, 0.7664353847503662, 0.5200193524360657, 0.7814029455184937, 0.4908669888973236, 0.7968623042106628, 0.4586855173110962, 0.8126866221427917, 0.42346981167793274, 0.828679084777832, 0.614531397819519, 0.6641597151756287, 0.3599458336830139, 0.8538905382156372, 0.32425886392593384, 0.8671514391899109, 0.7131369113922119, 0.4587548077106476, 0.7602694034576416, 0.518530011177063, 0.7551196813583374, 0.4863862693309784, 0.7743653655052185, 0.546653687953949, 0.6964569687843323, 0.5653630495071411, 0.6847819089889526, 0.42160478234291077, 0.8076634407043457, 0.6012789011001587, 0.6533136367797852, 0.6126033663749695, 0.3540980815887451, 0.17488938570022583, 0.9134647250175476, 0.15185877680778503, 0.9264314770698547, 0.877343475818634, 0.7660918831825256, 0.45997706055641174, 0.8052680492401123, 0.6057941317558289, 0.6759114265441895, 0.658988893032074], "advantages": [-14.482077598571777, -15.685440063476562, -16.98165512084961, -18.689483642578125, -18.35429573059082, -18.276588439941406, -19.51108169555664, -19.524442672729492, -19.298851013183594, -20.619993209838867, -21.804187774658203, -22.00367546081543, -23.159025192260742, -23.438030242919922, -23.278175354003906, -24.731821060180664, -24.560253143310547, -26.09242820739746, -27.209142684936523, -28.211423873901367, -28.946826934814453, -29.51181411743164, -30.523115158081055, -31.168453216552734, -31.123394012451172, -32.69401168823242, -33.637882232666016, 20.555456161499023, 21.328594207763672, 20.655712127685547, 20.952741622924805, 22.122591018676758, 20.547502517700195, 21.625835418701172, 20.1485652923584, 21.139678955078125, 19.73833656311035, 19.436079025268555, 20.202980041503906, 19.461645126342773, 20.312252044677734, 19.54850959777832, 19.672834396362305, 19.43801498413086, 20.432655334472656, 19.625638961791992, 19.747753143310547, 20.64727210998535, 19.414600372314453, 19.302186965942383, 20.448312759399414, 19.580795288085938, 20.900094985961914, 20.00322723388672, 20.08972930908203, 20.226919174194336, 20.29640007019043, 21.07607650756836, 22.30028533935547, 23.660932540893555, 21.360206604003906, 19.681779861450195, 18.878206253051758, 18.979501724243164, 20.185815811157227, 19.196428298950195, 19.033777236938477, 19.167083740234375, 18.99104118347168, 19.130474090576172, 18.940969467163086, 19.370500564575195, 18.56631088256836, 18.66836929321289, 19.941455841064453, 18.854915618896484, 20.249412536621094, 19.129140853881836, 20.67262840270996, 19.51805877685547, 19.27872085571289, 19.62356185913086, 19.35845947265625, 19.774532318115234, 19.477773666381836, 19.772727966308594, 19.255674362182617, 19.758981704711914, 19.394580841064453, 19.983083724975586, 19.57109832763672, 20.263437271118164, 19.792551040649414, 19.88861846923828, 19.666452407836914, 20.51788330078125, 19.93822479248047, 20.935327529907227, 20.2713680267334, 21.44013023376465, 20.682174682617188, 20.348600387573242, 20.754009246826172, 20.31786346435547, 20.135793685913086, 19.991016387939453, 20.579177856445312, 20.002666473388672, 20.730066299438477, 20.051679611206055, 19.57231330871582, 19.79802894592285, 19.25499153137207, 19.55126190185547, 18.948070526123047, 18.723281860351562, 18.39723014831543, 18.68284797668457, 18.07526206970215, 18.35774803161621, 17.734115600585938, 17.988117218017578, 18.638629913330078, 19.84379005432129, 19.021364212036133, 18.15662384033203, 18.96457862854004, 18.048799514770508, 18.881301879882812, 20.101469039916992, 19.232481002807617, 18.16534996032715, 17.01949119567871, 17.629920959472656, 18.522127151489258, 17.44660186767578, 18.339542388916016, 19.467432022094727, 18.582555770874023, 19.7237606048584, 18.806705474853516, 19.966487884521484, 19.018878936767578, 17.74654197692871, 16.18673324584961, 17.09309196472168, 15.582658767700195, 16.426761627197266, 17.364768981933594, 18.413944244384766, 17.486553192138672, 16.23153305053711, 14.662501335144043, 15.478194236755371, 16.339202880859375, 15.091856002807617, 15.916193962097168, 14.656871795654297, 15.441458702087402, 14.170373916625977, 14.91200065612793, 13.62952995300293, 14.324494361877441, 13.031535148620605, 13.675199508666992, 12.373804092407227, 12.959850311279297, 13.62219524383545, 12.558680534362793, 11.226810455322266, 11.694637298583984, 10.371880531311035, 10.763226509094238, 11.249979019165039, 12.206489562988281, 10.967809677124023, 9.777645111083984, 10.17342758178711, 8.961209297180176, 9.299959182739258, 10.088408470153809, 8.743889808654785, 9.506643295288086, 8.090642929077148, 6.752964019775391, 6.965818405151367, 7.63548469543457, 6.176755905151367, 6.823762893676758, 8.26770305633545, 10.404308319091797, 8.390480995178223, 10.522939682006836, 8.481830596923828, 6.279304027557373, 4.1629557609558105, 5.3500237464904785, 3.158806800842285, 1.3051939010620117, 1.8467903137207031]} diff --git a/rllib/tests/data/cartpole_small/output-2019-02-03_20-27-20_worker-0_0.json b/rllib/tests/data/cartpole/small.json similarity index 100% rename from rllib/tests/data/cartpole_small/output-2019-02-03_20-27-20_worker-0_0.json rename to rllib/tests/data/cartpole/small.json
[rllib] MARWIL tuned cartpole example (and my own experiments) produce nan rewards only. <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? + Reproduction I have a custom example that produces offline data and picks it up with MARWIL for training. I observed that I get `nan` reward values for my example every time, so I went a step back and used your cartpole example: https://github.com/ray-project/ray/blob/cd5a207d69cdaf05b47d956c18e89d928585eec7/rllib/tuned_examples/marwil/cartpole-marwil.yaml I'm following the exact steps there, i.e. first run ``` ./train.py --run=PPO --env=CartPole-v0 \ --stop='{"timesteps_total": 50000}' \ --config='{"output": "/tmp/out", "batch_mode": "complete_episodes"}' ``` followed by ``` rllib train -f cartpole-marwil.yaml ``` I did this both on my currently preferred stable version `0.8.5`, as well as on the `0.9.0.dev0` wheel. The result is this: ``` == Status == Memory usage on this node: 19.4/32.0 GiB Using FIFO scheduling algorithm. Resources requested: 0/12 CPUs, 0/0 GPUs, 0.0/9.96 GiB heap, 0.0/3.42 GiB objects Result logdir: /Users/maxpumperla/ray_results/cartpole-marwil Number of trials: 2 (2 TERMINATED) +--------------------------------+------------+-------+--------+--------+------------------+--------+----------+ | Trial name | status | loc | beta | iter | total time (s) | ts | reward | |--------------------------------+------------+-------+--------+--------+------------------+--------+----------| | MARWIL_CartPole-v0_7af06_00000 | TERMINATED | | 0 | 2206 | 58.5661 | 500007 | nan | | MARWIL_CartPole-v0_7af06_00001 | TERMINATED | | 1 | 2248 | 58.6117 | 500286 | nan | +--------------------------------+------------+-------+--------+--------+------------------+--------+----------+ ``` Also, I've noticed that your MARWIL unit test is a pure smoke test and doesn't check reward values, but I didn't run that locally. Maybe it produces nan values as well. In any case I'd appreciate any input here, as we'd love to use MARWIL for our "real" use case, in which we see the same behaviour.
Hi @maxpumperla , thanks for filing this. Yes, we don't actually have an actual MARWIL learning test, that's true. I'll take a look. I remember that the NaNs are b/c of the offline learning setup, but I'll confirm that this is a non-issue and get back to you. Either way, we shouldn't be displaying NaN and instead evaluate on a live env and report actual rewards. Ok, confirmed. All NaNs. But makes sense, if input is a file, then there are no samplers, hence no actual env. We should probably have sampling for the evaluation workers, though to indicate actual learning progress. ... Got it: Alter the rllib/agents/marwil/tests/test_marwil.py file as follows: ``` def test_marwil_compilation(self): """Test whether a MARWILTrainer can be built with all frameworks.""" config = marwil.DEFAULT_CONFIG.copy() config["num_workers"] = 0 # Run locally. config["evaluation_num_workers"] = 1 config["evaluation_interval"] = 1 # <- this is switched off in MARWIL's default config (probably shouldn't be) config["evaluation_config"] = {"input": "sampler"} # <- make sure evaluation workers have an actual Env config["input"] = ["[your file with pre-recorded data]"] num_iterations = 2000 # Test for all frameworks. for _ in framework_iterator(config, frameworks="torch"): trainer = marwil.MARWILTrainer(config=config, env="CartPole-v0") for i in range(num_iterations): print(trainer.train()) # <- in the returned dict, look for the "evaluation" key, which holds a sub-dict with the correct rewards from the evaluation workers' envs in it. trainer.stop() ``` @maxpumperla Will discuss internally, whether we should change the default settings of MARWIL. Also, I'm wondering, why tune would not print out the evaluation results (it prints out the "normal" workers' results, hence NaN). I'm closing this issue. Feel free to re-open in case the above solution does not work on your end. Again, we'll make sure this becomes more user-friendly in a next version.
2020-07-12T14:34:01
ray-project/ray
9,461
ray-project__ray-9461
[ "8821" ]
222635b63ff2a62c14bd3611f8c79ffba38e390c
diff --git a/python/ray/tune/analysis/experiment_analysis.py b/python/ray/tune/analysis/experiment_analysis.py --- a/python/ray/tune/analysis/experiment_analysis.py +++ b/python/ray/tune/analysis/experiment_analysis.py @@ -201,6 +201,11 @@ class ExperimentAnalysis(Analysis): """ def __init__(self, experiment_checkpoint_path, trials=None): + experiment_checkpoint_path = os.path.expanduser( + experiment_checkpoint_path) + if not os.path.isfile(experiment_checkpoint_path): + raise ValueError( + "{} is not a valid file.".format(experiment_checkpoint_path)) with open(experiment_checkpoint_path) as f: _experiment_state = json.load(f) self._experiment_state = _experiment_state
diff --git a/python/ray/tune/tests/test_experiment_analysis_mem.py b/python/ray/tune/tests/test_experiment_analysis_mem.py --- a/python/ray/tune/tests/test_experiment_analysis_mem.py +++ b/python/ray/tune/tests/test_experiment_analysis_mem.py @@ -1,12 +1,16 @@ +import json import unittest import shutil import tempfile +import os import random import pandas as pd +import pytest import numpy as np import ray -from ray.tune import run, Trainable, sample_from, Analysis, grid_search +from ray.tune import (run, Trainable, sample_from, Analysis, + ExperimentAnalysis, grid_search) from ray.tune.examples.async_hyperband_example import MyTrainableClass @@ -37,14 +41,36 @@ def load_checkpoint(self, checkpoint_path): pass self.MockTrainable = MockTrainable + self.test_dir = tempfile.mkdtemp() ray.init(local_mode=False, num_cpus=1) def tearDown(self): shutil.rmtree(self.test_dir, ignore_errors=True) ray.shutdown() + def testInit(self): + experiment_checkpoint_path = os.path.join(self.test_dir, + "experiment_state.json") + checkpoint_data = { + "checkpoints": [{ + "trainable_name": "MockTrainable", + "logdir": "/mock/test/MockTrainable_0_id=3_2020-07-12" + }] + } + + with open(experiment_checkpoint_path, "w") as f: + f.write(json.dumps(checkpoint_data)) + + experiment_analysis = ExperimentAnalysis(experiment_checkpoint_path) + self.assertEqual(len(experiment_analysis._checkpoints), 1) + self.assertTrue(experiment_analysis.trials is None) + + def testInitException(self): + experiment_checkpoint_path = os.path.join(self.test_dir, "mock.json") + with pytest.raises(ValueError): + ExperimentAnalysis(experiment_checkpoint_path) + def testCompareTrials(self): - self.test_dir = tempfile.mkdtemp() scores = np.asarray(list(self.MockTrainable.scores_dict.values())) scores_all = scores.flatten("F") scores_last = scores_all[5:] @@ -126,7 +152,7 @@ def testDataframe(self): analysis = Analysis(self.test_dir) df = analysis.dataframe() self.assertTrue(isinstance(df, pd.DataFrame)) - self.assertEquals(df.shape[0], self.num_samples * 2) + self.assertEqual(df.shape[0], self.num_samples * 2) def testBestLogdir(self): analysis = Analysis(self.test_dir) @@ -134,17 +160,16 @@ def testBestLogdir(self): self.assertTrue(logdir.startswith(self.test_dir)) logdir2 = analysis.get_best_logdir(self.metric, mode="min") self.assertTrue(logdir2.startswith(self.test_dir)) - self.assertNotEquals(logdir, logdir2) + self.assertNotEqual(logdir, logdir2) def testBestConfigIsLogdir(self): analysis = Analysis(self.test_dir) for metric, mode in [(self.metric, "min"), (self.metric, "max")]: logdir = analysis.get_best_logdir(metric, mode=mode) best_config = analysis.get_best_config(metric, mode=mode) - self.assertEquals(analysis.get_all_configs()[logdir], best_config) + self.assertEqual(analysis.get_all_configs()[logdir], best_config) if __name__ == "__main__": - import pytest import sys sys.exit(pytest.main(["-v", __file__]))
[tune] ExperimentAnalysis doesn't expand user as docs imply ### What is the problem? The Ray docs give an example where a path with the "~" operator is included: https://docs.ray.io/en/master/tune/api_docs/analysis.html#experimentanalysis-tune-experimentanalysis ``` >>> tune.run(my_trainable, name="my_exp", local_dir="~/tune_results") >>> analysis = ExperimentAnalysis( >>> experiment_checkpoint_path="~/tune_results/my_exp/state.json") ``` This throws an error for me: ``` from ray.tune import ExperimentAnalysis analysis = ExperimentAnalysis("~/rayexp/experiment_state-2020-06-06_20-22-19.json") 203 def __init__(self, experiment_checkpoint_path, trials=None): --> 204 with open(experiment_checkpoint_path) as f: 205 _experiment_state = json.load(f) 206 self._experiment_state = _experiment_state FileNotFoundError: [Errno 2] No such file or directory: '~/rayexp/experiment_state-2020-06-06_20-22-19.json' ``` Whereas the corresponding expanded version does not: ``` from ray.tune import ExperimentAnalysis from os.path import expanduser analysis = ExperimentAnalysis(expanduser("~/rayexp/experiment_state-2020-06-06_20-22-19.json")) ``` *Ray version and other system information (Python version, TensorFlow version, OS):* Ray 0.8.5 Python 3.8 Ubuntu 18 ### Reproduction (REQUIRED) See description above.
2020-07-14T02:04:02
ray-project/ray
9,479
ray-project__ray-9479
[ "9441" ]
ca4f3b79db968cd8ad31efb1170b8ba2c7bc38ba
diff --git a/python/ray/serve/request_params.py b/python/ray/serve/request_params.py --- a/python/ray/serve/request_params.py +++ b/python/ray/serve/request_params.py @@ -42,7 +42,7 @@ def adjust_relative_slo_ms(self) -> float: return current_time_ms + slo_ms def ray_serialize(self): - return pickle.dumps(self.__dict__, protocol=5) + return pickle.dumps(self.__dict__) @staticmethod def ray_deserialize(value): diff --git a/python/ray/serve/router.py b/python/ray/serve/router.py --- a/python/ray/serve/router.py +++ b/python/ray/serve/router.py @@ -3,10 +3,10 @@ from collections import defaultdict, deque import time from typing import DefaultDict, List +import pickle import blist -import ray.cloudpickle as pickle from ray.exceptions import RayTaskError import ray @@ -50,7 +50,7 @@ def ray_serialize(self): # worker without removing async_future. clone = copy.copy(self).__dict__ clone.pop("async_future") - return pickle.dumps(clone, protocol=5) + return pickle.dumps(clone) @staticmethod def ray_deserialize(value):
diff --git a/python/ray/serve/tests/test_regression.py b/python/ray/serve/tests/test_regression.py new file mode 100644 --- /dev/null +++ b/python/ray/serve/tests/test_regression.py @@ -0,0 +1,38 @@ +import numpy as np +import requests + +from ray import serve + + +def test_np_in_composed_model(serve_instance): + # https://github.com/ray-project/ray/issues/9441 + # AttributeError: 'bytes' object has no attribute 'readonly' + # in cloudpickle _from_numpy_buffer + + def sum_model(_request, data=None): + return np.sum(data) + + class ComposedModel: + def __init__(self): + self.model = serve.get_handle("sum_model") + + async def __call__(self, _request): + data = np.ones((10, 10)) + result = await self.model.remote(data=data) + return result + + serve.create_backend("sum_model", sum_model) + serve.create_endpoint("sum_model", backend="sum_model") + serve.create_backend("model", ComposedModel) + serve.create_endpoint( + "model", backend="model", route="/model", methods=['GET']) + + result = requests.get("http://127.0.0.1:8000/model") + assert result.status_code == 200 + assert result.json() == 100.0 + + +if __name__ == "__main__": + import sys + import pytest + sys.exit(pytest.main(["-v", "-s", __file__]))
[serve] Composed models with numpy arguments ### What is the problem? I tried modifying the [composing multiple models](https://docs.ray.io/en/master/serve/advanced.html?highlight=pipeline#composing-multiple-models) examples to pass a numpy array to another endpoint. However, I get the following error: ``` 2020-07-13 22:21:37,329 ERROR worker.py:987 -- Possible unhandled error from worker: ray::RayServeWorker_sum_model.handle_request() (pid=22037, ip=192.168.1.31) File "python/ray/_raylet.pyx", line 410, in ray._raylet.execute_task File "python/ray/_raylet.pyx", line 424, in ray._raylet.execute_task File "python/ray/_raylet.pyx", line 1212, in ray._raylet.CoreWorker.run_async_func_in_event_loop File "/home/users/daniel/.local/conda/lib/python3.7/concurrent/futures/_base.py", line 428, in result return self.__get_result() File "/home/users/daniel/.local/conda/lib/python3.7/concurrent/futures/_base.py", line 384, in __get_result raise self._exception File "python/ray/_raylet.pyx", line 422, in deserialize_args File "/home/users/daniel/.cache/pypoetry/virtualenvs/ray-test-mjIDN_hR-py3.7/lib/python3.7/site-packages/ray/serialization.py", line 312, in deserialize_objects self._deserialize_object(data, metadata, object_id)) File "/home/users/daniel/.cache/pypoetry/virtualenvs/ray-test-mjIDN_hR-py3.7/lib/python3.7/site-packages/ray/serialization.py", line 252, in _deserialize_object return self._deserialize_msgpack_data(data, metadata) File "/home/users/daniel/.cache/pypoetry/virtualenvs/ray-test-mjIDN_hR-py3.7/lib/python3.7/site-packages/ray/serialization.py", line 233, in _deserialize_msgpack_data python_objects = self._deserialize_pickle5_data(pickle5_data) File "/home/users/daniel/.cache/pypoetry/virtualenvs/ray-test-mjIDN_hR-py3.7/lib/python3.7/site-packages/ray/serialization.py", line 221, in _deserialize_pickle5_data obj = pickle.loads(in_band) File "/home/users/daniel/.cache/pypoetry/virtualenvs/ray-test-mjIDN_hR-py3.7/lib/python3.7/site-packages/ray/serve/router.py", line 53, in ray_deserialize kwargs = pickle.loads(value) File "/home/users/daniel/.cache/pypoetry/virtualenvs/ray-test-mjIDN_hR-py3.7/lib/python3.7/site-packages/ray/cloudpickle/cloudpickle_fast.py", line 448, in _numpy_frombuffer array.setflags(write=isinstance(buffer, bytearray) or not buffer.readonly) AttributeError: 'bytes' object has no attribute 'readonly' ``` Ray version: 0.8.6 Python version: 3.7.7 Numpy version: 1.19.0 ### Reproduction (REQUIRED) ``` import time from ray import serve import requests import numpy as np serve.init() def sum_model(_request, data=None): return np.sum(data) class ComposedModel: def __init__(self): self.model = serve.get_handle("sum_model") async def __call__(self, _request): data = np.ones((10, 10)) result = await self.model.remote(data=data) return result serve.create_backend("sum_model", sum_model) serve.create_endpoint("sum_model", backend="sum_model") serve.create_backend("model", ComposedModel) serve.create_endpoint("model", backend="model", route="/model", methods=['GET']) while True: result = requests.get('http://127.0.0.1:8000/model') print(result) time.sleep(1) ``` - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-07-14T22:31:01
ray-project/ray
9,516
ray-project__ray-9516
[ "9218" ]
2f674728a683ec7d466917afc6327308f2483ec1
diff --git a/rllib/evaluation/rollout_worker.py b/rllib/evaluation/rollout_worker.py --- a/rllib/evaluation/rollout_worker.py +++ b/rllib/evaluation/rollout_worker.py @@ -9,9 +9,6 @@ TYPE_CHECKING, TypeVar import ray -from ray.util.debug import log_once, disable_log_once_globally, \ - enable_periodic_logging -from ray.util.iter import ParallelIteratorWorker from ray.rllib.env.atari_wrappers import wrap_deepmind, is_atari from ray.rllib.env.base_env import BaseEnv from ray.rllib.env.env_context import EnvContext @@ -21,17 +18,17 @@ from ray.rllib.env.vector_env import VectorEnv from ray.rllib.evaluation.sampler import AsyncSampler, SyncSampler from ray.rllib.evaluation.rollout_metrics import RolloutMetrics -from ray.rllib.policy.sample_batch import MultiAgentBatch, DEFAULT_POLICY_ID -from ray.rllib.policy.policy import Policy -from ray.rllib.policy.tf_policy import TFPolicy -from ray.rllib.policy.torch_policy import TorchPolicy +from ray.rllib.models import ModelCatalog +from ray.rllib.models.preprocessors import NoPreprocessor, Preprocessor from ray.rllib.offline import NoopOutput, IOContext, OutputWriter, InputReader from ray.rllib.offline.off_policy_estimator import OffPolicyEstimator, \ OffPolicyEstimate from ray.rllib.offline.is_estimator import ImportanceSamplingEstimator from ray.rllib.offline.wis_estimator import WeightedImportanceSamplingEstimator -from ray.rllib.models import ModelCatalog -from ray.rllib.models.preprocessors import NoPreprocessor, Preprocessor +from ray.rllib.policy.sample_batch import MultiAgentBatch, DEFAULT_POLICY_ID +from ray.rllib.policy.policy import Policy +from ray.rllib.policy.tf_policy import TFPolicy +from ray.rllib.policy.torch_policy import TorchPolicy from ray.rllib.utils import merge_dicts from ray.rllib.utils.annotations import DeveloperAPI from ray.rllib.utils.debug import summarize @@ -42,6 +39,9 @@ from ray.rllib.utils.types import EnvType, AgentID, PolicyID, EnvConfigDict, \ ModelConfigDict, TrainerConfigDict, SampleBatchType, ModelWeights, \ ModelGradients, MultiAgentPolicyConfigDict +from ray.util.debug import log_once, disable_log_once_globally, \ + enable_periodic_logging +from ray.util.iter import ParallelIteratorWorker if TYPE_CHECKING: from ray.rllib.agents.callbacks import DefaultCallbacks @@ -399,22 +399,30 @@ def make_env(vector_index): tf1.set_random_seed(seed) self.policy_map, self.preprocessors = \ self._build_policy_map(policy_dict, policy_config) - if (ray.is_initialized() - and ray.worker._mode() != ray.worker.LOCAL_MODE): - if not ray.get_gpu_ids(): - logger.debug( - "Creating policy evaluation worker {}".format( - worker_index) + - " on CPU (please ignore any CUDA init errors)") - elif not tf1.test.is_gpu_available(): - raise RuntimeError( - "GPUs were assigned to this worker by Ray, but " - "TensorFlow reports GPU acceleration is disabled. " - "This could be due to a bad CUDA or TF installation.") else: self.policy_map, self.preprocessors = self._build_policy_map( policy_dict, policy_config) + if (ray.is_initialized() and + ray.worker._mode() != ray.worker.LOCAL_MODE): + # Check available number of GPUs + if not ray.get_gpu_ids(): + logger.debug( + "Creating policy evaluation worker {}".format( + worker_index) + + " on CPU (please ignore any CUDA init errors)") + elif (policy_config["framework"] in ["tf2", "tf", "tfe"] and + not tf.config.list_physical_devices("GPU")) or \ + (policy_config["framework"] == "torch" and + not torch.cuda.is_available()): + raise RuntimeError( + "GPUs were assigned to this worker by Ray, but " + "your DL framework ({}) reports GPU acceleration is " + "disabled. This could be due to a bad CUDA- or {} " + "installation.".format( + policy_config["framework"], + policy_config["framework"])) + self.multiagent: bool = set( self.policy_map.keys()) != {DEFAULT_POLICY_ID} if self.multiagent: diff --git a/rllib/policy/torch_policy.py b/rllib/policy/torch_policy.py --- a/rllib/policy/torch_policy.py +++ b/rllib/policy/torch_policy.py @@ -4,6 +4,7 @@ import time from typing import Callable, Dict, List, Optional, Tuple, Union +import ray from ray.rllib.models.modelv2 import ModelV2 from ray.rllib.models.torch.torch_modelv2 import TorchModelV2 from ray.rllib.models.torch.torch_action_dist import TorchDistributionWrapper @@ -98,8 +99,10 @@ def __init__(self, """ self.framework = "torch" super().__init__(observation_space, action_space, config) - self.device = (torch.device("cuda") - if torch.cuda.is_available() else torch.device("cpu")) + if torch.cuda.is_available() and ray.get_gpu_ids(): + self.device = torch.device("cuda") + else: + self.device = torch.device("cpu") self.model = model.to(self.device) self.exploration = self._create_exploration() self.unwrapped_model = model # used to support DistributedDataParallel
[rllib] num_gpus=0 on a device with GPU uses GPU for Pytorch ### What is the problem? Ray will find a GPU and place the model (e.g. FCNet) on the GPU even when `num_gpus=0`. Stack Trace: ``` ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/rllib/agents/trainer.py:520: in train raise e ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/rllib/agents/trainer.py:506: in train result = Trainable.train(self) ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/tune/trainable.py:260: in train result = self._train() ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/rllib/agents/trainer_template.py:139: in _train return self._train_exec_impl() ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/rllib/agents/trainer_template.py:177: in _train_exec_impl res = next(self.train_exec_impl) ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/util/iter.py:731: in __next__ return next(self.built_iterator) ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/util/iter.py:752: in apply_foreach result = fn(item) ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/rllib/agents/maml/maml.py:143: in __call__ fetches = self.workers.local_worker().learn_on_batch(samples) ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/rllib/evaluation/rollout_worker.py:742: in learn_on_batch .learn_on_batch(samples) ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/rllib/policy/torch_policy.py:244: in learn_on_batch self._loss(self, self.model, self.dist_class, train_batch)) ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/rllib/agents/maml/maml_torch_policy.py:329: in maml_loss logits, state = model.from_batch(train_batch) ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/rllib/models/modelv2.py:224: in from_batch return self.__call__(input_dict, states, train_batch.get("seq_lens")) ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/rllib/models/modelv2.py:181: in __call__ res = self.forward(restored, state or [], seq_lens) ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/rllib/models/torch/fcnet.py:118: in forward self._features = self._hidden_layers(self._last_flat_in) ../../miniconda3/envs/ray/lib/python3.7/site-packages/torch/nn/modules/module.py:550: in __call__ result = self.forward(*input, **kwargs) ../../miniconda3/envs/ray/lib/python3.7/site-packages/torch/nn/modules/container.py:100: in forward input = module(input) ../../miniconda3/envs/ray/lib/python3.7/site-packages/torch/nn/modules/module.py:550: in __call__ result = self.forward(*input, **kwargs) ../../miniconda3/envs/ray/lib/python3.7/site-packages/ray/rllib/models/torch/misc.py:110: in forward return self._model(x) ../../miniconda3/envs/ray/lib/python3.7/site-packages/torch/nn/modules/module.py:550: in __call__ result = self.forward(*input, **kwargs) ../../miniconda3/envs/ray/lib/python3.7/site-packages/torch/nn/modules/container.py:100: in forward input = module(input) ../../miniconda3/envs/ray/lib/python3.7/site-packages/torch/nn/modules/module.py:550: in __call__ result = self.forward(*input, **kwargs) ../../miniconda3/envs/ray/lib/python3.7/site-packages/torch/nn/modules/linear.py:87: in forward return F.linear(input, self.weight, self.bias) RuntimeError: Expected object of device type cuda but got device type cpu for argument #2 'mat1' in call to _th_addmm ``` ### Reproduction (REQUIRED) Run this in `ray/rlllib`: `pytest -v -s agents/maml/tests/test_maml.py` on a machine with a GPU. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
@michaelzhiluo Looking at this now ...
2020-07-16T09:03:33
ray-project/ray
9,517
ray-project__ray-9517
[ "9036" ]
2f674728a683ec7d466917afc6327308f2483ec1
diff --git a/python/ray/tune/checkpoint_manager.py b/python/ray/tune/checkpoint_manager.py --- a/python/ray/tune/checkpoint_manager.py +++ b/python/ray/tune/checkpoint_manager.py @@ -109,6 +109,10 @@ def on_checkpoint(self, checkpoint): return old_checkpoint = self.newest_persistent_checkpoint + + if old_checkpoint.value == checkpoint.value: + return + self.newest_persistent_checkpoint = checkpoint # Remove the old checkpoint if it isn't one of the best ones.
diff --git a/python/ray/tune/tests/test_trial_scheduler_pbt.py b/python/ray/tune/tests/test_trial_scheduler_pbt.py new file mode 100644 --- /dev/null +++ b/python/ray/tune/tests/test_trial_scheduler_pbt.py @@ -0,0 +1,90 @@ +import numpy as np +import os +import pickle +import random +import unittest +import sys + +from ray import tune +from ray.tune.schedulers import PopulationBasedTraining + + +class MockTrainable(tune.Trainable): + def setup(self, config): + self.iter = 0 + self.a = config["a"] + self.b = config["b"] + self.c = config["c"] + + def step(self): + self.iter += 1 + return {"mean_accuracy": (self.a - self.iter) * self.b} + + def save_checkpoint(self, tmp_checkpoint_dir): + checkpoint_path = os.path.join(tmp_checkpoint_dir, "model.mock") + with open(checkpoint_path, "wb") as fp: + pickle.dump((self.a, self.b), fp) + return tmp_checkpoint_dir + + def load_checkpoint(self, tmp_checkpoint_dir): + checkpoint_path = os.path.join(tmp_checkpoint_dir, "model.mock") + with open(checkpoint_path, "rb") as fp: + self.a, self.b = pickle.load(fp) + + +class MockParam(object): + def __init__(self, params): + self._params = params + self._index = 0 + + def __call__(self, *args, **kwargs): + val = self._params[self._index % len(self._params)] + self._index += 1 + return val + + +class PopulationBasedTrainingResumeTest(unittest.TestCase): + def testPermutationContinuation(self): + """ + Tests continuation of runs after permutation. + Sometimes, runs were continued from deleted checkpoints. + This deterministic initialisation would fail when the + fix was not applied. + See issues #9036, #9036 + """ + scheduler = PopulationBasedTraining( + time_attr="training_iteration", + metric="mean_accuracy", + mode="max", + perturbation_interval=1, + log_config=True, + hyperparam_mutations={"c": lambda: 1}) + + param_a = MockParam([10, 20, 30, 40]) + param_b = MockParam([1.2, 0.9, 1.1, 0.8]) + + random.seed(100) + np.random.seed(1000) + tune.run( + MockTrainable, + config={ + "a": tune.sample_from(lambda _: param_a()), + "b": tune.sample_from(lambda _: param_b()), + "c": 1 + }, + fail_fast=True, + num_samples=20, + global_checkpoint_period=1, + checkpoint_freq=1, + checkpoint_at_end=True, + keep_checkpoints_num=1, + checkpoint_score_attr="min-training_iteration", + scheduler=scheduler, + name="testPermutationContinuation", + stop={"training_iteration": 5}) + + +if __name__ == "__main__": + import pytest + + sys.exit(pytest.main(["-v", __file__]))
[tune] Population-based training: broken when using keep_checkpoint_num When using **population-based** training TUNE stops after some times throwing the following error: `There are paused trials, but no more pending trials with sufficient resources.` This is caused by not finding the latest checkpoint: ``` Failure # 1 (occurred at 2020-06-19_11-26-36) Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/ray/tune/ray_trial_executor.py", line 294, in start_trial self._start_trial(trial, checkpoint, train=train) File "/usr/local/lib/python3.6/dist-packages/ray/tune/ray_trial_executor.py", line 235, in _start_trial self.restore(trial, checkpoint) File "/usr/local/lib/python3.6/dist-packages/ray/tune/ray_trial_executor.py", line 673, in restore data_dict = TrainableUtil.pickle_checkpoint(value) File "/usr/local/lib/python3.6/dist-packages/ray/tune/trainable.py", line 62, in pickle_checkpoint checkpoint_dir = TrainableUtil.find_checkpoint_dir(checkpoint_path) File "/usr/local/lib/python3.6/dist-packages/ray/tune/trainable.py", line 87, in find_checkpoint_dir raise FileNotFoundError("Path does not exist", checkpoint_path) FileNotFoundError: [Errno Path does not exist] /content/TRASH_TUNE_PBT_oversampling_mimic_densenet121/TUNE_Model_0_2020-06-19_11-24-215xncry9c/checkpoint_6/ ``` The error appears to be somewhat random since it only appears after quite some iterations The error can be reproduced in this [colab notebook](https://colab.research.google.com/drive/1-o896bEUm7DTvS24Do0btlqbSHre49MH?usp=sharing). **It is not a COLAB related issue since the same problem arises on our own server.** @richardliaw Is this related to #8772 ?
Yeah this seems like an issue. We'll take a look into this. This is the issue (just for reference): ``` from ray.tune.schedulers import PopulationBasedTraining scheduler = PopulationBasedTraining( time_attr='training_iteration', metric='mean_accuracy', mode='max', perturbation_interval=1, log_config=True, hyperparam_mutations={ "lr": lambda: 10 ** np.random.uniform(-2, -5) }) train_config = {'lr': 0.01} analysis = tune.run( TUNE_Model, # Reference to the model class. config=train_config, fail_fast=True, # Stop after first error num_samples=4, # Num of different hyperparameter configurations search_alg=None, # Search algotihm like Hyperopt, Nevergrad... resources_per_trial={"gpu": 1}, # Requested resources per TRIAL. If set to fraction, total_GPUS/res_p_trial can be run in parallel global_checkpoint_period=np.inf, # Do not save checkpoints based on time interval checkpoint_freq = 1, # Save checkpoint every time the checkpoint_score_attr improves checkpoint_at_end = True, keep_checkpoints_num = 2, # Keep only the best checkpoint checkpoint_score_attr = 'mean_accuracy', # Metric used to compare checkpoints verbose=1, scheduler=scheduler, # Stop bad trials early name="TRASH_TUNE_PBT_oversampling_mimic_densenet121", local_dir="/content", stop={ # Stop a single trial if one of the conditions are met "mean_accuracy": 0.85, "training_iteration": 15}) ``` @Cryptiex for a temporary workaround, I think you can just do: ```python def _save(self, tmp_checkpoint_dir): """ Saves the model checkpoint depending of checkpoint_freq and checkpoint_at_end set in tune.run. Arguments: tmp_checkpoint_dir {str} -- [Folder where the checkpoints get saved.] Returns: [str] -- [tmp_checkpoint_dir] """ checkpoint_path = os.path.join(tmp_checkpoint_dir, "model.pth") torch.save(self.model.state_dict(), checkpoint_path) return checkpoint_path # CHANGE THIS def _restore(self, checkpoint_path): # CHANGE THIS """ Restore model. """ self.model.load_state_dict(torch.load(checkpoint_path)) I think the temporary fix actually is to comment this out: ``` keep_checkpoints_num = 2, # Keep only the best checkpoint ``` Thanks for the fast reply! Commenting out **keep_checkpoints_num** worked for me as a temporary fix. Any idea where this is coming from? Does the PBT scheduler reference an already deleted checkpoint? Yeah, I think it's something about not properly bookkeeping the checkpoints on disk. Any updates on this? The temporary fix seems not to work as indicated by #8441 and my own experiences
2020-07-16T09:16:32
ray-project/ray
9,521
ray-project__ray-9521
[ "9295" ]
ac39e23145336fa7d450dd170e3e12ffc3d72d83
diff --git a/rllib/policy/sample_batch.py b/rllib/policy/sample_batch.py --- a/rllib/policy/sample_batch.py +++ b/rllib/policy/sample_batch.py @@ -438,8 +438,8 @@ def timeslices(self, k: int) -> List["MultiAgentBatch"]: steps = [] for policy_id, batch in self.policy_batches.items(): for row in batch.rows(): - steps.append((row[SampleBatch.EPS_ID], row["t"], policy_id, - row)) + steps.append((row[SampleBatch.EPS_ID], row["t"], + row["agent_index"], policy_id, row)) steps.sort() finished_slices = [] @@ -458,7 +458,7 @@ def finish_slice(): # For each unique env timestep. for _, group in itertools.groupby(steps, lambda x: x[:2]): # Accumulate into the current slice. - for _, _, policy_id, row in group: + for _, _, _, policy_id, row in group: cur_slice[policy_id].add_values(**row) cur_slice_size += 1 # Slice has reached target number of env steps.
[RLlib] Mutiagent learning: can't combine replay lockstep and multiple agents controlled by the same policy. Hello, While implementing MADDPG for PyTorch I noticed that it is not possible to combine the lockstep replay mode configuration and having multiple agents controlled by the same policy. This is due to the implementation of MultiAgentBatch as a dict of `PolicyID -> SampleBatch`. In the contrib TensorFlow implementation this issue was circumvented using parameter sharing. Another solution which is framework agnostic would be to group the agents. I think that this combination should be supported without needing to group the agents. However, I suspect that it would require significant code change. Maybe another solution exists? Also, the documentation should be updated alongside with the code that checks that the config is valid. Thanks (Amazing work btw)
Hmm, I think this should be ok. The batch will indeed be keyed by policy id, but you can look into the "agent_id" field of the batch to see which experiences correspond with which original agent. Okk I will look into that next week and let you know. Thx
2020-07-16T16:40:43
ray-project/ray
9,525
ray-project__ray-9525
[ "9307" ]
94fcd43593a7b7675ff857cbfd6d189965899a65
diff --git a/python/ray/tune/progress_reporter.py b/python/ray/tune/progress_reporter.py --- a/python/ray/tune/progress_reporter.py +++ b/python/ray/tune/progress_reporter.py @@ -5,7 +5,7 @@ from ray.tune.result import (EPISODE_REWARD_MEAN, MEAN_ACCURACY, MEAN_LOSS, TRAINING_ITERATION, TIME_TOTAL_S, TIMESTEPS_TOTAL) -from ray.tune.utils import flatten_dict +from ray.tune.utils import unflattened_lookup try: from collections.abc import Mapping @@ -466,9 +466,9 @@ def _get_trial_info(trial, parameters, metrics): parameters (list[str]): Names of trial parameters to include. metrics (list[str]): Names of metrics to include. """ - result = flatten_dict(trial.last_result) - config = flatten_dict(trial.config) + result = trial.last_result + config = trial.config trial_info = [str(trial), trial.status, str(trial.location)] - trial_info += [config.get(param) for param in parameters] - trial_info += [result.get(metric) for metric in metrics] + trial_info += [unflattened_lookup(param, config) for param in parameters] + trial_info += [unflattened_lookup(metric, result) for metric in metrics] return trial_info diff --git a/python/ray/tune/utils/__init__.py b/python/ray/tune/utils/__init__.py --- a/python/ray/tune/utils/__init__.py +++ b/python/ray/tune/utils/__init__.py @@ -1,6 +1,6 @@ from ray.tune.utils.util import deep_update, flatten_dict, get_pinned_object, \ - merge_dicts, pin_in_object_store, UtilMonitor, validate_save_restore, \ - warn_if_slow + merge_dicts, pin_in_object_store, unflattened_lookup, UtilMonitor, \ + validate_save_restore, warn_if_slow __all__ = [ "deep_update", @@ -8,6 +8,7 @@ "get_pinned_object", "merge_dicts", "pin_in_object_store", + "unflattened_lookup", "UtilMonitor", "validate_save_restore", "warn_if_slow", diff --git a/python/ray/tune/utils/util.py b/python/ray/tune/utils/util.py --- a/python/ray/tune/utils/util.py +++ b/python/ray/tune/utils/util.py @@ -2,7 +2,7 @@ import logging import threading import time -from collections import defaultdict +from collections import defaultdict, deque, Mapping, Sequence from threading import Thread import numpy as np @@ -216,6 +216,27 @@ def flatten_dict(dt, delimiter="/"): return dt +def unflattened_lookup(flat_key, lookup, delimiter="/", default=None): + """ + Unflatten `flat_key` and iteratively look up in `lookup`. E.g. + `flat_key="a/0/b"` will try to return `lookup["a"][0]["b"]`. + """ + keys = deque(flat_key.split(delimiter)) + base = lookup + while keys: + key = keys.popleft() + try: + if isinstance(base, Mapping): + base = base[key] + elif isinstance(base, Sequence): + base = base[int(key)] + else: + raise KeyError() + except KeyError: + return default + return base + + def _to_pinnable(obj): """Converts obj to a form that can be pinned in object store memory.
diff --git a/python/ray/tune/tests/test_progress_reporter.py b/python/ray/tune/tests/test_progress_reporter.py --- a/python/ray/tune/tests/test_progress_reporter.py +++ b/python/ray/tune/tests/test_progress_reporter.py @@ -22,15 +22,15 @@ EXPECTED_RESULT_2 = """Result logdir: /foo Number of trials: 5 (1 PENDING, 3 RUNNING, 1 TERMINATED) -+--------------+------------+-------+-----+-----+ -| Trial name | status | loc | a | b | -|--------------+------------+-------+-----+-----| -| 00000 | TERMINATED | here | 0 | 0 | -| 00001 | PENDING | here | 1 | 2 | -| 00002 | RUNNING | here | 2 | 4 | -| 00003 | RUNNING | here | 3 | 6 | -| 00004 | RUNNING | here | 4 | 8 | -+--------------+------------+-------+-----+-----+""" ++--------------+------------+-------+-----+-----+---------+---------+ +| Trial name | status | loc | a | b | n/k/0 | n/k/1 | +|--------------+------------+-------+-----+-----+---------+---------| +| 00000 | TERMINATED | here | 0 | 0 | 0 | 0 | +| 00001 | PENDING | here | 1 | 2 | 1 | 2 | +| 00002 | RUNNING | here | 2 | 4 | 2 | 4 | +| 00003 | RUNNING | here | 3 | 6 | 3 | 6 | +| 00004 | RUNNING | here | 4 | 8 | 4 | 8 | ++--------------+------------+-------+-----+-----+---------+---------+""" EXPECTED_RESULT_3 = """Result logdir: /foo Number of trials: 5 (1 PENDING, 3 RUNNING, 1 TERMINATED) @@ -246,21 +246,29 @@ def testProgressStr(self): t.trial_id = "%05d" % i t.local_dir = "/foo" t.location = "here" - t.config = {"a": i, "b": i * 2} - t.evaluated_params = t.config + t.config = {"a": i, "b": i * 2, "n": {"k": [i, 2 * i]}} + t.evaluated_params = { + "a": i, + "b": i * 2, + "n/k/0": i, + "n/k/1": 2 * i + } t.last_result = { "config": { "a": i, - "b": i * 2 + "b": i * 2, + "n": { + "k": [i, 2 * i] + } }, "metric_1": i / 2, "metric_2": i / 4 } t.__str__ = lambda self: self.trial_id trials.append(t) - # One metric, all parameters + # One metric, two parameters prog1 = trial_progress_str( - trials, ["metric_1"], None, fmt="psql", max_rows=3) + trials, ["metric_1"], ["a", "b"], fmt="psql", max_rows=3) print(prog1) assert prog1 == EXPECTED_RESULT_1
[Tune] status tables during training aren't showing values for tunable, nested config parameters ### What is the problem? Ray 0.8.6. Python 3.7.7. Running this script (actually run with ipython): ```python from ray import tune ray.init() tune.run( "PPO", stop={"episode_reward_mean": 400}, config={ "env": "CartPole-v1", "num_gpus": 0, "num_workers": 1, # "lr": tune.grid_search([0.01, 0.001, 0.0001]), "model": { 'fcnet_hiddens': [tune.grid_search([20, 40, 60, 80]), tune.grid_search([20, 40, 60, 80])] }, "eager": False, }, ) ``` It appears that Tune easily drives the 16 different combinations for the two elements of `fcnet_hidden` (number of weights for two NN layers). All combinations easily reach the reward of 400, but some do so with lower iterations and/or elapsed times, which is what I want to determine. The bug is in the status table printed. It has two columns for the tuning parameters, `model/fcnet_hiddens/0` and `model/fcnet_hiddens/1`, but they are empty. I can determine the actual values by looking at the `experiment_tag`, e.g., ` experiment_tag: 8_fcnet_hiddens_0=20,fcnet_hiddens_1=60`
Taking a look at this this week. I can reproduce the issue locally. ``` == Status == Memory usage on this node: 10.5/16.0 GiB Using FIFO scheduling algorithm. Resources requested: 8/8 CPUs, 0/0 GPUs, 0.0/4.49 GiB heap, 0.0/1.51 GiB objects Number of trials: 16 (4 RUNNING, 12 TERMINATED) +-----------------------------+------------+--------------------+-------------------------+-------------------------+--------+------------------+--------+----------+ | Trial name | status | loc | model/fcnet_hiddens/0 | model/fcnet_hiddens/1 | iter | total time (s) | ts | reward | |-----------------------------+------------+--------------------+-------------------------+-------------------------+--------+------------------+--------+----------| | PPO_CartPole-v1_e7acd_00000 | TERMINATED | | | | 18 | 220.755 | 72000 | 406.54 | | PPO_CartPole-v1_e7acd_00001 | TERMINATED | | | | 31 | 371.998 | 124000 | 403.33 | | PPO_CartPole-v1_e7acd_00002 | TERMINATED | | | | 18 | 220.732 | 72000 | 406.52 | | PPO_CartPole-v1_e7acd_00003 | TERMINATED | | | | 15 | 189.082 | 60000 | 409.03 | | PPO_CartPole-v1_e7acd_00004 | TERMINATED | | | | 19 | 233.262 | 76000 | 410.45 | | PPO_CartPole-v1_e7acd_00005 | TERMINATED | | | | 22 | 260.052 | 88000 | 412.14 | | PPO_CartPole-v1_e7acd_00006 | TERMINATED | | | | 30 | 354.216 | 120000 | 401.67 | | PPO_CartPole-v1_e7acd_00007 | TERMINATED | | | | 18 | 211.58 | 72000 | 403.37 | | PPO_CartPole-v1_e7acd_00008 | TERMINATED | | | | 19 | 222.288 | 76000 | 416.59 | | PPO_CartPole-v1_e7acd_00009 | TERMINATED | | | | 19 | 218.809 | 76000 | 406.62 | | PPO_CartPole-v1_e7acd_00010 | TERMINATED | | | | 16 | 531.601 | 64000 | 400.38 | | PPO_CartPole-v1_e7acd_00011 | TERMINATED | | | | 21 | 4509.18 | 84000 | 415.27 | | PPO_CartPole-v1_e7acd_00012 | RUNNING | 192.168.2.66:49528 | | | 16 | 4449.35 | 64000 | 383.28 | | PPO_CartPole-v1_e7acd_00013 | RUNNING | 192.168.2.66:49553 | | | 12 | 4406.25 | 48000 | 292.06 | | PPO_CartPole-v1_e7acd_00014 | RUNNING | 192.168.2.66:49578 | | | 5 | 3502.22 | 20000 | 110.89 | | PPO_CartPole-v1_e7acd_00015 | RUNNING | | | | | | | | +-----------------------------+------------+--------------------+-------------------------+-------------------------+--------+------------------+--------+----------+ ``` the trial config contains the `model/fcnet_hiddens` key e.g. `{'env': 'CartPole-v1', 'num_gpus': 0, 'num_workers': 1, 'eager': False, 'model/fcnet_hiddens': [20, 20]}` whereas `trial.evaluated_params` contains the flattened keys e.g. `{'model/fcnet_hiddens/0': 20, 'model/fcnet_hiddens/1': 20}` instead. `progress_reporter` uses `trial.evaluated_params` to generate the header and looks into `trial.config` for the config values
2020-07-16T18:14:09
ray-project/ray
9,537
ray-project__ray-9537
[ "8913" ]
78dfed268391fb5ea927a5094d9c171b9da8c052
diff --git a/python/ray/tune/progress_reporter.py b/python/ray/tune/progress_reporter.py --- a/python/ray/tune/progress_reporter.py +++ b/python/ray/tune/progress_reporter.py @@ -87,7 +87,7 @@ def __init__(self, max_progress_rows=20, max_error_rows=20, max_report_frequency=5): - self._metric_columns = metric_columns or self.DEFAULT_COLUMNS + self._metric_columns = metric_columns or self.DEFAULT_COLUMNS.copy() self._parameter_columns = parameter_columns or [] self._max_progress_rows = max_progress_rows self._max_error_rows = max_error_rows
[Tune] ProgressReporter default_columns is shared across instances ### What is the problem? When adding a new column to a ProgressReporter that used the `DEFAULT_COLUMNS` during its creation, this column will be added to the actual default dictionary. This can cause several unexpected behaviors. For instance, when creating several experiments with custom columns for each of them, using the same column multiple times will break (see reproduction). ### Reproduction (REQUIRED) ```python from ray.tune import CLIReporter reporter = CLIReporter() reporter.add_metric_column('custom column') reporter = CLIReporter() reporter.add_metric_column('custom column') ``` Thows a `ValueError: Column custom column already exists.` - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html). ### Solution A solution is to copy this default parameter each time it is used [here](https://github.com/ray-project/ray/blob/4b31b383f379dad5514f4d93a9776db6b3caa38c/python/ray/tune/progress_reporter.py#L90). I can send a PR is this is okay for you.
Sorry; I totally missed your comment at the bottom. A PR would be great!
2020-07-17T10:29:17
ray-project/ray
9,547
ray-project__ray-9547
[ "9541" ]
bfa06052828f83ab790e6b6bbfa5b56edb42b45e
diff --git a/python/ray/node.py b/python/ray/node.py --- a/python/ray/node.py +++ b/python/ray/node.py @@ -394,8 +394,10 @@ def _make_inc_temp(self, suffix="", prefix="", directory_name=None): raise FileExistsError(errno.EEXIST, "No usable temporary filename found") - def get_log_file_names(self, name, unique=False): - """Generate partially randomized filenames for log files. + def get_log_file_handles(self, name, unique=False): + """Open log files with partially randomized filenames, returning the + file handles. If output redirection has been disabled, no files will + be opened and `(None, None)` will be returned. Args: name (str): descriptive string for this log file. @@ -403,7 +405,8 @@ def get_log_file_names(self, name, unique=False): ensure the returned filename is not already used. Returns: - A tuple of two file names for redirecting (stdout, stderr). + A tuple of two file handles for redirecting (stdout, stderr), or + `(None, None)` if output redirection is disabled. """ redirect_output = self._ray_params.redirect_output @@ -414,6 +417,21 @@ def get_log_file_names(self, name, unique=False): if not redirect_output: return None, None + log_stdout, log_stderr = self._get_log_file_names(name, unique=unique) + return open_log(log_stdout), open_log(log_stderr) + + def _get_log_file_names(self, name, unique=False): + """Generate partially randomized filenames for log files. + + Args: + name (str): descriptive string for this log file. + unique (bool): if true, a counter will be attached to `name` to + ensure the returned filename is not already used. + + Returns: + A tuple of two file names for redirecting (stdout, stderr). + """ + if unique: log_stdout = self._make_inc_temp( suffix=".out", prefix=name, directory_name=self._logs_dir) @@ -501,15 +519,11 @@ def start_reaper_process(self): def start_redis(self): """Start the Redis servers.""" assert self._redis_address is None - redis_out_name, redis_err_name = self.get_log_file_names( - "redis", unique=True) - redis_log_files = [(open_log(redis_out_name), - open_log(redis_err_name))] + redis_log_files = [self.get_log_file_handles("redis", unique=True)] for i in range(self._ray_params.num_redis_shards): - shard_out_name, shard_err_name = self.get_log_file_names( - "redis-shard_{}".format(i), unique=True) - redis_log_files.append((open_log(shard_out_name), - open_log(shard_err_name))) + redis_log_files.append( + self.get_log_file_handles( + "redis-shard_{}".format(i), unique=True)) (self._redis_address, redis_shards, process_infos) = ray.services.start_redis( @@ -531,10 +545,8 @@ def start_redis(self): def start_log_monitor(self): """Start the log monitor.""" - log_out_name, log_err_name = self.get_log_file_names( + stdout_file, stderr_file = self.get_log_file_handles( "log_monitor", unique=True) - stdout_file, stderr_file = open_log(log_out_name), open_log( - log_err_name) process_info = ray.services.start_log_monitor( self.redis_address, self._logs_dir, @@ -549,10 +561,8 @@ def start_log_monitor(self): def start_reporter(self): """Start the reporter.""" - reporter_out_name, reporter_err_name = self.get_log_file_names( + stdout_file, stderr_file = self.get_log_file_handles( "reporter", unique=True) - stdout_file, stderr_file = (open_log(reporter_out_name), - open_log(reporter_err_name)) process_info = ray.services.start_reporter( self.redis_address, self._ray_params.metrics_agent_port, @@ -574,10 +584,8 @@ def start_dashboard(self, require_dashboard): if we fail to start the dashboard. Otherwise it will print a warning if we fail to start the dashboard. """ - dashboard_out_name, dashboard_err_name = self.get_log_file_names( + stdout_file, stderr_file = self.get_log_file_handles( "dashboard", unique=True) - stdout_file, stderr_file = (open_log(dashboard_out_name), - open_log(dashboard_err_name)) self._webui_url, process_info = ray.services.start_dashboard( require_dashboard, self._ray_params.dashboard_host, @@ -598,10 +606,8 @@ def start_dashboard(self, require_dashboard): def start_plasma_store(self): """Start the plasma store.""" - plasma_out_name, plasma_err_name = self.get_log_file_names( + stdout_file, stderr_file = self.get_log_file_handles( "plasma_store", unique=True) - stdout_file, stderr_file = (open_log(plasma_out_name), - open_log(plasma_err_name)) process_info = ray.services.start_plasma_store( self.get_resource_spec(), self._plasma_store_socket_name, @@ -620,10 +626,8 @@ def start_plasma_store(self): def start_gcs_server(self): """Start the gcs server. """ - gcs_out_name, gcs_err_name = self.get_log_file_names( + stdout_file, stderr_file = self.get_log_file_handles( "gcs_server", unique=True) - stdout_file, stderr_file = (open_log(gcs_out_name), - open_log(gcs_err_name)) process_info = ray.services.start_gcs_server( self._redis_address, stdout_file=stdout_file, @@ -647,10 +651,8 @@ def start_raylet(self, use_valgrind=False, use_profiler=False): use_profiler (bool): True if we should start the process in the valgrind profiler. """ - raylet_out_name, raylet_err_name = self.get_log_file_names( + stdout_file, stderr_file = self.get_log_file_handles( "raylet", unique=True) - stdout_file, stderr_file = (open_log(raylet_out_name), - open_log(raylet_err_name)) process_info = ray.services.start_raylet( self._redis_address, self._node_ip_address, @@ -713,7 +715,7 @@ def get_job_redirected_log_file(self, else: name = "worker-{}".format(ray.utils.binary_to_hex(worker_id)) - worker_stdout_file, worker_stderr_file = self.get_log_file_names( + worker_stdout_file, worker_stderr_file = self._get_log_file_names( name, unique=False) return worker_stdout_file, worker_stderr_file @@ -723,10 +725,8 @@ def start_worker(self): def start_monitor(self): """Start the monitor.""" - monitor_out_name, monitor_err_name = self.get_log_file_names( + stdout_file, stderr_file = self.get_log_file_handles( "monitor", unique=True) - stdout_file, stderr_file = (open_log(monitor_out_name), - open_log(monitor_err_name)) process_info = ray.services.start_monitor( self._redis_address, stdout_file=stdout_file, diff --git a/python/ray/utils.py b/python/ray/utils.py --- a/python/ray/utils.py +++ b/python/ray/utils.py @@ -391,6 +391,10 @@ def setup_logger(logging_level, logging_format): def open_log(path, **kwargs): + """ + Opens the log file at `path`, with the provided kwargs being given to + `open`. + """ kwargs.setdefault("buffering", 1) kwargs.setdefault("mode", "a") kwargs.setdefault("encoding", "utf-8")
[Core] Turning off log redirection to files breaks all Ray service startups. <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? Turning off log redirection to files breaks all Ray process startups (redis, GCS server, log monitor, reporter, plasma store, raylet, etc.) ``` File "/app/descarteslabs/services/workbench/deploy/head-image-dev.binary.runfiles/requirements_py3_pypi__ray_0_9_0_dev0/ray-0.9.0.dev0.data/purelib/ray/scripts/scripts.py", line 470, in start ray_params, head=True, shutdown_at_exit=block, spawn_reaper=block) File "/app/descarteslabs/services/workbench/deploy/head-image-dev.binary.runfiles/requirements_py3_pypi__ray_0_9_0_dev0/ray-0.9.0.dev0.data/purelib/ray/node.py", line 194, in __init__ self.start_head_processes() File "/app/descarteslabs/services/workbench/deploy/head-image-dev.binary.runfiles/requirements_py3_pypi__ray_0_9_0_dev0/ray-0.9.0.dev0.data/purelib/ray/node.py", line 746, in start_head_processes self.start_redis() File "/app/descarteslabs/services/workbench/deploy/head-image-dev.binary.runfiles/requirements_py3_pypi__ray_0_9_0_dev0/ray-0.9.0.dev0.data/purelib/ray/node.py", line 506, in start_redis redis_log_files = [(open_log(redis_out_name), File "/app/descarteslabs/services/workbench/deploy/head-image-dev.binary.runfiles/requirements_py3_pypi__ray_0_9_0_dev0/ray-0.9.0.dev0.data/purelib/ray/utils.py", line 397, in open_log return open(path, **kwargs) TypeError: expected str, bytes or os.PathLike object, not NoneType ``` *Ray version and other system information (Python version, TensorFlow version, OS):* This bug was introduced in [this PR](https://github.com/ray-project/ray/pull/8885), so it shouldn't be affecting any releases. The issue is that `get_log_file_names` will return [`None, None` if output redirection is turned off](https://github.com/ray-project/ray/blob/f080aa6ce33ed34f686e868d97b85632c8a8f894/python/ray/node.py#L414-L415), while all callers of that function now pass the returned log file names directly to `open_log`, e.g. [here](https://github.com/ray-project/ray/blob/f080aa6ce33ed34f686e868d97b85632c8a8f894/python/ray/node.py#L506-L507). ### Reproduction: Start Ray with the environment variable `GLOG_logtostderr` set to 1, or with `redirect_output` or `redirect_worker_output` set to `False`. I've confirmed that this bug exists on the latest build off master. If we cannot run your script, we cannot fix your issue. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
I'll probably be submitting a fix soon. Awesome! Thanks @clarkzinzow !!
2020-07-17T17:50:29
ray-project/ray
9,561
ray-project__ray-9561
[ "9553" ]
cf719dd470f61918f0de8550c885360a78497e59
diff --git a/python/ray/node.py b/python/ray/node.py --- a/python/ray/node.py +++ b/python/ray/node.py @@ -647,7 +647,8 @@ def start_raylet(self, use_valgrind=False, use_profiler=False): use_profiler (bool): True if we should start the process in the valgrind profiler. """ - raylet_out_name, raylet_err_name = self.get_log_file_names("raylet") + raylet_out_name, raylet_err_name = self.get_log_file_names( + "raylet", unique=True) stdout_file, stderr_file = (open_log(raylet_out_name), open_log(raylet_err_name)) process_info = ray.services.start_raylet(
[core] Raylets in local cluster do not log to separate files <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? *Ray version and other system information (Python version, TensorFlow version, OS):* 0.9dev When multiple raylets are started locally, it used to be that they would each log to a separate file, e.g., `raylet.1.out`, `raylet.2.out`, etc. Now, they all log to `raylet.out`, which makes it harder to debug. ### Reproduction (REQUIRED) Any checked in pytest that starts a local cluster. For example, `RAY_BACKEND_LOG_LEVEL=debug pytest -sv python/ray/tests/test_advanced.py::test_wait_cluster` - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
I believe this was caused by #8885, @wuisawesome. @stephanie-wang That was another regression I noticed, adding a `unique=True` argument to `get_log_file_names` [here](https://github.com/ray-project/ray/blob/5881417ec4a993194ba6d64b156ad6b5ffbca97c/python/ray/node.py#L650) should do the trick: ``` raylet_out_name, raylet_err_name = self.get_log_file_names("raylet", unique=True) ``` > @stephanie-wang That was another regression I noticed, adding a `unique=True` argument to `get_log_file_names` [here](https://github.com/ray-project/ray/blob/5881417ec4a993194ba6d64b156ad6b5ffbca97c/python/ray/node.py#L650) should do the trick: > > ``` > raylet_out_name, raylet_err_name = self.get_log_file_names("raylet", unique=True) > ``` Ah, nice find! Nice catch, I may have made this same mistake in a few places. @clarkzinzow do you want to submit a PR for this? If not, I'll try to get around to it in a week or so. @wuisawesome Sure I can submit a PR for that, I'm already in this code for [another PR](https://github.com/ray-project/ray/pull/9547) so it's not much of a context switch.
2020-07-18T00:54:34
ray-project/ray
9,572
ray-project__ray-9572
[ "9568" ]
ce3f5427394af288f9699bed3d479c47181b252f
diff --git a/rllib/train.py b/rllib/train.py --- a/rllib/train.py +++ b/rllib/train.py @@ -180,22 +180,23 @@ def run(args, parser): parser.error("the following arguments are required: --run") if not exp.get("env") and not exp.get("config", {}).get("env"): parser.error("the following arguments are required: --env") - if args.eager: - exp["config"]["framework"] = "tfe" - elif args.torch: + + if args.torch: exp["config"]["framework"] = "torch" - else: - exp["config"]["framework"] = "tf" + elif args.eager: + exp["config"]["framework"] = "tfe" + + if args.trace: + if exp["config"]["framework"] not in ["tf2", "tfe"]: + raise ValueError("Must enable --eager to enable tracing.") + exp["config"]["eager_tracing"] = True + if args.v: exp["config"]["log_level"] = "INFO" verbose = 2 if args.vv: exp["config"]["log_level"] = "DEBUG" verbose = 3 - if args.trace: - if exp["config"]["framework"] != "tfe": - raise ValueError("Must enable --eager to enable tracing.") - exp["config"]["eager_tracing"] = True if args.ray_num_nodes: cluster = Cluster()
[rllib] Train script overwrites framework field in experiment config argument [rllib] ### What is the problem? Framework setting passed to training script via config argument is ignored, because training script uses separate arguments for this field (`--torch`, `--eager`). If none of these args is passed to the script the framework is set to `tf`, regardless of options passed in `--config` argument. ### Expected behavior Script should use default value for `framework` field only if it is not already present in `config` dict (it can be passed as cmd argument or be loaded from file). ### Reproduction (REQUIRED) Run `rllib train --run DQN --env CartPole-v0 --config "{\"framework\": \"tfe\"}"` - script uses `tf` instead of `tfe` - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-07-19T18:59:50
ray-project/ray
9,966
ray-project__ray-9966
[ "9967" ]
84b7240c4bcbd9373ec18db8be483a3e41a4f02a
diff --git a/rllib/models/tf/visionnet.py b/rllib/models/tf/visionnet.py --- a/rllib/models/tf/visionnet.py +++ b/rllib/models/tf/visionnet.py @@ -11,15 +11,17 @@ class VisionNetwork(TFModelV2): def __init__(self, obs_space, action_space, num_outputs, model_config, name): + if not model_config.get("conv_filters"): + model_config["conv_filters"] = _get_filter_config(obs_space.shape) + super(VisionNetwork, self).__init__(obs_space, action_space, num_outputs, model_config, name) - activation = get_activation_fn(model_config.get("conv_activation")) - filters = model_config.get("conv_filters") - if not filters: - filters = _get_filter_config(obs_space.shape) - no_final_linear = model_config.get("no_final_linear") - vf_share_layers = model_config.get("vf_share_layers") + activation = get_activation_fn( + self.model_config.get("conv_activation"), framework="tf") + filters = self.model_config["conv_filters"] + no_final_linear = self.model_config.get("no_final_linear") + vf_share_layers = self.model_config.get("vf_share_layers") inputs = tf.keras.layers.Input( shape=obs_space.shape, name="observations") @@ -73,6 +75,16 @@ def __init__(self, obs_space, action_space, num_outputs, model_config, padding="same", data_format="channels_last", name="conv_out")(last_layer) + + if conv_out.shape[1] != 1 or conv_out.shape[2] != 1: + raise ValueError( + "Given `conv_filters` ({}) do not result in a [B, 1, " + "1, {} (`num_outputs`)] shape (but in {})! Please " + "adjust your Conv2D stack such that the dims 1 and 2 " + "are both 1.".format( + self.model_config["conv_filters"], + self.num_outputs, list(conv_out.shape))) + # num_outputs not known -> Flatten, then set self.num_outputs # to the resulting number of nodes. else: diff --git a/rllib/models/torch/visionnet.py b/rllib/models/torch/visionnet.py --- a/rllib/models/torch/visionnet.py +++ b/rllib/models/torch/visionnet.py @@ -15,17 +15,18 @@ class VisionNetwork(TorchModelV2, nn.Module): def __init__(self, obs_space, action_space, num_outputs, model_config, name): + if not model_config.get("conv_filters"): + model_config["conv_filters"] = _get_filter_config(obs_space.shape) + TorchModelV2.__init__(self, obs_space, action_space, num_outputs, model_config, name) nn.Module.__init__(self) activation = get_activation_fn( - model_config.get("conv_activation"), framework="torch") - filters = model_config.get("conv_filters") - if not filters: - filters = _get_filter_config(obs_space.shape) - no_final_linear = model_config.get("no_final_linear") - vf_share_layers = model_config.get("vf_share_layers") + self.model_config.get("conv_activation"), framework="torch") + filters = self.model_config["conv_filters"] + no_final_linear = self.model_config.get("no_final_linear") + vf_share_layers = self.model_config.get("vf_share_layers") # Whether the last layer is the output of a Flattened (rather than # a n x (1,1) Conv2D). @@ -152,8 +153,17 @@ def forward(self, input_dict, state, seq_lens): if not self.last_layer_is_flattened: if self._logits: conv_out = self._logits(conv_out) + if conv_out.shape[2] != 1 or conv_out.shape[3] != 1: + raise ValueError( + "Given `conv_filters` ({}) do not result in a [B, {} " + "(`num_outputs`), 1, 1] shape (but in {})! Please adjust " + "your Conv2D stack such that the last 2 dims are both " + "1.".format( + self.model_config["conv_filters"], self.num_outputs, + list(conv_out.shape))) logits = conv_out.squeeze(3) logits = logits.squeeze(2) + return logits, state else: return conv_out, state
[RLlib] Add informative error message when bad Conv2D stack is used with fixed `num_outputs` (no flattening at end). We should add an informative error message to visionnet.py (torch and tf), when a bad Conv2D stack (`conv_filters` in model config) is configured by the user, with a fixed `num_outputs` (no flattening at end). In this case, the output of the last Conv2D layer must always be [B, num_outputs, 1, 1]. ``` import gym import numpy as np import ray from ray import tune from ray.rllib.models import ModelCatalog from ray.tune.registry import register_env from ray.rllib.agents.ppo import PPOTrainer, ppo from ray.tune import CLIReporter from ray.rllib.models.torch.torch_modelv2 import TorchModelV2 from ray.rllib.models.torch.fcnet import FullyConnectedNetwork from ray.rllib.models.torch.visionnet import VisionNetwork from ray.rllib.utils.framework import try_import_torch ray.init() torch, nn = try_import_torch() #### ENV class JSS(gym.Env): def __init__(self, env_config): self.problem_size = 16 self.action_space = gym.spaces.Discrete(self.problem_size + 1) self.observation_space = gym.spaces.Dict({ 'real_obs': gym.spaces.Box(low=0.0, high=1.0, shape=(self.problem_size, 8, 1), dtype=np.float), 'action_mask': gym.spaces.Box(low=0, high=1, shape=(self.action_space.n, ), dtype=np.int), }) self.i = 0 def reset(self): self.i = 0 legal_action = np.ones((self.action_space.n,), dtype=np.int) legal_action[0] = 0 output = { 'real_obs': np.zeros((self.problem_size, 8, 1), dtype=np.float), 'action_mask': legal_action } return output def step(self, action: int): self.i += 1 legal_action = np.ones((self.action_space.n, ), dtype=np.int) legal_action[self.problem_size] = 0 legal_action[self.i % self.problem_size + 1] = 0 output = { 'real_obs': np.zeros((self.problem_size, 8, 1), dtype=np.float), 'action_mask': legal_action } return output def render(self, mode='human'): pass def env_creator(env_config): return JSS(env_config) register_env("jss_env", env_creator) ## MODEL # THIS ONE WORKS class FCMaskedActionsModel(TorchModelV2, nn.Module): def __init__(self, obs_space, action_space, num_outputs, model_config, name): nn.Module.__init__(self) super(FCMaskedActionsModel, self).__init__(obs_space, action_space, action_space.n, model_config, name) true_obs_space = gym.spaces.MultiBinary(n=obs_space.shape[0] - action_space.n) self.action_model = FullyConnectedNetwork( obs_space=true_obs_space, action_space=action_space, num_outputs=action_space.n, model_config=model_config, name=name + "action_model") def forward(self, input_dict, state, seq_lens): # Extract the available actions tensor from the observation. action_mask = input_dict["obs"]["action_mask"] # Compute the predicted action embedding raw_actions, state = self.action_model({ "obs": input_dict["obs"]["real_obs"] }) inf_mask = torch.clamp(torch.log(action_mask), min=torch.finfo(raw_actions.dtype).min) actions = raw_actions + inf_mask return actions, state def value_function(self): return self.action_model.value_function() #THIS ONE HAS A SIZE MISMATCH PROBLEM WITH THE LOGIT LAYER class ConvMaskedActionsModel(TorchModelV2, nn.Module): def __init__(self, obs_space, action_space, num_outputs, model_config, name): nn.Module.__init__(self) super(ConvMaskedActionsModel, self).__init__(obs_space, action_space, action_space.n, model_config, name) true_obs_space = obs_space.original_space['real_obs'] self.action_model = VisionNetwork( obs_space=true_obs_space, action_space=action_space, num_outputs=action_space.n, model_config=model_config, name=name + "action_model_conv") def forward(self, input_dict, state, seq_lens): # Extract the available actions tensor from the observation. action_mask = input_dict["obs"]["action_mask"] # Compute the predicted action embedding raw_actions, state = self.action_model({ "obs": input_dict["obs"]["real_obs"] }) inf_mask = torch.clamp(torch.log(action_mask), min=torch.finfo(raw_actions.dtype).min) actions = raw_actions + inf_mask return actions, state def value_function(self): return self.action_model.value_function() ModelCatalog.register_custom_model("fc_masked_model", FCMaskedActionsModel) ModelCatalog.register_custom_model("conv_masked_model", ConvMaskedActionsModel) config = ppo.DEFAULT_CONFIG.copy() config['framework'] = 'torch' config['env'] = 'jss_env' config['model'] = { "conv_filters": [[4, [4, 4], 1], [8, [4, 4], 1]], "custom_model": "conv_masked_model", # Nonlinearity for fully connected net (tanh, relu) "fcnet_activation": "tanh", } config['train_batch_size'] = 1024 config['rollout_fragment_length'] = 1000 config['sgd_minibatch_size'] = 256 config['lr'] = 1e-4 config['evaluation_interval'] = None config['metrics_smoothing_episodes'] = 1000 stop = { "time_total_s": 60, } reporter = CLIReporter(max_progress_rows=15) reporter.add_metric_column("episode_reward_max") analysis = tune.run(PPOTrainer, config=config, stop=stop, progress_reporter=reporter) print("Best config: ", analysis.get_best_config(metric="episode_reward_max")) ray.shutdown() ```
2020-08-06T20:15:58
ray-project/ray
10,078
ray-project__ray-10078
[ "10077" ]
739933e5b84291dea54c29fc6acc982a3d42a52f
diff --git a/python/ray/autoscaler/command_runner.py b/python/ray/autoscaler/command_runner.py --- a/python/ray/autoscaler/command_runner.py +++ b/python/ray/autoscaler/command_runner.py @@ -556,6 +556,7 @@ def run( ssh_options_override_ssh_key=ssh_options_override_ssh_key) def run_rsync_up(self, source, target): + # TODO(ilr) Expose this to before NodeUpdater::sync_file_mounts protected_path = target if target.find("/root") == 0: target = target.replace("/root", "/tmp/root") diff --git a/python/ray/autoscaler/updater.py b/python/ray/autoscaler/updater.py --- a/python/ray/autoscaler/updater.py +++ b/python/ray/autoscaler/updater.py @@ -145,8 +145,9 @@ def do_sync(remote_path, local_path, allow_non_existing_paths=False): with LogTimer(self.log_prefix + "Synced {} to {}".format(local_path, remote_path)): - self.cmd_runner.run("mkdir -p {}".format( - os.path.dirname(remote_path))) + self.cmd_runner.run( + "mkdir -p {}".format(os.path.dirname(remote_path)), + run_env="host") sync_cmd(local_path, remote_path) if remote_path not in nolog_paths: @@ -188,7 +189,7 @@ def wait_ready(self, deadline): "{}Waiting for remote shell...", self.log_prefix) - self.cmd_runner.run("uptime") + self.cmd_runner.run("uptime", run_env="host") cli_logger.old_debug(logger, "Uptime succeeded.") cli_logger.success("Success.") return True
Docker + AWS fails with no such container error Latest dev version of ray ``` (vanilla_ray_venv) richard@richard-desktop:~/improbable/vanillas/ray/python/ray/autoscaler/aws$ ray up aws_gpu_dummy.yaml 2020-08-12 20:12:39,383 INFO config.py:268 -- _configure_iam_role: Role not specified for head node, using arn:aws:iam::179622923911:instance-profile/ray-autoscaler-v1 2020-08-12 20:12:39,612 INFO config.py:346 -- _configure_key_pair: KeyName not specified for nodes, using ray-autoscaler_us-east-1 2020-08-12 20:12:39,745 INFO config.py:407 -- _configure_subnet: SubnetIds not specified for head node, using [('subnet-f737f791', 'us-east-1a')] 2020-08-12 20:12:39,746 INFO config.py:417 -- _configure_subnet: SubnetId not specified for workers, using [('subnet-f737f791', 'us-east-1a')] 2020-08-12 20:12:40,358 INFO config.py:590 -- _create_security_group: Created new security group ray-autoscaler-richard_cluster_gpu_dummy (sg-0061ca6aff182c1bf) 2020-08-12 20:12:40,739 INFO config.py:444 -- _configure_security_group: SecurityGroupIds not specified for head node, using ray-autoscaler-richard_cluster_gpu_dummy (sg-0061ca6aff182c1bf) 2020-08-12 20:12:40,739 INFO config.py:454 -- _configure_security_group: SecurityGroupIds not specified for workers, using ray-autoscaler-richard_cluster_gpu_dummy (sg-0061ca6aff182c1bf) This will create a new cluster [y/N]: y 2020-08-12 20:12:42,619 INFO commands.py:531 -- get_or_create_head_node: Launching new head node... 2020-08-12 20:12:42,620 INFO node_provider.py:326 -- NodeProvider: calling create_instances with subnet-f737f791 (count=1). 2020-08-12 20:12:44,032 INFO node_provider.py:354 -- NodeProvider: Created instance [id=i-0729c7a86355d5ff8, name=pending, info=pending] 2020-08-12 20:12:44,223 INFO commands.py:570 -- get_or_create_head_node: Updating files on head node... 2020-08-12 20:12:44,320 INFO command_runner.py:331 -- NodeUpdater: i-0729c7a86355d5ff8: Waiting for IP... 2020-08-12 20:12:54,409 INFO command_runner.py:331 -- NodeUpdater: i-0729c7a86355d5ff8: Waiting for IP... 2020-08-12 20:12:54,534 INFO log_timer.py:27 -- NodeUpdater: i-0729c7a86355d5ff8: Got IP [LogTimer=10310ms] 2020-08-12 20:12:54,534 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && command -v docker' Warning: Permanently added '3.226.253.119' (ECDSA) to the list of known hosts. /usr/bin/docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:14:04,587 INFO updater.py:71 -- NodeUpdater: i-0729c7a86355d5ff8: Updating to 6b5fc8ee8c5dcdf3cfabe0bf90ba4e844f65a7c9 2020-08-12 20:14:04,587 INFO updater.py:180 -- NodeUpdater: i-0729c7a86355d5ff8: Waiting for remote shell... 2020-08-12 20:14:04,587 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' 2020-08-12 20:14:04,950 INFO log_timer.py:27 -- AWSNodeProvider: Set tag ray-node-status=waiting-for-ssh on ['i-0729c7a86355d5ff8'] [LogTimer=361ms] Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:14:21,222 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:14:26,417 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:14:31,610 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:14:36,798 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:14:41,986 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:14:47,170 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:14:52,358 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:14:57,554 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:15:02,750 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:15:07,938 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:15:13,126 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:15:18,307 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:15:23,494 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. Shared connection to 3.226.253.119 closed. 2020-08-12 20:19:01,502 INFO command_runner.py:468 -- NodeUpdater: i-0729c7a86355d5ff8: Running ssh -tt -i /home/richard/.ssh/ray-autoscaler_us-east-1.pem -o StrictHostKeyChecking=no -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ExitOnForwardFailure=yes -o ServerAliveInterval=5 -o ServerAliveCountMax=3 -o ControlMaster=auto -o ControlPath=/tmp/ray_ssh_6ae199a93c/cfde1c79f1/%C -o ControlPersist=10s -o ConnectTimeout=120s [email protected] bash --login -c -i 'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && docker exec -it pytorch_docker /bin/bash -c '"'"'bash --login -c -i '"'"'"'"'"'"'"'"'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && uptime'"'"'"'"'"'"'"'"''"'"' ' Error: No such container: pytorch_docker Shared connection to 3.226.253.119 closed. 2020-08-12 20:19:06,689 INFO log_timer.py:27 -- NodeUpdater: i-0729c7a86355d5ff8: Got remote shell [LogTimer=302102ms] 2020-08-12 20:19:06,690 INFO log_timer.py:27 -- NodeUpdater: i-0729c7a86355d5ff8: Applied config 6b5fc8ee8c5dcdf3cfabe0bf90ba4e844f65a7c9 [LogTimer=302103ms] 2020-08-12 20:19:06,690 ERROR updater.py:88 -- NodeUpdater: i-0729c7a86355d5ff8: Error executing: Unable to connect to node Exception in thread Thread-2: Traceback (most recent call last): File "/usr/lib/python3.6/threading.py", line 916, in _bootstrap_inner self.run() File "/home/richard/improbable/vanillas/ray/python/ray/autoscaler/updater.py", line 76, in run self.do_update() File "/home/richard/improbable/vanillas/ray/python/ray/autoscaler/updater.py", line 232, in do_update self.wait_ready(deadline) File "/home/richard/improbable/vanillas/ray/python/ray/autoscaler/updater.py", line 224, in wait_ready assert False, "Unable to connect to node" AssertionError: Unable to connect to node 2020-08-12 20:19:06,962 ERROR commands.py:650 -- get_or_create_head_node: Updating 3.226.253.119 failed 2020-08-12 20:19:07,002 INFO log_timer.py:27 -- AWSNodeProvider: Set tag ray-node-status=update-failed on ['i-0729c7a86355d5ff8'] [LogTimer=312ms] ``` YAML file for repo attached. ``` # An unique identifier for the head node and workers of this cluster. cluster_name: richard_cluster_gpu_dummy # The minimum number of workers nodes to launch in addition to the head # node. This number should be >= 0. min_workers: 1 # The maximum number of workers nodes to launch in addition to the head # node. This takes precedence over min_workers. max_workers: 5 # The initial number of worker nodes to launch in addition to the head # node. When the cluster is first brought up (or when it is refreshed with a # subsequent `ray up`) this number of nodes will be started. initial_workers: 1 # Whether or not to autoscale aggressively. If this is enabled, if at any point # we would start more workers, we start at least enough to bring us to # initial_workers. autoscaling_mode: default # This executes all commands on all nodes in the docker efcontainer, # and opens all the necessary ports to support the Ray cluster. # Empty string means disabled. docker: image: "pytorch/pytorch:latest" # e.g., tensorflow/tensorflow:1.5.0-py3 container_name: "pytorch_docker" # e.g. ray_docker # If true, pulls latest version of image. Otherwise, `docker run` will only pull the image # if no cached version is present. pull_before_run: True run_options: [] # - $([ -d /proc/driver ] && echo -n --runtime-nvidia) # Use the nvidia runtime only if nvidia gpu's are installed worker_run_options: - --runtime=nvidia # Extra options to pass into "docker run" # Example of running a GPU head with CPU workers # head_image: "tensorflow/tensorflow:1.13.1-py3" # head_run_options: # - --runtime=nvidia # worker_image: "ubuntu:18.04" # worker_run_options: [] # The autoscaler will scale up the cluster to this target fraction of resource # usage. For example, if a cluster of 10 nodes is 100% busy and # target_utilization is 0.8, it would resize the cluster to 13. This fraction # can be decreased to increase the aggressiveness of upscaling. # This value must be less than 1.0 for scaling to happen. target_utilization_fraction: 0.8 # If a node is idle for this many minutes, it will be removed. idle_timeout_minutes: 5 # Cloud-provider specific configuration. provider: type: aws region: us-east-1 # Availability zone(s), comma-separated, that nodes may be launched in. # Nodes are currently spread between zones by a round-robin approach, # however this implementation detail should not be relied upon. availability_zone: us-east-1a, us-east-1b cache_stopped_nodes: False # How Ray will authenticate with newly launched nodes. auth: ssh_user: ubuntu # By default Ray creates a new private keypair, but you can also use your own. # If you do so, make sure to also set "KeyName" in the head and worker node # configurations below. # ssh_private_key: /path/to/your/key.pem # Provider-specific config for the head node, e.g. instance type. By default # Ray will auto-configure unspecified fields such as SubnetId and KeyName. # For more documentation on available fields, see: # http://boto3.readthedocs.io/en/latest/reference/services/ec2.html#EC2.ServiceResource.create_instances head_node: InstanceType: c4.2xlarge ImageId: ami-043f9aeaf108ebc37 # Deep Learning AMI (Ubuntu) Version 24.3 # You can provision additional disk space with a conf as follows BlockDeviceMappings: - DeviceName: /dev/sda1 Ebs: VolumeSize: 100 # Additional options in the boto docs. # Provider-specific config for worker nodes, e.g. instance type. By default # Ray will auto-configure unspecified fields such as SubnetId and KeyName. # For more documentation on available fields, see: # http://boto3.readthedocs.io/en/latest/reference/services/ec2.html#EC2.ServiceResource.create_instances worker_nodes: InstanceType: p3.2xlarge ImageId: ami-043f9aeaf108ebc37 # Deep Learning AMI (Ubuntu) Version 24.3 # Run workers on spot by default. Comment this out to use on-demand. InstanceMarketOptions: MarketType: spot # Additional options can be found in the boto docs, e.g. # SpotOptions: # MaxPrice: MAX_HOURLY_PRICE # Additional options in the boto docs. # Files or directories to copy to the head and worker nodes. The format is a # dictionary from REMOTE_PATH: LOCAL_PATH, e.g. file_mounts: { } # List of commands that will be run before `setup_commands`. If docker is # enabled, these commands will run outside the container and before docker # is setup. initialization_commands: [] # List of shell commands to run to set up nodes. setup_commands: # Note: if you're developing Ray, you probably want to create an AMI that # has your Ray repo pre-cloned. Then, you can replace the pip installs # below with a git checkout <your_sha> (and possibly a recompile). - pip install -U https://s3-us-west-2.amazonaws.com/ray-wheels/latest/ray-0.9.0.dev0-cp37-cp37m-manylinux1_x86_64.whl # - pip install -U https://s3-us-west-2.amazonaws.com/ray-wheels/latest/ray-0.9.0.dev0-cp37-cp37m-manylinux1_x86_64.whl # Consider uncommenting these if you also want to run apt-get commands during setup # - sudo pkill -9 apt-get || true # - sudo pkill -9 dpkg || true # - sudo dpkg --configure -a # Custom commands that will be run on the head node after common setup. head_setup_commands: - pip install boto3 # 1.4.8 adds InstanceMarketOptions # Custom commands that will be run on worker nodes after common setup. worker_setup_commands: - pip install boto3 # 1.4.8 adds InstanceMarketOptions # Command to start ray on the head node. You don't need to change this. head_start_ray_commands: - ray stop - ulimit -n 65536; ray start --num-cpus=0 --head --port=6379 --object-manager-port=8076 --autoscaling-config=~/ray_bootstrap_config.yaml # Command to start ray on worker nodes. You don't need to change this. worker_start_ray_commands: - ray stop - ulimit -n 65536; ray start --address=$RAY_HEAD_IP:6379 --object-manager-port=8076 ```
2020-08-13T01:01:46
ray-project/ray
10,097
ray-project__ray-10097
[ "9558" ]
0effcda3e4fce48de26dcb7ab64ac236eaa543e0
diff --git a/python/ray/autoscaler/command_runner.py b/python/ray/autoscaler/command_runner.py --- a/python/ray/autoscaler/command_runner.py +++ b/python/ray/autoscaler/command_runner.py @@ -189,10 +189,16 @@ def run_rsync_up(self, source, target): logger.warning(self.log_prefix + "rsync failed: '{}'. Falling back to 'kubectl cp'" .format(e)) - self.process_runner.check_call(self.kubectl + [ - "cp", source, "{}/{}:{}".format(self.namespace, self.node_id, - target) - ]) + self.run_cp_up(source, target) + + def run_cp_up(self, source, target): + if target.startswith("~"): + target = "/root" + target[1:] + + self.process_runner.check_call(self.kubectl + [ + "cp", source, "{}/{}:{}".format(self.namespace, self.node_id, + target) + ]) def run_rsync_down(self, source, target): if target.startswith("~"): @@ -209,10 +215,16 @@ def run_rsync_down(self, source, target): logger.warning(self.log_prefix + "rsync failed: '{}'. Falling back to 'kubectl cp'" .format(e)) - self.process_runner.check_call(self.kubectl + [ - "cp", "{}/{}:{}".format(self.namespace, self.node_id, source), - target - ]) + self.run_cp_down(source, target) + + def run_cp_down(self, source, target): + if target.startswith("~"): + target = "/root" + target[1:] + + self.process_runner.check_call(self.kubectl + [ + "cp", "{}/{}:{}".format(self.namespace, self.node_id, source), + target + ]) def remote_shell_command_str(self): return "{} exec -it {} bash".format(" ".join(self.kubectl), diff --git a/python/ray/tune/integration/kubernetes.py b/python/ray/tune/integration/kubernetes.py new file mode 100644 --- /dev/null +++ b/python/ray/tune/integration/kubernetes.py @@ -0,0 +1,169 @@ +import kubernetes +import subprocess + +from ray import services, logger +from ray.autoscaler.command_runner import KubernetesCommandRunner +from ray.tune.syncer import NodeSyncer +from ray.tune.sync_client import SyncClient + + +def NamespacedKubernetesSyncer(namespace, use_rsync=False): + """Wrapper to return a ``KubernetesSyncer`` for a Kubernetes namespace. + + Args: + namespace (str): Kubernetes namespace. + use_rsync (bool): Use ``rsync`` if True or ``kubectl cp`` + if False. If True, ``rsync`` will need to be + installed in the Kubernetes pods for this to work. + If False, ``tar`` will need to be installed instead. + + Returns: A ``KubernetesSyncer`` class to be passed to + ``tune.run(sync_to_driver)``. + + Example: + + .. code-block:: python + + from ray.tune.integration.kubernetes import NamespacedKubernetesSyncer + tune.run(train, + sync_to_driver=NamespacedKubernetesSyncer("ray")) + + """ + + class _NamespacedKubernetesSyncer(KubernetesSyncer): + _namespace = namespace + _use_rsync = use_rsync + + return _NamespacedKubernetesSyncer + + +class KubernetesSyncer(NodeSyncer): + """KubernetesSyncer used for synchronization between Kubernetes pods. + + This syncer extends the node syncer, but is usually instantiated + without a custom sync client. The sync client defaults to + ``KubernetesSyncClient`` instead. + + KubernetesSyncer uses the default namespace ``ray``. You should + probably use ``NamespacedKubernetesSyncer`` to return a class + with a custom namespace instead. + """ + + _namespace = "ray" + _use_rsync = False + + def __init__(self, local_dir, remote_dir, sync_client=None): + self.local_ip = services.get_node_ip_address() + self.local_node = self._get_kubernetes_node_by_ip(self.local_ip) + self.worker_ip = None + self.worker_node = None + + sync_client = sync_client or KubernetesSyncClient( + namespace=self.__class__._namespace, + use_rsync=self.__class__._use_rsync) + + super(NodeSyncer, self).__init__(local_dir, remote_dir, sync_client) + + def set_worker_ip(self, worker_ip): + self.worker_ip = worker_ip + self.worker_node = self._get_kubernetes_node_by_ip(worker_ip) + + def _get_kubernetes_node_by_ip(self, node_ip): + """Return node name by internal or external IP""" + kubernetes.config.load_incluster_config() + api = kubernetes.client.CoreV1Api() + pods = api.list_namespaced_pod(self._namespace) + for pod in pods.items: + if pod.status.host_ip == node_ip or \ + pod.status.pod_ip == node_ip: + return pod.metadata.name + + logger.error( + "Could not find Kubernetes pod name for IP {}".format(node_ip)) + return None + + @property + def _remote_path(self): + return (self.worker_node, self._remote_dir) + + +class KubernetesSyncClient(SyncClient): + """KubernetesSyncClient to be used by KubernetesSyncer. + + This client takes care of executing the synchronization + commands for Kubernetes clients. In its ``sync_down`` and + ``sync_up`` commands, it expects tuples for the source + and target, respectively, for compatibility with the + KubernetesCommandRunner. + + Args: + namespace (str): Namespace in which the pods live. + use_rsync (bool): Use ``rsync`` if True or ``kubectl cp`` + if False. If True, ``rsync`` will need to be + installed in the Kubernetes pods for this to work. + If False, ``tar`` will need to be installed instead. + process_runner: How commands should be called. + Defaults to ``subprocess``. + + """ + + def __init__(self, namespace, use_rsync=False, process_runner=subprocess): + self.namespace = namespace + self.use_rsync = use_rsync + self._process_runner = process_runner + self._command_runners = {} + + def _create_command_runner(self, node_id): + """Create a command runner for one Kubernetes node""" + return KubernetesCommandRunner( + log_prefix="KubernetesSyncClient: {}:".format(node_id), + namespace=self.namespace, + node_id=node_id, + auth_config=None, + process_runner=self._process_runner) + + def _get_command_runner(self, node_id): + """Create command runner if it doesn't exist""" + # Todo(krfricke): These cached runners are currently + # never cleaned up. They are cheap so this shouldn't + # cause much problems, but should be addressed if + # the SyncClient is used more extensively in the future. + if node_id not in self._command_runners: + command_runner = self._create_command_runner(node_id) + self._command_runners[node_id] = command_runner + return self._command_runners[node_id] + + def sync_up(self, source, target): + """Here target is a tuple (target_node, target_dir)""" + target_node, target_dir = target + + # Add trailing slashes for rsync + source += "/" if not source.endswith("/") else "" + target_dir += "/" if not target_dir.endswith("/") else "" + + command_runner = self._get_command_runner(target_node) + if self.use_rsync: + command_runner.run_rsync_up(source, target_dir) + else: + command_runner.run_cp_up(source, target_dir) + return True + + def sync_down(self, source, target): + """Here source is a tuple (source_node, source_dir)""" + source_node, source_dir = source + + # Add trailing slashes for rsync + source_dir += "/" if not source_dir.endswith("/") else "" + target += "/" if not target.endswith("/") else "" + + command_runner = self._get_command_runner(source_node) + if self.use_rsync: + command_runner.run_rsync_down(source_dir, target) + else: + command_runner.run_cp_down(source_dir, target) + return True + + def delete(self, target): + """No delete function because it is only used by + the KubernetesSyncer, which doesn't call delete.""" + return True diff --git a/python/ray/tune/syncer.py b/python/ray/tune/syncer.py --- a/python/ray/tune/syncer.py +++ b/python/ray/tune/syncer.py @@ -3,6 +3,7 @@ import os import time +from inspect import isclass from shlex import quote from ray import ray_constants @@ -285,6 +286,9 @@ def get_node_syncer(local_dir, remote_dir=None, sync_function=None): key = (local_dir, remote_dir) if key in _syncers: return _syncers[key] + elif isclass(sync_function) and issubclass(sync_function, Syncer): + _syncers[key] = sync_function(local_dir, remote_dir, None) + return _syncers[key] elif not remote_dir or sync_function is False: sync_client = NOOP elif sync_function and sync_function is not True:
diff --git a/python/ray/tune/tests/test_integration_kubernetes.py b/python/ray/tune/tests/test_integration_kubernetes.py new file mode 100644 --- /dev/null +++ b/python/ray/tune/tests/test_integration_kubernetes.py @@ -0,0 +1,147 @@ +import unittest + +from ray.autoscaler.command_runner import KUBECTL_RSYNC +from ray.tune.integration.kubernetes import KubernetesSyncer, \ + KubernetesSyncClient +from ray.tune.syncer import NodeSyncer + + +class _MockProcessRunner: + def __init__(self): + self.history = [] + + def check_call(self, command): + self.history.append(command) + return True + + +class _MockLookup: + def __init__(self, node_ips): + self.node_to_ip = {} + self.ip_to_node = {} + for node, ip in node_ips.items(): + self.node_to_ip[node] = ip + self.ip_to_node[ip] = node + + def get_ip(self, node): + return self.node_to_ip[node] + + def get_node(self, ip): + return self.ip_to_node[ip] + + def __call__(self, ip): + return self.ip_to_node[ip] + + +def _create_mock_syncer(namespace, lookup, use_rsync, process_runner, local_ip, + local_dir, remote_dir): + class _MockSyncer(KubernetesSyncer): + _namespace = namespace + _use_rsync = use_rsync + _get_kubernetes_node_by_ip = lookup + + def __init__(self, local_dir, remote_dir, sync_client): + self.local_ip = local_ip + self.local_node = self._get_kubernetes_node_by_ip(self.local_ip) + self.worker_ip = None + self.worker_node = None + + sync_client = sync_client + super(NodeSyncer, self).__init__(local_dir, remote_dir, + sync_client) + + return _MockSyncer( + local_dir, + remote_dir, + sync_client=KubernetesSyncClient( + namespace=namespace, + use_rsync=use_rsync, + process_runner=process_runner)) + + +class KubernetesIntegrationTest(unittest.TestCase): + def setUp(self): + self.namespace = "test_ray" + self.lookup = _MockLookup({ + "head": "1.0.0.0", + "w1": "1.0.0.1", + "w2": "1.0.0.2" + }) + self.process_runner = _MockProcessRunner() + self.local_dir = "/tmp/local" + self.remote_dir = "/tmp/remote" + + def tearDown(self): + pass + + def testKubernetesCpUpDown(self): + syncer = _create_mock_syncer( + self.namespace, self.lookup, False, self.process_runner, + self.lookup.get_ip("head"), self.local_dir, self.remote_dir) + + syncer.set_worker_ip(self.lookup.get_ip("w1")) + + # Test sync up. Should add / to the dirs and call kubectl cp + syncer.sync_up() + + self.assertEqual(self.process_runner.history[-1], [ + "kubectl", "-n", self.namespace, "cp", self.local_dir + "/", + "{}/{}:{}".format(self.namespace, "w1", self.remote_dir + "/") + ]) + + # Test sync down. + syncer.sync_down() + self.assertEqual(self.process_runner.history[-1], [ + "kubectl", "-n", self.namespace, "cp", "{}/{}:{}".format( + self.namespace, + "w1", + self.remote_dir + "/", + ), self.local_dir + "/" + ]) + + # Sync to same node should be ignored + syncer.set_worker_ip(self.lookup.get_ip("head")) + syncer.sync_up() + self.assertTrue(len(self.process_runner.history) == 2) + + syncer.sync_down() + self.assertTrue(len(self.process_runner.history) == 2) + + def testKubernetesRsyncUpDown(self): + syncer = _create_mock_syncer( + self.namespace, self.lookup, True, self.process_runner, + self.lookup.get_ip("head"), self.local_dir, self.remote_dir) + + syncer.set_worker_ip(self.lookup.get_ip("w1")) + + # Test sync up. Should add / to the dirs and call rsync + syncer.sync_up() + self.assertEqual(self.process_runner.history[-1][0], KUBECTL_RSYNC) + self.assertEqual(self.process_runner.history[-1][-2], + self.local_dir + "/") + self.assertEqual( + self.process_runner.history[-1][-1], "{}@{}:{}".format( + "w1", self.namespace, self.remote_dir + "/")) + + # Test sync down. + syncer.sync_down() + self.assertEqual(self.process_runner.history[-1][0], KUBECTL_RSYNC) + self.assertEqual( + self.process_runner.history[-1][-2], "{}@{}:{}".format( + "w1", self.namespace, self.remote_dir + "/")) + self.assertEqual(self.process_runner.history[-1][-1], + self.local_dir + "/") + + # Sync to same node should be ignored + syncer.set_worker_ip(self.lookup.get_ip("head")) + syncer.sync_up() + self.assertTrue(len(self.process_runner.history) == 2) + + syncer.sync_down() + self.assertTrue(len(self.process_runner.history) == 2) + + +if __name__ == "__main__": + import pytest + import sys + sys.exit(pytest.main(["-v", __file__]))
[tune] Ray Cluster deployed in Kubernetes unable to sync checkpoint directories to the driver node ### What is the problem? Getting the following error (Stacktrace attached): ``` 2020-07-17 21:53:48,101 ERROR trial_runner.py:550 -- Trial TrainExample_fd24b_00001: Error handling checkpoint /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/ Traceback (most recent call last): File "/opt/conda/lib/python3.6/site-packages/ray/tune/trial_runner.py", line 546, in _process_trial_save trial.on_checkpoint(trial.saving_to) File "/opt/conda/lib/python3.6/site-packages/ray/tune/trial.py", line 448, in on_checkpoint self, checkpoint.value)) ray.tune.error.TuneError: Trial TrainExample_fd24b_00001: Checkpoint path /root/ray_results/TrainExample/TrainExample_1_randomforestclassifier__n_estimators=5_2020-07-17_21-53-462l3hkjfs/checkpoint_1/ not found after successful sync down. ``` This is output of `ray_random_forest.py`. From investigation it looks like the checkpoint directories ie. **checkpoint_1, checkpoint_2** that were created in other nodes (workers) of the cluster do not get synced back to the driver node (from where I am running the python script). Rest of the files in the Trial directories seem to be in sync. This problem is not reproducible by running it on a single-machine with a cluster started up using bare `init()`. My guess is because all the workers point to the same file-system, and thus, the `os.path.exist()` returns True (This is reference to the line of code which checks and throws the above error) Taking a look at `validate_save_restore()` to validate a Trainable, I have written a bare-bone script which proves that `checkpoint` directories are not synced across workers. The stacktrace from `checkpoint_validation_failed.py` ``` Traceback (most recent call last): File "code/checkpoint_validation_failed.py", line 115, in <module> _main() File "code/checkpoint_validation_failed.py", line 94, in _main print(f"VALIDATE TRAINABLE CLASS: {validate_save_restore(TrainExample)}") File "code/checkpoint_validation_failed.py", line 67, in validate_save_restore restore_check = ray.get(trainable_2.restore.remote(old_trainable)) File "/opt/conda/lib/python3.6/site-packages/ray/worker.py", line 1526, in get raise value.as_instanceof_cause() ray.exceptions.RayTaskError(FileNotFoundError): ray::TrainExample.restore() (pid=2704, ip=172.17.0.8) File "python/ray/_raylet.pyx", line 471, in ray._raylet.execute_task File "python/ray/_raylet.pyx", line 424, in ray._raylet.execute_task.function_executor File "/opt/conda/lib/python3.6/site-packages/ray/tune/trainable.py", line 444, in restore with open(checkpoint_path + ".tune_metadata", "rb") as f: FileNotFoundError: [Errno 2] No such file or directory: '/root/ray_results/2020-07-17_21-43-17dbvo01tp/checkpoint_3/.tune_metadata' command terminated with exit code 1 ``` Here, Worker X is trying to restore checkpoint created by Worker Y. *Ray version and other system information (Python version, TensorFlow version, OS):* Setup: - Ray cluster running in K8s. - Ray version `pip list | grep ray ray 0.9.0.dev0` - Docker image: ` REPOSITORY TAG IMAGE ID CREATED SIZE rayproject/autoscaler latest 3f8d747cb8af 7 days ago 5.25GB ` Playing with `sync_to_driver`, `sync_on_checkpoint` parameters did not help. Thanks! ### Reproduction (REQUIRED) Step 1. Spin-up a Ray cluster using `ray-cluster.yaml` in a namespace called `ray`. ie 1. `kubectl create -f ray-namespace.yaml` 2. `kubectl apply -f ray-cluster.yaml` Step 2. Copy the python script to any worker node. 1. `kubectl -n ray get pods`. 2. Copy WORKER_1_POD_NAME 3. Exec into pod `kubectl -n ray exec -it <WORKER_1_POD_NAME> -- bin/bash` 4. `mkdir code` 5. Copy the python script `kubectl -n ray cp checkpoint_validation_failed.py <WORKER_1_POD_NAME>:/code/checkpoint_validation_failed.py` 6. Exit pod 7. Execute script `kubectl -n ray exec <WORKER_1_POD_NAME> -- bin/bash -c "python code/checkpoint_validation_failed.py` You can do the same steps to run `ray_random_forest.py`. [ray_issue_related_files.zip](https://github.com/ray-project/ray/files/4940566/ray_issue_related_files.zip) The above zip contains: 1. ray-cluster.yaml 2. ray-namespace.yaml 3. checkpoint_validation_failed.py 4. ray_random_forest.py If we cannot run your script, we cannot fix your issue. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
Do you have an NFS for your kubernetes cluster? That might save you a lot of trouble. Hello @richardliaw, Thank you for a quick reply. No, I do not have NFS for the cluster. Does it have to be a NFS? How does ray sync files like 1. events.out.tfevents.1595022826.ray-worker-6d9f6b9b8d-g4zvk 2. params.json 3. params.pkl 4. progress.csv 5. result.json But not `checkpoint_1`, `checkpoint_2` etc.? The files are actually generated on the driver, not on the node. Do you have access to S3? S3 could work too. Okay. So what gets synced on `sync_on_checkpoint`? sync_on_checkpoint (bool) – Force sync-down of trial **checkpoint** to driver. If set to False, checkpoint syncing from worker to driver is asynchronous and best-effort. This does not affect persistent storage syncing. Defaults to True. In the above documentation snippet I am interpreting **checkpoint** as the `checkpoint_N` directories described above. Would the suggested S3 solution involve implementing DurableTrainable rather than Trainable? Also I wanted to clarify, is the above issue a bug in Ray Tune and a network/cloud storage a work-around for it? Checkpoints (directories) are synced on sync_on_checkpoint, yep. The above is just a limitation of our file-transfer mechanism; we use rsync to move files between VMs, but I think that mechanism breaks down on kubernetes. The two approaches to workaround this limitation are to use 1. a networked storage or 2. to implement your own file-transfer protocol (check out `sync_to_driver`). Does that make sense? Let me know if you have any further questions! Thank you Richard for this clarification. Upon searching more, I found this: https://serverfault.com/questions/741670/rsync-files-to-a-kubernetes-pod I see similar solution applied to `KubernetesCommandRunner` in Ray. Maybe we can do the same for sync-checkpoint mechanism to resolve this issue. Or, we just implement it as a custom file-transfer protocol and make it part of the pallet. On a side note, I have taken your suggestion to use a NFS. It has unblocked me from aggregating all my checkpoints. Thanks once again. Great! yeah, KubernetesCommandRunner is what we will want to use. More than happy to accept a PR if you want to take a quick pass! Otherwise, we'll try to get to this in the next couple of weeks. I spent sometime last-week to understand if I can fix it, but for me it was not a going to be a quick pass. So I will defer it to the team. Thanks once again Richard. :)
2020-08-13T10:21:24
ray-project/ray
10,122
ray-project__ray-10122
[ "10117" ]
c7adb464e49db7309cf4084d698dae7968b8e2e3
diff --git a/rllib/offline/off_policy_estimator.py b/rllib/offline/off_policy_estimator.py --- a/rllib/offline/off_policy_estimator.py +++ b/rllib/offline/off_policy_estimator.py @@ -1,10 +1,13 @@ from collections import namedtuple import logging +import numpy as np + from ray.rllib.policy.sample_batch import MultiAgentBatch, SampleBatch from ray.rllib.policy import Policy from ray.rllib.utils.annotations import DeveloperAPI from ray.rllib.offline.io_context import IOContext +from ray.rllib.utils.numpy import convert_to_numpy from ray.rllib.utils.typing import TensorType, SampleBatchType from typing import List @@ -53,7 +56,7 @@ def estimate(self, batch: SampleBatchType): raise NotImplementedError @DeveloperAPI - def action_prob(self, batch: SampleBatchType) -> TensorType: + def action_prob(self, batch: SampleBatchType) -> np.ndarray: """Returns the probs for the batch actions for the current policy.""" num_state_inputs = 0 @@ -61,13 +64,13 @@ def action_prob(self, batch: SampleBatchType) -> TensorType: if k.startswith("state_in_"): num_state_inputs += 1 state_keys = ["state_in_{}".format(i) for i in range(num_state_inputs)] - log_likelihoods = self.policy.compute_log_likelihoods( + log_likelihoods: TensorType = self.policy.compute_log_likelihoods( actions=batch[SampleBatch.ACTIONS], obs_batch=batch[SampleBatch.CUR_OBS], state_batches=[batch[k] for k in state_keys], prev_action_batch=batch.data.get(SampleBatch.PREV_ACTIONS), prev_reward_batch=batch.data.get(SampleBatch.PREV_REWARDS)) - return log_likelihoods + return convert_to_numpy(log_likelihoods) @DeveloperAPI def process(self, batch: SampleBatchType):
[rllib] Off Policy Estimation breaks with GPU on PyTorch (MARWIL) (Offline API) ### What is the problem? When I use MARWIL with PyTorch and num_gpus: 1, I get an error when computing metrics. This happens because in off policy estimation it uses the torch tensors on gpu instead of numpy arrays. Particularly, I use "input_evaluation": ["is", "wis"] and the error goes away when "input_evaluation": ["simulation"] ``` Traceback (most recent call last): File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\ray\tune\trial_runner.py", line 497, in _process_trial result = self.trial_executor.fetch_result(trial) File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\ray\tune\ray_trial_executor.py", line 434, in fetch_result result = ray.get(trial_future[0], DEFAULT_GET_TIMEOUT) File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\ray\worker.py", line 1553, in get raise value.as_instanceof_cause() ray.exceptions.RayTaskError(AttributeError): ray::MARWIL.train() (pid=9136, ip=10.0.0.18) File "python\ray\_raylet.pyx", line 474, in ray._raylet.execute_task File "python\ray\_raylet.pyx", line 427, in ray._raylet.execute_task.function_executor File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\ray\function_manager.py", line 567, in actor_method_executor raise e File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\ray\function_manager.py", line 559, in actor_method_executor method_returns = method(actor, *args, **kwargs) File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\ray\rllib\agents\trainer.py", line 522, in train raise e File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\ray\rllib\agents\trainer.py", line 508, in train result = Trainable.train(self) File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\ray\tune\trainable.py", line 337, in train result = self.step() File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\ray\rllib\agents\trainer_template.py", line 110, in step res = next(self.train_exec_impl) File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\ray\util\iter.py", line 758, in __next__ return next(self.built_iterator) File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\ray\util\iter.py", line 793, in apply_foreach result = fn(item) File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\ray\rllib\execution\metric_ops.py", line 87, in __call__ res = summarize_episodes(episodes, orig_episodes) File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\ray\rllib\evaluation\metrics.py", line 173, in summarize_episodes metrics[k] = np.mean(v_list) File "<__array_function__ internals>", line 6, in mean File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\numpy\core\fromnumeric.py", line 3335, in mean out=out, **kwargs) File "C:\Users\Julius\Anaconda3\envs\ray\lib\site-packages\numpy\core\_methods.py", line 161, in _mean ret = ret.dtype.type(ret / rcount) AttributeError: 'torch.dtype' object has no attribute 'type' ``` ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): If we cannot run your script, we cannot fix your issue. - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
Hmm I think we are missing a call to `.cpu().numpy()` somewhere and are trying to calculate a np.mean() over torch tensors representing a single element. Do you think you could look into this issue?
2020-08-14T23:29:31
ray-project/ray
10,221
ray-project__ray-10221
[ "8894" ]
d206fbbc991dff60d152151ef19af5412f6774b1
diff --git a/python/ray/worker.py b/python/ray/worker.py --- a/python/ray/worker.py +++ b/python/ray/worker.py @@ -1806,7 +1806,21 @@ def decorator(function_or_class): if max_task_retries is not None: raise ValueError("The keyword 'max_task_retries' is not " "allowed for remote functions.") - + if num_return_vals is not None and (not isinstance( + num_return_vals, int) or num_return_vals < 0): + raise ValueError( + "The keyword 'num_return_vals' only accepts 0 or a" + " positive integer") + if max_retries is not None and (not isinstance(max_retries, int) + or max_retries < -1): + raise ValueError( + "The keyword 'max_retries' only accepts 0, -1 or a" + " positive integer") + if max_calls is not None and (not isinstance(max_calls, int) + or max_calls < 0): + raise ValueError( + "The keyword 'max_calls' only accepts 0 or a positive" + " integer") return ray.remote_function.RemoteFunction( Language.PYTHON, function_or_class, None, num_cpus, num_gpus, memory, object_store_memory, resources, num_return_vals, @@ -1820,7 +1834,16 @@ def decorator(function_or_class): if max_calls is not None: raise TypeError("The keyword 'max_calls' is not " "allowed for actors.") - + if max_restarts is not None and (not isinstance(max_restarts, int) + or max_restarts < -1): + raise ValueError( + "The keyword 'max_restarts' only accepts -1, 0 or a" + " positive integer") + if max_task_retries is not None and (not isinstance( + max_task_retries, int) or max_task_retries < -1): + raise ValueError( + "The keyword 'max_task_retries' only accepts -1, 0 or a" + " positive integer") return ray.actor.make_actor(function_or_class, num_cpus, num_gpus, memory, object_store_memory, resources, max_restarts, max_task_retries)
diff --git a/python/ray/tests/test_basic.py b/python/ray/tests/test_basic.py --- a/python/ray/tests/test_basic.py +++ b/python/ray/tests/test_basic.py @@ -111,6 +111,60 @@ def method(self): assert ray.get([id1, id2, id3, id4]) == [0, 1, "test", 2] +def test_invalid_arguments(shutdown_only): + ray.init(num_cpus=2) + + for opt in [np.random.randint(-100, -1), np.random.uniform(0, 1)]: + with pytest.raises( + ValueError, + match="The keyword 'num_return_vals' only accepts 0 or a" + " positive integer"): + + @ray.remote(num_return_vals=opt) + def g1(): + return 1 + + for opt in [np.random.randint(-100, -2), np.random.uniform(0, 1)]: + with pytest.raises( + ValueError, + match="The keyword 'max_retries' only accepts 0, -1 or a" + " positive integer"): + + @ray.remote(max_retries=opt) + def g2(): + return 1 + + for opt in [np.random.randint(-100, -1), np.random.uniform(0, 1)]: + with pytest.raises( + ValueError, + match="The keyword 'max_calls' only accepts 0 or a positive" + " integer"): + + @ray.remote(max_calls=opt) + def g3(): + return 1 + + for opt in [np.random.randint(-100, -2), np.random.uniform(0, 1)]: + with pytest.raises( + ValueError, + match="The keyword 'max_restarts' only accepts -1, 0 or a" + " positive integer"): + + @ray.remote(max_restarts=opt) + class A1: + x = 1 + + for opt in [np.random.randint(-100, -2), np.random.uniform(0, 1)]: + with pytest.raises( + ValueError, + match="The keyword 'max_task_retries' only accepts -1, 0 or a" + " positive integer"): + + @ray.remote(max_task_retries=opt) + class A2: + x = 1 + + def test_many_fractional_resources(shutdown_only): ray.init(num_cpus=2, num_gpus=2, resources={"Custom": 2})
No proper error message when num_return_vals is negative or not integral <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? *Ray version and other system information (Python version, TensorFlow version, OS):* *Ubuntu 18.04.4 LTS* *Python 3.7.7* *ray-0.9.0.dev0-cp37-cp37m-manylinux1_x86_64.whl* * For non integral values of `x`, eg. `x = 3.2` in the code below, ray doesn't give a proper error message. Instead the worker dies. This also happens when `x` is not equal to the true num return values, eg. `x = 5`. ` ray.exceptions.RayWorkerError: The worker died unexpectedly while executing this task. ` * For negative values of `x`, eg. `x = -2` the kernel dies altogether. ` Aborted (core dumped)` Although one would not expect the user to enter such values. But still. ### Reproduction ``` import ray ray.init() def f(a): return a, 2 x = <some_value> g = ray.remote(num_return_vals = x)(f) print(ray.get(g.remote(3))) ``` - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
@vinamrabenara thanks for catching this!
2020-08-20T10:26:48
ray-project/ray
10,323
ray-project__ray-10323
[ "10319" ]
b14c56e599ea34dd88838c820151063b1aeba0e8
diff --git a/python/ray/dashboard/node_stats.py b/python/ray/dashboard/node_stats.py --- a/python/ray/dashboard/node_stats.py +++ b/python/ray/dashboard/node_stats.py @@ -60,14 +60,22 @@ def __init__(self, redis_address, redis_password=None): def _insert_log_counts(self): for ip, logs_by_pid in self._logs.items(): hostname = self._ip_to_hostname[ip] - logs_by_pid = {pid: len(logs) for pid, logs in logs_by_pid.items()} - self._node_stats[hostname]["log_count"] = logs_by_pid + if hostname in self._node_stats: + logs_by_pid = { + pid: len(logs) + for pid, logs in logs_by_pid.items() + } + self._node_stats[hostname]["log_count"] = logs_by_pid def _insert_error_counts(self): for ip, errs_by_pid in self._errors.items(): hostname = self._ip_to_hostname[ip] - errs_by_pid = {pid: len(errs) for pid, errs in errs_by_pid.items()} - self._node_stats[hostname]["error_count"] = errs_by_pid + if hostname in self._node_stats: + errs_by_pid = { + pid: len(errs) + for pid, errs in errs_by_pid.items() + } + self._node_stats[hostname]["error_count"] = errs_by_pid def _purge_outdated_stats(self): def current(then, now):
[dashboard] When autoscaler adds/removes worker, dashboard crashes <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? This is originally reported by Pieterjan for 0.8.7 but also appears to be in issue in the latest nightly. The issue would appear to occur across ### Reproduction (REQUIRED) This is not a script, but given I will author the fix, I think this is sufficient: - Start a cluster with autoscaling enabled - Increase workload so that an additional worker or workers are spawned - When one of those workers is removed, their log entries still exist on the backend - The following stack trace occurs and crashes the dashboard: `/ray/dashboard/node_stats.py", line 63, in _insert_log_counts self._node_stats[hostname]["log_count"] = logs_by_pid` - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-08-25T18:27:35
ray-project/ray
10,443
ray-project__ray-10443
[ "10228" ]
eb025ea8cb27583e8ef6287f5654f23d1ab270ef
diff --git a/rllib/utils/exploration/stochastic_sampling.py b/rllib/utils/exploration/stochastic_sampling.py --- a/rllib/utils/exploration/stochastic_sampling.py +++ b/rllib/utils/exploration/stochastic_sampling.py @@ -72,5 +72,5 @@ def _get_torch_exploration_action(action_dist, explore): logp = action_dist.sampled_action_logp() else: action = action_dist.deterministic_sample() - logp = torch.zeros((action.size()[0], ), dtype=torch.float32) + logp = torch.zeros_like(action_dist.sampled_action_logp()) return action, logp
[rllib] _get_torch_exploration_action doesn't support tuple action dist <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### System information * **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**: Mac OS 10.15.4 * **Ray installed from (source or binary)**: binary (via pip) * **Ray version**: 0.8.6., but nothing seems to have changed on master * **Python version**: 3.7 ### What is the problem? When using tuple action distributions (as advised in #6372) and exploration is disabled, the line: https://github.com/ray-project/ray/blob/a462ae2747afbeb9047e443cd51e67e3fe0b49e6/rllib/utils/exploration/stochastic_sampling.py#L75 from `_get_torch_exploration_action` raises the following exception: ``` AttributeError: 'tuple' object has no attribute 'size' ``` A simple fix that supports any type of distribution would be: ```python logp = torch.zeros_like(action_dist.sampled_action_logp()) ``` I can submit a PR if it helps. ### Reproduction (REQUIRED) Exact command to reproduce: python `rllib_cartpole.py` for the following file ```python import gym.envs.classic_control from gym.spaces import Tuple, Discrete import ray from ray import tune class CustomCartpole(gym.envs.classic_control.CartPoleEnv): """Add a dimension to the cartpole action space that is ignored.""" def __init__(self, env_config): super().__init__() # if override_actions is false this is just the Cartpole environment self.override_actions = env_config['override_actions'] if self.override_actions: # 2 is the environment's normal action space # 4 is just a dummy number to give it an extra dimension self.original_action_space = self.action_space self.action_space = Tuple([Discrete(2), Discrete(4)]) self.tuple_action_space = self.action_space def step(self, action): # call the cartpole environment with the original action if self.override_actions: self.action_space = self.original_action_space return super().step(action[0]) else: return super().step(action) def main(): ray.init() tune.run( "PPO", stop={"episode_reward_mean": 50}, config={ "env": CustomCartpole, "env_config": {'override_actions': True}, "num_gpus": 0, "num_workers": 1, "eager": False, "evaluation_interval": 1, "evaluation_config": { "explore": False, }, "framework": "torch", }, ) if __name__ == '__main__': main() ``` - [x] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
The proposed fix makes sense to me. We could alternatively try to get the batch dimension of the tuple, but I don't see an existing helper method for that, so your proposal is probably simpler. And yeah, a PR would be great!
2020-08-31T11:43:42
ray-project/ray
10,449
ray-project__ray-10449
[ "10371" ]
715ee8dfc95ab44909f800e8b917e7e44082a6ec
diff --git a/python/ray/actor.py b/python/ray/actor.py --- a/python/ray/actor.py +++ b/python/ray/actor.py @@ -100,7 +100,27 @@ def __call__(self, *args, **kwargs): def remote(self, *args, **kwargs): return self._remote(args, kwargs) - def _remote(self, args=None, kwargs=None, num_returns=None): + def options(self, **options): + """Convenience method for executing an actor method call with options. + + Same arguments as func._remote(), but returns a wrapped function + that a non-underscore .remote() can be called on. + + Examples: + # The following two calls are equivalent. + >>> actor.my_method._remote(args=[x, y], name="foo", num_returns=2) + >>> actor.my_method.options(name="foo", num_returns=2).remote(x, y) + """ + + func_cls = self + + class FuncWrapper: + def remote(self, *args, **kwargs): + return func_cls._remote(args=args, kwargs=kwargs, **options) + + return FuncWrapper() + + def _remote(self, args=None, kwargs=None, name="", num_returns=None): if num_returns is None: num_returns = self._num_returns @@ -112,6 +132,7 @@ def invocation(args, kwargs): self._method_name, args=args, kwargs=kwargs, + name=name, num_returns=num_returns) # Apply the decorator if there is one. @@ -317,8 +338,10 @@ def _ray_from_modified_class(cls, modified_class, class_id, max_restarts, max_task_retries, num_cpus, num_gpus, memory, object_store_memory, resources): for attribute in [ - "remote", "_remote", "_ray_from_modified_class", - "_ray_from_function_descriptor" + "remote", + "_remote", + "_ray_from_modified_class", + "_ray_from_function_descriptor", ]: if hasattr(modified_class, attribute): logger.warning("Creating an actor from class " @@ -679,6 +702,7 @@ def _actor_method_call(self, method_name, args=None, kwargs=None, + name="", num_returns=None): """Method execution stub for an actor handle. @@ -691,6 +715,7 @@ def _actor_method_call(self, method_name: The name of the actor method to execute. args: A list of arguments for the actor method. kwargs: A dictionary of keyword arguments for the actor method. + name (str): The name to give the actor method call task. num_returns (int): The number of return values for the method. Returns: @@ -724,7 +749,7 @@ def _actor_method_call(self, object_refs = worker.core_worker.submit_actor_task( self._ray_actor_language, self._ray_actor_id, function_descriptor, - list_args, num_returns, self._ray_actor_method_cpus) + list_args, name, num_returns, self._ray_actor_method_cpus) if len(object_refs) == 1: object_refs = object_refs[0] diff --git a/python/ray/remote_function.py b/python/ray/remote_function.py --- a/python/ray/remote_function.py +++ b/python/ray/remote_function.py @@ -152,7 +152,8 @@ def _remote(self, resources=None, max_retries=None, placement_group=None, - placement_group_bundle_index=-1): + placement_group_bundle_index=-1, + name=""): """Submit the remote function for execution.""" worker = ray.worker.global_worker worker.check_connected() @@ -212,7 +213,7 @@ def invocation(args, kwargs): "Cross language remote function " \ "cannot be executed locally." object_refs = worker.core_worker.submit_task( - self._language, self._function_descriptor, list_args, + self._language, self._function_descriptor, list_args, name, num_returns, resources, max_retries, placement_group.id, placement_group_bundle_index)
diff --git a/java/test/src/main/java/io/ray/test/TaskNameTest.java b/java/test/src/main/java/io/ray/test/TaskNameTest.java new file mode 100644 --- /dev/null +++ b/java/test/src/main/java/io/ray/test/TaskNameTest.java @@ -0,0 +1,19 @@ +package io.ray.test; + +import io.ray.api.Ray; +import org.testng.Assert; +import org.testng.annotations.Test; + +/** Task Name Test. */ +public class TaskNameTest extends BaseTest { + + private static int testFoo() { + return 0; + } + + /** Test setting task name at task submission time. */ + @Test + public void testSetName() { + Assert.assertEquals(0, (int) Ray.task(TaskNameTest::testFoo).setName("foo").remote().get()); + } +} diff --git a/python/ray/tests/test_actor.py b/python/ray/tests/test_actor.py --- a/python/ray/tests/test_actor.py +++ b/python/ray/tests/test_actor.py @@ -14,6 +14,10 @@ import ray.test_utils import ray.cluster_utils +# NOTE: We have to import setproctitle after ray because we bundle setproctitle +# with ray. +import setproctitle + def test_caching_actors(shutdown_only): # Test defining actors before ray.init() has been called. @@ -673,6 +677,33 @@ def method3(self): assert ray.get([id3a, id3b, id3c]) == [1, 2, 3] +def test_options_num_returns(ray_start_regular_shared): + @ray.remote + class Foo: + def method(self): + return 1, 2 + + f = Foo.remote() + + obj = f.method.remote() + assert ray.get(obj) == (1, 2) + + obj1, obj2 = f.method.options(num_returns=2).remote() + assert ray.get([obj1, obj2]) == [1, 2] + + +def test_options_name(ray_start_regular_shared): + @ray.remote + class Foo: + def method(self, name): + assert setproctitle.getproctitle() == f"ray::{name}" + + f = Foo.remote() + + ray.get(f.method.options(name="foo").remote("foo")) + ray.get(f.method.options(name="bar").remote("bar")) + + def test_define_actor(ray_start_regular_shared): @ray.remote class Test: diff --git a/python/ray/tests/test_advanced_3.py b/python/ray/tests/test_advanced_3.py --- a/python/ray/tests/test_advanced_3.py +++ b/python/ray/tests/test_advanced_3.py @@ -35,7 +35,7 @@ def attempt_to_load_balance(remote_function, [remote_function.remote(*args) for _ in range(total_tasks)]) names = set(locations) counts = [locations.count(name) for name in names] - logger.info("Counts are {}.".format(counts)) + logger.info(f"Counts are {counts}.") if (len(names) == num_nodes and all(count >= minimum_count for count in counts)): break @@ -346,6 +346,28 @@ def unique_1(): ray.get(unique_1.remote()) +def test_ray_task_name_setproctitle(ray_start_2_cpus): + method_task_name = "foo" + + @ray.remote + class UniqueName: + def __init__(self): + assert setproctitle.getproctitle() == "ray::UniqueName.__init__()" + + def f(self): + assert setproctitle.getproctitle() == f"ray::{method_task_name}" + + task_name = "bar" + + @ray.remote + def unique_1(): + assert task_name in setproctitle.getproctitle() + + actor = UniqueName.remote() + ray.get(actor.f.options(name=method_task_name).remote()) + ray.get(unique_1.options(name=task_name).remote()) + + @pytest.mark.skipif( os.getenv("TRAVIS") is None, reason="This test should only be run on Travis.") @@ -508,7 +530,7 @@ def test_invalid_unicode_in_worker_log(shutdown_only): # Wait till first worker log file is created. while True: - log_file_paths = glob.glob("{}/worker*.out".format(logs_dir)) + log_file_paths = glob.glob(f"{logs_dir}/worker*.out") if len(log_file_paths) == 0: time.sleep(0.2) else: @@ -546,13 +568,13 @@ def f(self): # Make sure no log files are in the "old" directory before the actors # are killed. - assert len(glob.glob("{}/old/worker*.out".format(logs_dir))) == 0 + assert len(glob.glob(f"{logs_dir}/old/worker*.out")) == 0 # Now kill the actors so the files get moved to logs/old/. [a.__ray_terminate__.remote() for a in actors] while True: - log_file_paths = glob.glob("{}/old/worker*.out".format(logs_dir)) + log_file_paths = glob.glob(f"{logs_dir}/old/worker*.out") if len(log_file_paths) > 0: with open(log_file_paths[0], "r") as f: assert "function f finished\n" in f.readlines() @@ -641,7 +663,7 @@ def test_gpu_info_parsing(): """ constraints_dict = resource_spec._constraints_from_gpu_info(info_string) expected_dict = { - "{}V100".format(ray_constants.RESOURCE_CONSTRAINT_PREFIX): 1 + f"{ray_constants.RESOURCE_CONSTRAINT_PREFIX}V100": 1, } assert constraints_dict == expected_dict @@ -658,7 +680,7 @@ def test_gpu_info_parsing(): """ constraints_dict = resource_spec._constraints_from_gpu_info(info_string) expected_dict = { - "{}T4".format(ray_constants.RESOURCE_CONSTRAINT_PREFIX): 1 + f"{ray_constants.RESOURCE_CONSTRAINT_PREFIX}T4": 1, } assert constraints_dict == expected_dict diff --git a/src/ray/core_worker/test/core_worker_test.cc b/src/ray/core_worker/test/core_worker_test.cc --- a/src/ray/core_worker/test/core_worker_test.cc +++ b/src/ray/core_worker/test/core_worker_test.cc @@ -226,7 +226,7 @@ bool CoreWorkerTest::WaitForDirectCallActorState(const ActorID &actor_id, bool w int CoreWorkerTest::GetActorPid(const ActorID &actor_id, std::unordered_map<std::string, double> &resources) { std::vector<std::unique_ptr<TaskArg>> args; - TaskOptions options{1, resources}; + TaskOptions options{"", 1, resources}; std::vector<ObjectID> return_ids; RayFunction func{Language::PYTHON, ray::FunctionDescriptorBuilder::BuildPython( "GetWorkerPid", "", "", "")}; @@ -308,7 +308,7 @@ void CoreWorkerTest::TestActorTask(std::unordered_map<std::string, double> &reso args.emplace_back(new TaskArgByValue( std::make_shared<RayObject>(buffer2, nullptr, std::vector<ObjectID>()))); - TaskOptions options{1, resources}; + TaskOptions options{"", 1, resources}; std::vector<ObjectID> return_ids; RayFunction func(ray::Language::PYTHON, ray::FunctionDescriptorBuilder::BuildPython( "MergeInputArgsAsOutput", "", "", "")); @@ -350,7 +350,7 @@ void CoreWorkerTest::TestActorTask(std::unordered_map<std::string, double> &reso args.emplace_back(new TaskArgByValue( std::make_shared<RayObject>(buffer2, nullptr, std::vector<ObjectID>()))); - TaskOptions options{1, resources}; + TaskOptions options{"", 1, resources}; std::vector<ObjectID> return_ids; RayFunction func(ray::Language::PYTHON, ray::FunctionDescriptorBuilder::BuildPython( "MergeInputArgsAsOutput", "", "", "")); @@ -412,7 +412,7 @@ void CoreWorkerTest::TestActorRestart( args.emplace_back(new TaskArgByValue( std::make_shared<RayObject>(buffer1, nullptr, std::vector<ObjectID>()))); - TaskOptions options{1, resources}; + TaskOptions options{"", 1, resources}; std::vector<ObjectID> return_ids; RayFunction func(ray::Language::PYTHON, ray::FunctionDescriptorBuilder::BuildPython( "MergeInputArgsAsOutput", "", "", "")); @@ -455,7 +455,7 @@ void CoreWorkerTest::TestActorFailure( args.emplace_back(new TaskArgByValue( std::make_shared<RayObject>(buffer1, nullptr, std::vector<ObjectID>()))); - TaskOptions options{1, resources}; + TaskOptions options{"", 1, resources}; std::vector<ObjectID> return_ids; RayFunction func(ray::Language::PYTHON, ray::FunctionDescriptorBuilder::BuildPython( "MergeInputArgsAsOutput", "", "", "")); @@ -539,12 +539,12 @@ TEST_F(ZeroNodeTest, TestTaskSpecPerf) { RAY_LOG(INFO) << "start creating " << num_tasks << " PushTaskRequests"; rpc::Address address; for (int i = 0; i < num_tasks; i++) { - TaskOptions options{1, resources}; + TaskOptions options{"", 1, resources}; std::vector<ObjectID> return_ids; auto num_returns = options.num_returns; TaskSpecBuilder builder; - builder.SetCommonTaskSpec(RandomTaskId(), function.GetLanguage(), + builder.SetCommonTaskSpec(RandomTaskId(), options.name, function.GetLanguage(), function.GetFunctionDescriptor(), job_id, RandomTaskId(), 0, RandomTaskId(), address, num_returns, resources, resources, PlacementGroupID::Nil()); @@ -587,7 +587,7 @@ TEST_F(SingleNodeTest, TestDirectActorTaskSubmissionPerf) { args.emplace_back(new TaskArgByValue( std::make_shared<RayObject>(buffer, nullptr, std::vector<ObjectID>()))); - TaskOptions options{1, resources}; + TaskOptions options{"", 1, resources}; std::vector<ObjectID> return_ids; RayFunction func(ray::Language::PYTHON, ray::FunctionDescriptorBuilder::BuildPython( "MergeInputArgsAsOutput", "", "", "")); diff --git a/src/ray/core_worker/test/direct_task_transport_test.cc b/src/ray/core_worker/test/direct_task_transport_test.cc --- a/src/ray/core_worker/test/direct_task_transport_test.cc +++ b/src/ray/core_worker/test/direct_task_transport_test.cc @@ -324,9 +324,10 @@ TaskSpecification BuildTaskSpec(const std::unordered_map<std::string, double> &r const ray::FunctionDescriptor &function_descriptor) { TaskSpecBuilder builder; rpc::Address empty_address; - builder.SetCommonTaskSpec(TaskID::Nil(), Language::PYTHON, function_descriptor, - JobID::Nil(), TaskID::Nil(), 0, TaskID::Nil(), empty_address, - 1, resources, resources, PlacementGroupID::Nil()); + builder.SetCommonTaskSpec(TaskID::Nil(), "dummy_task", Language::PYTHON, + function_descriptor, JobID::Nil(), TaskID::Nil(), 0, + TaskID::Nil(), empty_address, 1, resources, resources, + PlacementGroupID::Nil()); return builder.Build(); } diff --git a/src/ray/core_worker/test/mock_worker.cc b/src/ray/core_worker/test/mock_worker.cc --- a/src/ray/core_worker/test/mock_worker.cc +++ b/src/ray/core_worker/test/mock_worker.cc @@ -50,8 +50,8 @@ class MockWorker { "", // driver_name "", // stdout_file "", // stderr_file - std::bind(&MockWorker::ExecuteTask, this, _1, _2, _3, _4, _5, _6, - _7), // task_execution_callback + std::bind(&MockWorker::ExecuteTask, this, _1, _2, _3, _4, _5, _6, _7, + _8), // task_execution_callback nullptr, // check_signals nullptr, // gc_collect nullptr, // spill_objects @@ -71,7 +71,8 @@ class MockWorker { void RunTaskExecutionLoop() { CoreWorkerProcess::RunTaskExecutionLoop(); } private: - Status ExecuteTask(TaskType task_type, const RayFunction &ray_function, + Status ExecuteTask(TaskType task_type, const std::string task_name, + const RayFunction &ray_function, const std::unordered_map<std::string, double> &required_resources, const std::vector<std::shared_ptr<RayObject>> &args, const std::vector<ObjectID> &arg_reference_ids, diff --git a/src/ray/gcs/test/gcs_test_util.h b/src/ray/gcs/test/gcs_test_util.h --- a/src/ray/gcs/test/gcs_test_util.h +++ b/src/ray/gcs/test/gcs_test_util.h @@ -38,9 +38,10 @@ struct Mocker { auto actor_id = ActorID::Of(job_id, RandomTaskId(), 0); auto task_id = TaskID::ForActorCreationTask(actor_id); auto resource = std::unordered_map<std::string, double>(); - builder.SetCommonTaskSpec(task_id, Language::PYTHON, empty_descriptor, job_id, - TaskID::Nil(), 0, TaskID::Nil(), owner_address, 1, resource, - resource, PlacementGroupID::Nil()); + builder.SetCommonTaskSpec(task_id, name + ":" + empty_descriptor->CallString(), + Language::PYTHON, empty_descriptor, job_id, TaskID::Nil(), + 0, TaskID::Nil(), owner_address, 1, resource, resource, + PlacementGroupID::Nil()); builder.SetActorCreationTaskSpec(actor_id, max_restarts, {}, 1, detached, name); return builder.Build(); } diff --git a/src/ray/raylet/scheduling/cluster_task_manager_test.cc b/src/ray/raylet/scheduling/cluster_task_manager_test.cc --- a/src/ray/raylet/scheduling/cluster_task_manager_test.cc +++ b/src/ray/raylet/scheduling/cluster_task_manager_test.cc @@ -264,7 +264,7 @@ Task CreateTask(const std::unordered_map<std::string, double> &required_resource TaskID id = RandomTaskId(); JobID job_id = RandomJobId(); rpc::Address address; - spec_builder.SetCommonTaskSpec(id, Language::PYTHON, + spec_builder.SetCommonTaskSpec(id, "dummy_task", Language::PYTHON, FunctionDescriptorBuilder::BuildPython("", "", "", ""), job_id, TaskID::Nil(), 0, TaskID::Nil(), address, 0, required_resources, {}, PlacementGroupID::Nil()); @@ -489,4 +489,4 @@ int main(int argc, char **argv) { } } // namespace raylet -} // namespace ray \ No newline at end of file +} // namespace ray diff --git a/src/ray/raylet/task_dependency_manager_test.cc b/src/ray/raylet/task_dependency_manager_test.cc --- a/src/ray/raylet/task_dependency_manager_test.cc +++ b/src/ray/raylet/task_dependency_manager_test.cc @@ -65,7 +65,7 @@ static inline Task ExampleTask(const std::vector<ObjectID> &arguments, uint64_t num_returns) { TaskSpecBuilder builder; rpc::Address address; - builder.SetCommonTaskSpec(RandomTaskId(), Language::PYTHON, + builder.SetCommonTaskSpec(RandomTaskId(), "example_task", Language::PYTHON, FunctionDescriptorBuilder::BuildPython("", "", "", ""), JobID::Nil(), RandomTaskId(), 0, RandomTaskId(), address, num_returns, {}, {}, PlacementGroupID::Nil()); diff --git a/streaming/src/test/mock_actor.cc b/streaming/src/test/mock_actor.cc --- a/streaming/src/test/mock_actor.cc +++ b/streaming/src/test/mock_actor.cc @@ -496,8 +496,8 @@ class StreamingWorker { "", // driver_name "", // stdout_file "", // stderr_file - std::bind(&StreamingWorker::ExecuteTask, this, _1, _2, _3, _4, _5, _6, - _7), // task_execution_callback + std::bind(&StreamingWorker::ExecuteTask, this, _1, _2, _3, _4, _5, _6, _7, + _8), // task_execution_callback nullptr, // check_signals nullptr, // gc_collect nullptr, // spill_objects @@ -521,7 +521,8 @@ class StreamingWorker { } private: - Status ExecuteTask(TaskType task_type, const RayFunction &ray_function, + Status ExecuteTask(TaskType task_type, const std::string task_name, + const RayFunction &ray_function, const std::unordered_map<std::string, double> &required_resources, const std::vector<std::shared_ptr<RayObject>> &args, const std::vector<ObjectID> &arg_reference_ids, diff --git a/streaming/src/test/queue_tests_base.h b/streaming/src/test/queue_tests_base.h --- a/streaming/src/test/queue_tests_base.h +++ b/streaming/src/test/queue_tests_base.h @@ -87,7 +87,7 @@ class StreamingQueueTestBase : public ::testing::TestWithParam<uint64_t> { args.emplace_back(new TaskArgByValue(std::make_shared<RayObject>( msg.ToBytes(), nullptr, std::vector<ObjectID>(), true))); std::unordered_map<std::string, double> resources; - TaskOptions options{0, resources}; + TaskOptions options{"", 0, resources}; std::vector<ObjectID> return_ids; RayFunction func{ray::Language::PYTHON, ray::FunctionDescriptorBuilder::BuildPython("", "", "init", "")}; @@ -103,7 +103,7 @@ class StreamingQueueTestBase : public ::testing::TestWithParam<uint64_t> { args.emplace_back(new TaskArgByValue( std::make_shared<RayObject>(buffer, nullptr, std::vector<ObjectID>(), true))); std::unordered_map<std::string, double> resources; - TaskOptions options{0, resources}; + TaskOptions options{"", 0, resources}; std::vector<ObjectID> return_ids; RayFunction func{ray::Language::PYTHON, ray::FunctionDescriptorBuilder::BuildPython( "", test, "execute_test", "")}; @@ -119,7 +119,7 @@ class StreamingQueueTestBase : public ::testing::TestWithParam<uint64_t> { args.emplace_back(new TaskArgByValue( std::make_shared<RayObject>(buffer, nullptr, std::vector<ObjectID>(), true))); std::unordered_map<std::string, double> resources; - TaskOptions options{1, resources}; + TaskOptions options{"", 1, resources}; std::vector<ObjectID> return_ids; RayFunction func{ray::Language::PYTHON, ray::FunctionDescriptorBuilder::BuildPython( "", "", "check_current_test_status", "")};
[Core] Submission-time task names. <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### Feature Request Allow the user to specify submission-time task names, such that the task name is used as the process title (for displaying in the dashboard and debugging) and possibly other areas (e.g. core worker stats). Using the task function name as the user-friendly identifier results in poor UX when many tasks are executed using the same remote function, especially when there exists a semantically meaningful name that could be attached to the task at submission time. ### API and Semantics The user can specify a task name at submission time: ```python @ray.remote def foo(x): return x + 1 foo.options(name="my_special_foo").remote(1) ``` Unlike Ray actor names, this would **not** "detach" the task, and global uniqueness of these task names would **not** be enforced. Given the existing semantics for `actor.options(name="foo").remote()`, I'm imagining that this new task name option would only apply for actor and non-actor tasks, not for actor creation. When viewing the machine view of the Ray dashboard, the current task (in the workers/PID column) for a given core worker would be given by this task name, if set. The task name could also be included in other metrics/stats payloads, such as the core worker stats. ### Possible Implementation I think that this semantically meaningful task name should be easy to propagate if there's a dedicated field on the `TaskSpec`. If the user supplies a `name` at task submission time, the `task_name` field is set to that; if not, it's set to the task function's name, the current behavior. Flow - setting field in task spec + using said field to set proctitle: 1. Add a `task_name` field to the `TaskSpec` [class](https://github.com/ray-project/ray/blob/0178d6318ead643fff6d9ed873c88f71fdac52e7/src/ray/common/task/task_spec.h#L30-L216)/[proto](https://github.com/ray-project/ray/blob/fe6daef85e1f3cd2aef12b39487124a6e2dcc2b9/src/ray/protobuf/common.proto#L147-L189) and to the [`TaskOptions` class](https://github.com/ray-project/ray/blob/0178d6318ead643fff6d9ed873c88f71fdac52e7/src/ray/core_worker/common.h#L54-L64). 2. Properly set it on the task spec in [`CoreWorker::SubmitTask`](https://github.com/ray-project/ray/blob/0aec4cbccb5862a7808f98b90696236da9e2a03e/src/ray/core_worker/core_worker.cc#L1257-L1287) and [`CoreWorker::SubmitActorTask`](https://github.com/ray-project/ray/blob/0aec4cbccb5862a7808f98b90696236da9e2a03e/src/ray/core_worker/core_worker.cc#L1398-L1436) to either the user-provided `task_name` if given or set it to the function's name if not (current behavior). 3. Pull the `task_name` off of the `TaskSpec` class in [`CoreWorker::ExecuteTask`](https://github.com/ray-project/ray/blob/0aec4cbccb5862a7808f98b90696236da9e2a03e/src/ray/core_worker/core_worker.cc#L1635-L1757) and pass to language-specific task execution handler. 4. Use said `task_name` to [set the proctitle in the language-specific task execution handlers](https://github.com/ray-project/ray/blob/0aec4cbccb5862a7808f98b90696236da9e2a03e/python/ray/_raylet.pyx#L389). ### Concerns/Open Questions 1. Should the submission-time task name be appended or prepended to the function name, yielding something like `some_module.some_func:{task_name}` or `{task_name}:some_module.some_func`, instead of replacing the function name? Or should the user be expected to craft such a task name in their code and use _that_ as the `.options(name=task_name)` argument if that's what they want? I'm thinking the latter, but having to do that crafting for actor methods may be annoying. 2. Will this be confusing given the existence of named actors and the already established semantics thereof? Should a similar concept be provided for actors, and if so, how could we delineate between detached actors and named actors?
Btw, from 1.0 (the next version), you can name an actor that is not detached!
2020-08-31T17:52:59
ray-project/ray
10,456
ray-project__ray-10456
[ "10451" ]
91535e910201dc403cb2b8644a75e1132bd712fe
diff --git a/python/ray/tune/__init__.py b/python/ray/tune/__init__.py --- a/python/ray/tune/__init__.py +++ b/python/ray/tune/__init__.py @@ -15,6 +15,8 @@ from ray.tune.sample import (function, sample_from, uniform, quniform, choice, randint, qrandint, randn, qrandn, loguniform, qloguniform) +from ray.tune.suggest import create_searcher +from ray.tune.schedulers import create_scheduler __all__ = [ "Trainable", "DurableTrainable", "TuneError", "grid_search", @@ -24,5 +26,5 @@ "loguniform", "qloguniform", "ExperimentAnalysis", "Analysis", "CLIReporter", "JupyterNotebookReporter", "ProgressReporter", "report", "get_trial_dir", "get_trial_name", "get_trial_id", "make_checkpoint_dir", - "save_checkpoint", "checkpoint_dir" + "save_checkpoint", "checkpoint_dir", "create_searcher", "create_scheduler" ] diff --git a/python/ray/tune/schedulers/__init__.py b/python/ray/tune/schedulers/__init__.py --- a/python/ray/tune/schedulers/__init__.py +++ b/python/ray/tune/schedulers/__init__.py @@ -7,6 +7,69 @@ from ray.tune.schedulers.pbt import (PopulationBasedTraining, PopulationBasedTrainingReplay) + +def create_scheduler( + scheduler, + metric="episode_reward_mean", + mode="max", + **kwargs, +): + """Instantiate a scheduler based on the given string. + + This is useful for swapping between different schedulers. + + Args: + scheduler (str): The scheduler to use. + metric (str): The training result objective value attribute. Stopping + procedures will use this attribute. + mode (str): One of {min, max}. Determines whether objective is + minimizing or maximizing the metric attribute. + **kwargs: Additional parameters. + These keyword arguments will be passed to the initialization + function of the chosen class. + Returns: + ray.tune.schedulers.trial_scheduler.TrialScheduler: The scheduler. + Example: + >>> scheduler = tune.create_scheduler('pbt') + """ + + def _import_async_hyperband_scheduler(): + from ray.tune.schedulers import AsyncHyperBandScheduler + return AsyncHyperBandScheduler + + def _import_median_stopping_rule_scheduler(): + from ray.tune.schedulers import MedianStoppingRule + return MedianStoppingRule + + def _import_hyperband_scheduler(): + from ray.tune.schedulers import HyperBandScheduler + return HyperBandScheduler + + def _import_hb_bohb_scheduler(): + from ray.tune.schedulers import HyperBandForBOHB + return HyperBandForBOHB + + def _import_pbt_search(): + from ray.tune.schedulers import PopulationBasedTraining + return PopulationBasedTraining + + SCHEDULER_IMPORT = { + "async_hyperband": _import_async_hyperband_scheduler, + "median_stopping_rule": _import_median_stopping_rule_scheduler, + "hyperband": _import_hyperband_scheduler, + "hb_bohb": _import_hb_bohb_scheduler, + "pbt": _import_pbt_search, + } + scheduler = scheduler.lower() + if scheduler not in SCHEDULER_IMPORT: + raise ValueError( + f"Search alg must be one of {list(SCHEDULER_IMPORT)}. " + f"Got: {scheduler}") + + SchedulerClass = SCHEDULER_IMPORT[scheduler]() + return SchedulerClass(metric=metric, mode=mode, **kwargs) + + __all__ = [ "TrialScheduler", "HyperBandScheduler", "AsyncHyperBandScheduler", "ASHAScheduler", "MedianStoppingRule", "FIFOScheduler", diff --git a/python/ray/tune/suggest/__init__.py b/python/ray/tune/suggest/__init__.py --- a/python/ray/tune/suggest/__init__.py +++ b/python/ray/tune/suggest/__init__.py @@ -5,6 +5,94 @@ from ray.tune.suggest.variant_generator import grid_search from ray.tune.suggest.repeater import Repeater + +def create_searcher( + search_alg, + metric="episode_reward_mean", + mode="max", + **kwargs, +): + """Instantiate a search algorithm based on the given string. + + This is useful for swapping between different search algorithms. + + Args: + search_alg (str): The search algorithm to use. + metric (str): The training result objective value attribute. Stopping + procedures will use this attribute. + mode (str): One of {min, max}. Determines whether objective is + minimizing or maximizing the metric attribute. + **kwargs: Additional parameters. + These keyword arguments will be passed to the initialization + function of the chosen class. + Returns: + ray.tune.suggest.Searcher: The search algorithm. + Example: + >>> search_alg = tune.create_searcher('ax') + """ + + def _import_ax_search(): + from ray.tune.suggest.ax import AxSearch + return AxSearch + + def _import_dragonfly_search(): + from ray.tune.suggest.dragonfly import DragonflySearch + return DragonflySearch + + def _import_skopt_search(): + from ray.tune.suggest.skopt import SkOptSearch + return SkOptSearch + + def _import_hyperopt_search(): + from ray.tune.suggest.hyperopt import HyperOptSearch + return HyperOptSearch + + def _import_bayesopt_search(): + from ray.tune.suggest.bayesopt import BayesOptSearch + return BayesOptSearch + + def _import_bohb_search(): + from ray.tune.suggest.bohb import TuneBOHB + return TuneBOHB + + def _import_nevergrad_search(): + from ray.tune.suggest.nevergrad import NevergradSearch + return NevergradSearch + + def _import_optuna_search(): + from ray.tune.suggest.optuna import OptunaSearch + return OptunaSearch + + def _import_zoopt_search(): + from ray.tune.suggest.zoopt import ZOOptSearch + return ZOOptSearch + + def _import_sigopt_search(): + from ray.tune.suggest.sigopt import SigOptSearch + return SigOptSearch + + SEARCH_ALG_IMPORT = { + "ax": _import_ax_search, + "dragonfly": _import_dragonfly_search, + "skopt": _import_skopt_search, + "hyperopt": _import_hyperopt_search, + "bayesopt": _import_bayesopt_search, + "bohb": _import_bohb_search, + "nevergrad": _import_nevergrad_search, + "optuna": _import_optuna_search, + "zoopt": _import_zoopt_search, + "sigopt": _import_sigopt_search, + } + search_alg = search_alg.lower() + if search_alg not in SEARCH_ALG_IMPORT: + raise ValueError( + f"Search alg must be one of {list(SEARCH_ALG_IMPORT)}. " + f"Got: {search_alg}") + + SearcherClass = SEARCH_ALG_IMPORT[search_alg]() + return SearcherClass(metric=metric, mode=mode, **kwargs) + + __all__ = [ "SearchAlgorithm", "Searcher", "BasicVariantGenerator", "SearchGenerator", "grid_search", "Repeater", "ConcurrencyLimiter"
diff --git a/python/ray/tune/tests/test_api.py b/python/ray/tune/tests/test_api.py --- a/python/ray/tune/tests/test_api.py +++ b/python/ray/tune/tests/test_api.py @@ -14,7 +14,8 @@ from ray.tune import (DurableTrainable, Trainable, TuneError, Stopper, EarlyStopping) from ray.tune import register_env, register_trainable, run_experiments -from ray.tune.schedulers import TrialScheduler, FIFOScheduler +from ray.tune.schedulers import (TrialScheduler, FIFOScheduler, + AsyncHyperBandScheduler) from ray.tune.trial import Trial from ray.tune.result import (TIMESTEPS_TOTAL, DONE, HOSTNAME, NODE_IP, PID, EPISODES_TOTAL, TRAINING_ITERATION, @@ -24,6 +25,8 @@ from ray.tune.experiment import Experiment from ray.tune.resources import Resources from ray.tune.suggest import grid_search +from ray.tune.suggest.hyperopt import HyperOptSearch +from ray.tune.suggest.ax import AxSearch from ray.tune.suggest._mock import _MockSuggestionAlgorithm from ray.tune.utils import (flatten_dict, get_pinned_object, pin_in_object_store) @@ -1105,6 +1108,30 @@ def train(config, reporter): self.assertIn("LOG_STDERR", content) +class ShimCreationTest(unittest.TestCase): + def testCreateScheduler(self): + kwargs = {"metric": "metric_foo", "mode": "min"} + + scheduler = "async_hyperband" + shim_scheduler = tune.create_scheduler(scheduler, **kwargs) + real_scheduler = AsyncHyperBandScheduler(**kwargs) + assert type(shim_scheduler) is type(real_scheduler) + + def testCreateSearcher(self): + kwargs = {"metric": "metric_foo", "mode": "min"} + + searcher_ax = "ax" + shim_searcher_ax = tune.create_searcher(searcher_ax, **kwargs) + real_searcher_ax = AxSearch(space=[], **kwargs) + assert type(shim_searcher_ax) is type(real_searcher_ax) + + searcher_hyperopt = "hyperopt" + shim_searcher_hyperopt = tune.create_searcher(searcher_hyperopt, + **kwargs) + real_searcher_hyperopt = HyperOptSearch({}, **kwargs) + assert type(shim_searcher_hyperopt) is type(real_searcher_hyperopt) + + if __name__ == "__main__": import pytest import sys
[tune] shim instantiation of search algorithms ### Describe your feature request allow creating a search algorithm for `tune.run` given: - (maybe; see option 1 below) a string defining the search algorithm to use - (maybe; see option 2 below) the search space configuration ### API there will be two ways to use shim instantiation (option 3 from the design doc): #### Option 1 (notice how the search algorithm definition doesn't require redefining the search configuration when initializing `HyperOptSearch`) > ```python > from ray import tune > from ray.tune.param import Float, Integer, Categorical > from ray.tune.param import Grid, Uniform, LogUniform, Normal > > space = { > "lr": Float(1e-4, 1e-1).LogUniform(), > "num_epochs": Integer(5, 10).Uniform(), > "temperature": Float().Normal(1, 0.1), > "batch_size": Categorical([32, 64]).Grid() > } > > tune.run( > trainable, > space, > search_alg=HyperOptSearch(metric="mean_loss")) > ``` #### Option 2 > ```python > from ray import tune > from ray.tune.param import Float, Integer, Categorical > from ray.tune.param import Grid, Uniform, LogUniform, Normal > > space = { > "lr": Float(1e-4, 1e-1).LogUniform(), > "num_epochs": Integer(5, 10).Uniform(), > "temperature": Float().Normal(1, 0.1), > "batch_size": Categorical([32, 64]).Grid() > } > > search_alg = tune.create_searcher("HyperOpt", space, > metric="mean_loss") > > tune.run( > trainable, > search_alg=search_alg) > ``` --- related: #9969, #10401, #10444 CC @richardliaw @krfricke
```python from ray import tune from ray.tune.param import Float, Integer, Categorical from ray.tune.param import Grid, Uniform, LogUniform, Normal space = { "lr": Float(1e-4, 1e-1).LogUniform(), "num_epochs": Integer(5, 10).Uniform(), "temperature": Float().Normal(1, 0.1), "batch_size": Categorical([32, 64]).Grid() } ############################### # Shim API - Option 3 ################################ tune.run( trainable, space, search_alg=tune.create_searcher("HyperOpt", metric="mean_loss")) # This should also work. # tune.run( # trainable, # space, # search_alg=HyperOptSearch(metric="mean_loss")) ``` untested; we should use an if-statement instead of `getattr`: ```python def create_searcher(search_alg: str, **kwargs): from ray.tune import suggest return getattr(suggest, f'{search_alg.lower()}.{search_alg}Search')(None, **kwargs) # in HyperOptSearch.__init__ we check if space is None, and if so, set HyperOptSearch.domain to None # in tune.run, we check if `HyperOptSearch.domain` is None, and if so, we set assign the proper config based on the config kwarg using PR #10444 ``` instead of this getattr, can you just create a dictionary? ```python def import_hyperopt_search(): from ray.tune.suggest import HyperOptSearch return HyperOptSearch SEARCH_ALG_IMPORT = { "hyperopt": import_hyperopt_search, "dragonfly": import_dragonfly_search, } ``` This avoids this dependency on naming convention. ```python def create_searcher(search_alg, **kwargs): # TODO: docstring def _import_ax_search(): from ray.tune.suggest.ax import AxSearch return AxSearch def _import_dragonfly_search(): from ray.tune.suggest.dragonfly import DragonflySearch return DragonflySearch def _import_skopt_search(): from ray.tune.suggest.skopt import SkOptSearch return SkOptSearch def _import_hyperopt_search(): from ray.tune.suggest import HyperOptSearch return HyperOptSearch def _import_bayesopt_search(): from ray.tune.suggest.bayesopt import BayesOptSearch return BayesOptSearch def _import_bohb_search(): from ray.tune.suggest.bohb import TuneBOHB return TuneBOHB def _import_nevergrad_search(): from ray.tune.suggest.nevergrad import NevergradSearch return NevergradSearch def _import_optuna_search(): from ray.tune.suggest.optuna import OptunaSearch return OptunaSearch def _import_zoopt_search(): from ray.tune.suggest.zoopt import ZOOptSearch return ZOOptSearch def _import_sigopt_search(): from ray.tune.suggest.sigopt import SigOptSearch return SigOptSearch SEARCH_ALG_IMPORT = { "ax": _import_ax_search, "dragonfly": _import_dragonfly_search, "skopt": _import_skopt_search, "hyperopt": _import_hyperopt_search, "bayesopt": _import_bayesopt_search, "bohb": _import_bohb_search, "nevergrad": _import_nevergrad_search, "optuna": _import_optuna_search, "zoopt": _import_zoopt_search, "sigopt": _import_sigopt_search, } return SEARCH_ALG_IMPORT[search_alg](**kwargs) ``` Look good to me! Consider moving some of the default arguments (i.e., metric, mode) also into the wrapper so that it's explicit and captured in the docstring. could you push a PR @sumanthratna and tag me when you do? Thanks!
2020-08-31T22:43:50
ray-project/ray
10,464
ray-project__ray-10464
[ "10408" ]
3f98a8bfcb0c47c8a70458f1251761b0f6ec8195
diff --git a/python/ray/autoscaler/autoscaler.py b/python/ray/autoscaler/autoscaler.py --- a/python/ray/autoscaler/autoscaler.py +++ b/python/ray/autoscaler/autoscaler.py @@ -205,11 +205,15 @@ def _update(self): self.resource_demand_scheduler.get_nodes_to_launch( self.provider.non_terminated_nodes(tag_filters={}), self.pending_launches.breakdown(), - resource_demand_vector)) + resource_demand_vector, + self.load_metrics.get_resource_utilization())) # TODO(ekl) also enforce max launch concurrency here? for node_type, count in to_launch: self.launch_new_node(count, node_type=node_type) + num_pending = self.pending_launches.value + nodes = self.workers() + # Launch additional nodes of the default type, if still needed. num_workers = len(nodes) + num_pending if num_workers < target_workers: @@ -483,7 +487,8 @@ def log_info_string(self, nodes, target): tmp += "\n" if self.resource_demand_scheduler: tmp += self.resource_demand_scheduler.debug_string( - nodes, self.pending_launches.breakdown()) + nodes, self.pending_launches.breakdown(), + self.load_metrics.get_resource_utilization()) if _internal_kv_initialized(): _internal_kv_put(DEBUG_AUTOSCALING_STATUS, tmp, overwrite=True) logger.info(tmp) diff --git a/python/ray/autoscaler/load_metrics.py b/python/ray/autoscaler/load_metrics.py --- a/python/ray/autoscaler/load_metrics.py +++ b/python/ray/autoscaler/load_metrics.py @@ -16,13 +16,14 @@ class LoadMetrics: can be removed. """ - def __init__(self): + def __init__(self, local_ip=None): self.last_used_time_by_ip = {} self.last_heartbeat_time_by_ip = {} self.static_resources_by_ip = {} self.dynamic_resources_by_ip = {} self.resource_load_by_ip = {} - self.local_ip = services.get_node_ip_address() + self.local_ip = services.get_node_ip_address( + ) if local_ip is None else local_ip self.waiting_bundles = [] self.infeasible_bundles = [] @@ -98,7 +99,10 @@ def get_node_resources(self): """ return self.static_resources_by_ip.values() - def get_resource_usage(self): + def get_resource_utilization(self): + return self.dynamic_resources_by_ip + + def _get_resource_usage(self): num_nodes = len(self.static_resources_by_ip) nodes_used = 0.0 num_nonidle = 0 @@ -145,7 +149,8 @@ def info_string(self): ["{}: {}".format(k, v) for k, v in sorted(self._info().items())]) def _info(self): - nodes_used, resources_used, resources_total = self.get_resource_usage() + nodes_used, resources_used, resources_total = self._get_resource_usage( + ) now = time.time() idle_times = [now - t for t in self.last_used_time_by_ip.values()] diff --git a/python/ray/autoscaler/resource_demand_scheduler.py b/python/ray/autoscaler/resource_demand_scheduler.py --- a/python/ray/autoscaler/resource_demand_scheduler.py +++ b/python/ray/autoscaler/resource_demand_scheduler.py @@ -39,11 +39,10 @@ def __init__(self, provider: NodeProvider, self.node_types = node_types self.max_workers = max_workers - # TODO(ekl) take into account existing utilization of node resources. We - # should subtract these from node resources prior to running bin packing. def get_nodes_to_launch(self, nodes: List[NodeID], pending_nodes: Dict[NodeType, int], - resource_demands: List[ResourceDict] + resource_demands: List[ResourceDict], + usage_by_ip: Dict[str, ResourceDict] ) -> List[Tuple[NodeType, int]]: """Given resource demands, return node types to add to the cluster. @@ -64,7 +63,8 @@ def get_nodes_to_launch(self, nodes: List[NodeID], return [] node_resources, node_type_counts = self.calculate_node_resources( - nodes, pending_nodes) + nodes, pending_nodes, usage_by_ip) + logger.info("Cluster resources: {}".format(node_resources)) logger.info("Node counts: {}".format(node_type_counts)) @@ -72,34 +72,43 @@ def get_nodes_to_launch(self, nodes: List[NodeID], logger.info("Resource demands: {}".format(resource_demands)) logger.info("Unfulfilled demands: {}".format(unfulfilled)) - nodes = get_nodes_for(self.node_types, node_type_counts, - self.max_workers - len(nodes), unfulfilled) + nodes = get_nodes_for( + self.node_types, node_type_counts, + self.max_workers - len(nodes) - sum(pending_nodes.values()), + unfulfilled) logger.info("Node requests: {}".format(nodes)) return nodes def calculate_node_resources( - self, nodes: List[NodeID], pending_nodes: Dict[NodeID, int] + self, nodes: List[NodeID], pending_nodes: Dict[NodeID, int], + usage_by_ip: Dict[str, ResourceDict] ) -> (List[ResourceDict], Dict[NodeType, int]): """Returns node resource list and node type counts.""" node_resources = [] node_type_counts = collections.defaultdict(int) - def add_node(node_type): + def add_node(node_type, existing_resource_usages=None): if node_type not in self.node_types: raise RuntimeError("Missing entry for node_type {} in " "available_node_types config: {}".format( node_type, self.node_types)) # Careful not to include the same dict object multiple times. - node_resources.append( - copy.deepcopy(self.node_types[node_type]["resources"])) + available = copy.deepcopy(self.node_types[node_type]["resources"]) + if existing_resource_usages: + for resource, used in existing_resource_usages.items(): + available[resource] -= used + + node_resources.append(available) node_type_counts[node_type] += 1 for node_id in nodes: tags = self.provider.node_tags(node_id) if TAG_RAY_USER_NODE_TYPE in tags: node_type = tags[TAG_RAY_USER_NODE_TYPE] - add_node(node_type) + ip = self.provider.internal_ip(node_id) + resources = usage_by_ip.get(ip, {}) + add_node(node_type, resources) for node_type, count in pending_nodes.items(): for _ in range(count): @@ -108,9 +117,11 @@ def add_node(node_type): return node_resources, node_type_counts def debug_string(self, nodes: List[NodeID], - pending_nodes: Dict[NodeID, int]) -> str: + pending_nodes: Dict[NodeID, int], + usage_by_ip: Dict[str, ResourceDict]) -> str: + print(f"{usage_by_ip}") node_resources, node_type_counts = self.calculate_node_resources( - nodes, pending_nodes) + nodes, pending_nodes, usage_by_ip) out = "Worker node types:" for node_type, count in node_type_counts.items():
diff --git a/python/ray/tests/test_multi_node_2.py b/python/ray/tests/test_multi_node_2.py --- a/python/ray/tests/test_multi_node_2.py +++ b/python/ray/tests/test_multi_node_2.py @@ -76,7 +76,7 @@ def setup_monitor(address): def verify_load_metrics(monitor, expected_resource_usage=None, timeout=30): while True: monitor.process_messages() - resource_usage = monitor.load_metrics.get_resource_usage() + resource_usage = monitor.load_metrics._get_resource_usage() if "memory" in resource_usage[1]: del resource_usage[1]["memory"] diff --git a/python/ray/tests/test_resource_demand_scheduler.py b/python/ray/tests/test_resource_demand_scheduler.py --- a/python/ray/tests/test_resource_demand_scheduler.py +++ b/python/ray/tests/test_resource_demand_scheduler.py @@ -14,7 +14,7 @@ from ray.autoscaler.commands import get_or_create_head_node from ray.autoscaler.tags import TAG_RAY_USER_NODE_TYPE, TAG_RAY_NODE_KIND from ray.autoscaler.resource_demand_scheduler import _utilization_score, \ - get_bin_pack_residual, get_nodes_for + get_bin_pack_residual, get_nodes_for, ResourceDemandScheduler from ray.test_utils import same_elements from time import sleep @@ -163,6 +163,23 @@ def test_get_nodes_respects_max_limit(): }] * 10) == [("m4.large", 2)] +def test_get_nodes_to_launch_limits(): + provider = MockProvider() + scheduler = ResourceDemandScheduler(provider, TYPES_A, 3) + + provider.create_node({}, {TAG_RAY_USER_NODE_TYPE: "p2.8xlarge"}, 2) + + nodes = provider.non_terminated_nodes({}) + + ips = provider.non_terminated_node_ips({}) + utilizations = {ip: {"GPU": 8} for ip in ips} + + to_launch = scheduler.get_nodes_to_launch(nodes, {"p2.8xlarge": 1}, [{ + "GPU": 8 + }] * 2, utilizations) + assert to_launch == [] + + class LoadMetricsTest(unittest.TestCase): def testResourceDemandVector(self): lm = LoadMetrics() @@ -264,6 +281,44 @@ def testScaleUpMinSanity(self): autoscaler.update() self.waitForNodes(2) + def testScaleUpIgnoreUsed(self): + config = MULTI_WORKER_CLUSTER.copy() + # Commenting out this line causes the test case to fail?!?! + config["min_workers"] = 0 + config["target_utilization_fraction"] = 1.0 + config_path = self.write_config(config) + self.provider = MockProvider() + self.provider.create_node({}, { + TAG_RAY_NODE_KIND: "head", + TAG_RAY_USER_NODE_TYPE: "p2.xlarge" + }, 1) + head_ip = self.provider.non_terminated_node_ips({})[0] + self.provider.finish_starting_nodes() + runner = MockProcessRunner() + lm = LoadMetrics(local_ip=head_ip) + autoscaler = StandardAutoscaler( + config_path, + lm, + max_failures=0, + process_runner=runner, + update_interval_s=0) + autoscaler.update() + self.waitForNodes(1) + lm.update(head_ip, {"CPU": 4, "GPU": 1}, {}, {}) + self.waitForNodes(1) + + lm.update( + head_ip, { + "CPU": 4, + "GPU": 1 + }, {"GPU": 1}, {}, + waiting_bundles=[{ + "GPU": 1 + }]) + autoscaler.update() + self.waitForNodes(2) + assert self.provider.mock_nodes[1].node_type == "p2.xlarge" + def testRequestBundlesAccountsForHeadNode(self): config = MULTI_WORKER_CLUSTER.copy() config["head_node_type"] = "p2.8xlarge"
[autoscaler] Multi node autoscaler doesn't take into account existing resource usage <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? When we calculate the resources available in the cluster, we should take into account the existing usage from `LoadMetrics`: https://github.com/ray-project/ray/blob/master/python/ray/autoscaler/resource_demand_scheduler.py#L54 Otherwise, we might not scale up even when tasks cannot be scheduled. This should be fixed: 1. Subtract the usage of the node when calculating node resources by looking at LoadMetrics. 2. Add unit tests for this scenario.
2020-09-01T07:31:15
ray-project/ray
10,499
ray-project__ray-10499
[ "10521" ]
ef18893fb50c97c2c0d9d7b8b951455683f89635
diff --git a/rllib/agents/ddpg/ddpg_tf_policy.py b/rllib/agents/ddpg/ddpg_tf_policy.py --- a/rllib/agents/ddpg/ddpg_tf_policy.py +++ b/rllib/agents/ddpg/ddpg_tf_policy.py @@ -1,4 +1,5 @@ from gym.spaces import Box +from functools import partial import logging import numpy as np @@ -290,7 +291,8 @@ def gradients_fn(policy, optimizer, loss): # Clip if necessary. if policy.config["grad_clip"]: - clip_func = tf.clip_by_norm + clip_func = partial( + tf.clip_by_norm, clip_norm=policy.config["grad_clip"]) else: clip_func = tf.identity diff --git a/rllib/agents/sac/sac_tf_policy.py b/rllib/agents/sac/sac_tf_policy.py --- a/rllib/agents/sac/sac_tf_policy.py +++ b/rllib/agents/sac/sac_tf_policy.py @@ -1,4 +1,5 @@ from gym.spaces import Box, Discrete +from functools import partial import logging import ray @@ -322,7 +323,8 @@ def gradients_fn(policy, optimizer, loss): # Clip if necessary. if policy.config["grad_clip"]: - clip_func = tf.clip_by_norm + clip_func = partial( + tf.clip_by_norm, clip_norm=policy.config["grad_clip"]) else: clip_func = tf.identity
[rllib] SAC and DDPG policy can't do `grad_clip` <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? *Ray version and other system information (Python version, TensorFlow version, OS):* Ray: 0.8.7 Python: 3.7 Tensorflow: 2.2.0 OS: Manjaro x86_64 ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): ````py from ray import tune from ray.rllib.agents.sac import SACTrainer from ray.rllib.agents.ddpg import DDPGTrainer tune.run(DDPGTrainer, config={"env": "CartPole-v0","grad_clip":40.0}) # tune.run(SACTrainer, config={"env": "CartPole-v0","grad_clip":40.0}) ```` This [PR](https://github.com/ray-project/ray/pull/10499) will fix this issue. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-09-02T13:07:23
ray-project/ray
10,504
ray-project__ray-10504
[ "10313" ]
ef18893fb50c97c2c0d9d7b8b951455683f89635
diff --git a/python/ray/autoscaler/autoscaler.py b/python/ray/autoscaler/autoscaler.py --- a/python/ray/autoscaler/autoscaler.py +++ b/python/ray/autoscaler/autoscaler.py @@ -1,5 +1,5 @@ -from collections import defaultdict -from typing import Optional +from collections import defaultdict, namedtuple +from typing import Any, Optional import copy import logging import math @@ -30,6 +30,12 @@ logger = logging.getLogger(__name__) +# Tuple of modified fields for the given node_id returned by should_update +# that will be passed into a NodeUpdaterThread. +UpdateInstructions = namedtuple( + "UpdateInstructions", + ["node_id", "init_commands", "start_ray_commands", "docker_config"]) + class StandardAutoscaler: """The autoscaling control loop for a Ray cluster. @@ -236,14 +242,15 @@ def _update(self): # problems. They should at a minimum be spawned as daemon threads. # See https://github.com/ray-project/ray/pull/5903 for more info. T = [] - for node_id, commands, ray_start in (self.should_update(node_id) - for node_id in nodes): + for node_id, commands, ray_start, docker_config in ( + self.should_update(node_id) for node_id in nodes): if node_id is not None: resources = self._node_resources(node_id) T.append( threading.Thread( target=self.spawn_updater, - args=(node_id, commands, ray_start, resources))) + args=(node_id, commands, ray_start, resources, + docker_config))) for t in T: t.start() for t in T: @@ -401,44 +408,56 @@ def recover_if_needed(self, node_id, now): updater.start() self.updaters[node_id] = updater - def _get_node_type_specific_commands(self, node_id: str, - commands_key: str): - commands = self.config[commands_key] + def _get_node_type_specific_fields(self, node_id: str, + fields_key: str) -> Any: + fields = self.config[fields_key] node_tags = self.provider.node_tags(node_id) if TAG_RAY_USER_NODE_TYPE in node_tags: node_type = node_tags[TAG_RAY_USER_NODE_TYPE] if node_type not in self.available_node_types: raise ValueError(f"Unknown node type tag: {node_type}.") node_specific_config = self.available_node_types[node_type] - if commands_key in node_specific_config: - commands = node_specific_config[commands_key] - return commands + if fields_key in node_specific_config: + fields = node_specific_config[fields_key] + return fields + + def _get_node_specific_docker_config(self, node_id): + docker_config = copy.deepcopy(self.config.get("docker", {})) + node_specific_docker = self._get_node_type_specific_fields( + node_id, "docker") + docker_config.update(node_specific_docker) + return docker_config def should_update(self, node_id): if not self.can_update(node_id): - return None, None, None # no update + return UpdateInstructions(None, None, None, None) # no update status = self.provider.node_tags(node_id).get(TAG_RAY_NODE_STATUS) if status == STATUS_UP_TO_DATE and self.files_up_to_date(node_id): - return None, None, None # no update + return UpdateInstructions(None, None, None, None) # no update successful_updated = self.num_successful_updates.get(node_id, 0) > 0 if successful_updated and self.config.get("restart_only", False): init_commands = [] ray_commands = self.config["worker_start_ray_commands"] elif successful_updated and self.config.get("no_restart", False): - init_commands = self._get_node_type_specific_commands( + init_commands = self._get_node_type_specific_fields( node_id, "worker_setup_commands") ray_commands = [] else: - init_commands = self._get_node_type_specific_commands( + init_commands = self._get_node_type_specific_fields( node_id, "worker_setup_commands") ray_commands = self.config["worker_start_ray_commands"] - return (node_id, init_commands, ray_commands) + docker_config = self._get_node_specific_docker_config(node_id) + return UpdateInstructions( + node_id=node_id, + init_commands=init_commands, + start_ray_commands=ray_commands, + docker_config=docker_config) def spawn_updater(self, node_id, init_commands, ray_start_commands, - node_resources): + node_resources, docker_config): updater = NodeUpdaterThread( node_id=node_id, provider_config=self.config["provider"], @@ -447,7 +466,7 @@ def spawn_updater(self, node_id, init_commands, ray_start_commands, cluster_name=self.config["cluster_name"], file_mounts=self.config["file_mounts"], initialization_commands=with_head_node_ip( - self._get_node_type_specific_commands( + self._get_node_type_specific_fields( node_id, "initialization_commands")), setup_commands=with_head_node_ip(init_commands), ray_start_commands=with_head_node_ip(ray_start_commands), @@ -457,7 +476,7 @@ def spawn_updater(self, node_id, init_commands, ray_start_commands, cluster_synced_files=self.config["cluster_synced_files"], process_runner=self.process_runner, use_internal_ip=True, - docker_config=self.config.get("docker"), + docker_config=docker_config, node_resources=node_resources) updater.start() self.updaters[node_id] = updater diff --git a/python/ray/autoscaler/command_runner.py b/python/ray/autoscaler/command_runner.py --- a/python/ray/autoscaler/command_runner.py +++ b/python/ray/autoscaler/command_runner.py @@ -705,9 +705,8 @@ def _docker_expand_user(self, string, any_char=False): def run_init(self, *, as_head, file_mounts): image = self.docker_config.get("image") - if image is None: - image = self.docker_config.get( - f"{'head' if as_head else 'worker'}_image") + image = self.docker_config.get( + f"{'head' if as_head else 'worker'}_image", image) self._check_docker_installed() if self.docker_config.get("pull_before_run", True):
diff --git a/python/ray/tests/test_resource_demand_scheduler.py b/python/ray/tests/test_resource_demand_scheduler.py --- a/python/ray/tests/test_resource_demand_scheduler.py +++ b/python/ray/tests/test_resource_demand_scheduler.py @@ -449,6 +449,86 @@ def testCommandPassing(self): runner.assert_not_has_call(self.provider.mock_nodes[2].internal_ip, "init_cmd") + def testDockerWorkers(self): + config = MULTI_WORKER_CLUSTER.copy() + config["available_node_types"]["p2.8xlarge"]["docker"] = { + "worker_image": "p2.8x_image:latest", + "worker_run_options": ["p2.8x-run-options"] + } + config["available_node_types"]["p2.xlarge"]["docker"] = { + "worker_image": "p2x_image:nightly" + } + config["docker"]["worker_run_options"] = ["standard-run-options"] + config["docker"]["image"] = "default-image:nightly" + config["docker"]["worker_image"] = "default-image:nightly" + # Commenting out this line causes the test case to fail?!?! + config["min_workers"] = 0 + config["max_workers"] = 10 + config_path = self.write_config(config) + self.provider = MockProvider() + runner = MockProcessRunner() + autoscaler = StandardAutoscaler( + config_path, + LoadMetrics(), + max_failures=0, + process_runner=runner, + update_interval_s=0) + assert len(self.provider.non_terminated_nodes({})) == 0 + autoscaler.update() + self.waitForNodes(0) + autoscaler.request_resources([{"CPU": 1}]) + autoscaler.update() + self.waitForNodes(1) + assert self.provider.mock_nodes[0].node_type == "m4.large" + autoscaler.request_resources([{"GPU": 8}]) + autoscaler.update() + self.waitForNodes(2) + assert self.provider.mock_nodes[1].node_type == "p2.8xlarge" + autoscaler.request_resources([{"GPU": 1}] * 9) + autoscaler.update() + self.waitForNodes(3) + assert self.provider.mock_nodes[2].node_type == "p2.xlarge" + autoscaler.update() + # Fill up m4, p2.8, p2 and request 2 more CPUs + autoscaler.request_resources([{ + "CPU": 2 + }, { + "CPU": 16 + }, { + "CPU": 32 + }, { + "CPU": 2 + }]) + autoscaler.update() + self.waitForNodes(4) + assert self.provider.mock_nodes[3].node_type == "m4.16xlarge" + autoscaler.update() + sleep(0.1) + runner.assert_has_call(self.provider.mock_nodes[1].internal_ip, + "p2.8x-run-options") + runner.assert_has_call(self.provider.mock_nodes[1].internal_ip, + "p2.8x_image:latest") + runner.assert_not_has_call(self.provider.mock_nodes[1].internal_ip, + "default-image:nightly") + runner.assert_not_has_call(self.provider.mock_nodes[1].internal_ip, + "standard-run-options") + + runner.assert_has_call(self.provider.mock_nodes[2].internal_ip, + "p2x_image:nightly") + runner.assert_has_call(self.provider.mock_nodes[2].internal_ip, + "standard-run-options") + runner.assert_not_has_call(self.provider.mock_nodes[2].internal_ip, + "p2.8x-run-options") + + runner.assert_has_call(self.provider.mock_nodes[3].internal_ip, + "default-image:nightly") + runner.assert_has_call(self.provider.mock_nodes[3].internal_ip, + "standard-run-options") + runner.assert_not_has_call(self.provider.mock_nodes[3].internal_ip, + "p2.8x-run-options") + runner.assert_not_has_call(self.provider.mock_nodes[3].internal_ip, + "p2x_image:nightly") + def testUpdateConfig(self): config = MULTI_WORKER_CLUSTER.copy() config_path = self.write_config(config)
[Autoscaler] Node specific docker options <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### Describe your feature request We should allow node specific docker options for the autoscaler (similar to provider options, commands, resources, etc.)
cc @ijrsvt This is probably post-ray1.0 @yiranwang52 FYI
2020-09-02T17:24:41
ray-project/ray
10,513
ray-project__ray-10513
[ "10404" ]
3eed3eca09bfbbb3955e1c7757d480d1e123aceb
diff --git a/python/ray/autoscaler/autoscaler.py b/python/ray/autoscaler/autoscaler.py --- a/python/ray/autoscaler/autoscaler.py +++ b/python/ray/autoscaler/autoscaler.py @@ -13,10 +13,11 @@ from ray.experimental.internal_kv import _internal_kv_put, \ _internal_kv_initialized from ray.autoscaler.node_provider import get_node_provider -from ray.autoscaler.tags import ( - TAG_RAY_LAUNCH_CONFIG, TAG_RAY_RUNTIME_CONFIG, - TAG_RAY_FILE_MOUNTS_CONTENTS, TAG_RAY_NODE_STATUS, TAG_RAY_NODE_KIND, - TAG_RAY_USER_NODE_TYPE, STATUS_UP_TO_DATE, NODE_KIND_WORKER) +from ray.autoscaler.tags import (TAG_RAY_LAUNCH_CONFIG, TAG_RAY_RUNTIME_CONFIG, + TAG_RAY_FILE_MOUNTS_CONTENTS, + TAG_RAY_NODE_STATUS, TAG_RAY_NODE_KIND, + TAG_RAY_USER_NODE_TYPE, STATUS_UP_TO_DATE, + NODE_KIND_WORKER, NODE_KIND_UNMANAGED) from ray.autoscaler.updater import NodeUpdaterThread from ray.autoscaler.node_launcher import NodeLauncher from ray.autoscaler.resource_demand_scheduler import ResourceDemandScheduler @@ -146,10 +147,12 @@ def _update(self): nodes = self.workers() # Check pending nodes immediately after fetching the number of running # nodes to minimize chance number of pending nodes changing after - # additional nodes are launched. + # additional nodes (managed and unmanaged) are launched. num_pending = self.pending_launches.value - self.load_metrics.prune_active_ips( - [self.provider.internal_ip(node_id) for node_id in nodes]) + self.load_metrics.prune_active_ips([ + self.provider.internal_ip(node_id) + for node_id in self.all_workers() + ]) target_workers = self.target_num_workers() if len(nodes) >= target_workers: @@ -165,8 +168,9 @@ def _update(self): nodes_to_terminate = [] for node_id in nodes: node_ip = self.provider.internal_ip(node_id) - if node_ip in last_used and last_used[node_ip] < horizon and \ - len(nodes) - len(nodes_to_terminate) > target_workers: + if (node_ip in last_used and last_used[node_ip] < horizon) and \ + (len(nodes) - len(nodes_to_terminate) + > target_workers): logger.info("StandardAutoscaler: " "{}: Terminating idle node".format(node_id)) nodes_to_terminate.append(node_id) @@ -182,11 +186,12 @@ def _update(self): # Terminate nodes if there are too many nodes_to_terminate = [] - while len(nodes) > self.config["max_workers"]: + while (len(nodes) - + len(nodes_to_terminate)) > self.config["max_workers"] and nodes: + to_terminate = nodes.pop() logger.info("StandardAutoscaler: " - "{}: Terminating unneeded node".format(nodes[-1])) - nodes_to_terminate.append(nodes[-1]) - nodes = nodes[:-1] + "{}: Terminating unneeded node".format(to_terminate)) + nodes_to_terminate.append(to_terminate) if nodes_to_terminate: self.provider.terminate_nodes(nodes_to_terminate) @@ -511,10 +516,17 @@ def launch_new_node(self, count: int, node_type: Optional[str]) -> None: config = copy.deepcopy(self.config) self.launch_queue.put((config, count, node_type)) + def all_workers(self): + return self.workers() + self.unmanaged_workers() + def workers(self): return self.provider.non_terminated_nodes( tag_filters={TAG_RAY_NODE_KIND: NODE_KIND_WORKER}) + def unmanaged_workers(self): + return self.provider.non_terminated_nodes( + tag_filters={TAG_RAY_NODE_KIND: NODE_KIND_UNMANAGED}) + def log_info_string(self, nodes, target): tmp = "Cluster status: " tmp += self.info_string(nodes, target) diff --git a/python/ray/autoscaler/resource_demand_scheduler.py b/python/ray/autoscaler/resource_demand_scheduler.py --- a/python/ray/autoscaler/resource_demand_scheduler.py +++ b/python/ray/autoscaler/resource_demand_scheduler.py @@ -14,7 +14,7 @@ from typing import List, Dict, Tuple from ray.autoscaler.node_provider import NodeProvider -from ray.autoscaler.tags import TAG_RAY_USER_NODE_TYPE +from ray.autoscaler.tags import TAG_RAY_USER_NODE_TYPE, NODE_KIND_UNMANAGED logger = logging.getLogger(__name__) @@ -90,9 +90,14 @@ def calculate_node_resources( def add_node(node_type, existing_resource_usages=None): if node_type not in self.node_types: - raise RuntimeError("Missing entry for node_type {} in " - "available_node_types config: {}".format( - node_type, self.node_types)) + logger.warn( + f"Missing entry for node_type {node_type} in " + f"cluster config: {self.node_types} under entry " + f"available_node_types. This node's resources will be " + f"ignored. If you are using an unmanaged node, manually " + f"set the user_node_type tag to \"{NODE_KIND_UNMANAGED}\"" + f"in your cloud provider's management console.") + return None # Careful not to include the same dict object multiple times. available = copy.deepcopy(self.node_types[node_type]["resources"]) if existing_resource_usages: diff --git a/python/ray/autoscaler/tags.py b/python/ray/autoscaler/tags.py --- a/python/ray/autoscaler/tags.py +++ b/python/ray/autoscaler/tags.py @@ -8,6 +8,7 @@ TAG_RAY_NODE_KIND = "ray-node-type" NODE_KIND_HEAD = "head" NODE_KIND_WORKER = "worker" +NODE_KIND_UNMANAGED = "unmanaged" # Tag for user defined node types (e.g., m4xl_spot). This is used for multi # node type clusters.
diff --git a/python/ray/tests/test_autoscaler.py b/python/ray/tests/test_autoscaler.py --- a/python/ray/tests/test_autoscaler.py +++ b/python/ray/tests/test_autoscaler.py @@ -597,6 +597,50 @@ def testAggressiveAutoscaling(self): autoscaler.update() self.waitForNodes(11) + def testUnmanagedNodes(self): + config = SMALL_CLUSTER.copy() + config["min_workers"] = 0 + config["max_workers"] = 20 + config["initial_workers"] = 0 + config["idle_timeout_minutes"] = 0 + config["autoscaling_mode"] = "aggressive" + config["target_utilization_fraction"] = 0.8 + config_path = self.write_config(config) + + self.provider = MockProvider() + self.provider.create_node({}, {TAG_RAY_NODE_KIND: "head"}, 1) + head_ip = self.provider.non_terminated_node_ips( + tag_filters={TAG_RAY_NODE_KIND: "head"}, )[0] + + self.provider.create_node({}, {TAG_RAY_NODE_KIND: "unmanaged"}, 1) + unmanaged_ip = self.provider.non_terminated_node_ips( + tag_filters={TAG_RAY_NODE_KIND: "unmanaged"}, )[0] + + runner = MockProcessRunner() + + lm = LoadMetrics() + lm.local_ip = head_ip + + autoscaler = StandardAutoscaler( + config_path, + lm, + max_launch_batch=5, + max_concurrent_launches=5, + max_failures=0, + process_runner=runner, + update_interval_s=0) + + autoscaler.update() + self.waitForNodes(2) + # This node has num_cpus=0 + lm.update(unmanaged_ip, {"CPU": 0}, {"CPU": 0}, {}) + autoscaler.update() + self.waitForNodes(2) + # 1 CPU task cannot be scheduled. + lm.update(unmanaged_ip, {"CPU": 0}, {"CPU": 0}, {"CPU": 1}) + autoscaler.update() + self.waitForNodes(3) + def testDelayedLaunch(self): config_path = self.write_config(SMALL_CLUSTER) self.provider = MockProvider()
[Autoscaler] Only recognizes workers it launches <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? *Ray version and other system information (Python version, TensorFlow version, OS):* 0.8.6+ ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): * Start a ray cluster with gpu workers (non-resource demand autoscaler), head should not have gpu's. * Manually add a node (not using autoscaler), with a custom resource {"custom": 1} ``` @ray.remote(num_gpus=1) def gpu_task(): pass @ray.remote(resources={"custom":1}) def launch_gpu(): # This blocks and never starts a gpu worker. ray.get(gpu_task.remote()) ray.get(launch_gpu.remote()) ``` If we cannot run your script, we cannot fix your issue. - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
The root cause of this is that `StandardAutoscaler::workers()` explicitely filters ``` def workers(self): return self.provider.non_terminated_nodes( tag_filters={TAG_RAY_NODE_KIND: NODE_KIND_WORKER}) ``` @wuisawesome Can you set the priority of this task? > The root cause of this is that `StandardAutoscaler::workers()` explicitely filters > > ``` > def workers(self): > return self.provider.non_terminated_nodes( > tag_filters={TAG_RAY_NODE_KIND: NODE_KIND_WORKER}) > ``` This is indeed the root cause. When this would be fixed, the autoscaler will however still try to update the manual node because the `ray-launch-config` and `ray-runtime-config` labels are not set on the manual node either. Additionally, the script you posted looks very similar to the thing I'm trying to do, but is still a little different. In my case the `gpu_task` task is not started from within an actor but simply from within python code that is running on the manually started worker. I don't know if that makes a difference or not...
2020-09-02T21:54:34
ray-project/ray
10,519
ray-project__ray-10519
[ "10378" ]
f38dba09b2cd07a27d1e9659176c8a60972bfd44
diff --git a/python/ray/util/dask/__init__.py b/python/ray/util/dask/__init__.py --- a/python/ray/util/dask/__init__.py +++ b/python/ray/util/dask/__init__.py @@ -1,3 +1,14 @@ from .scheduler import ray_dask_get, ray_dask_get_sync +from .callbacks import ( + RayDaskCallback, + local_ray_callbacks, + unpack_ray_callbacks, +) -__all__ = ["ray_dask_get", "ray_dask_get_sync"] +__all__ = [ + "ray_dask_get", + "ray_dask_get_sync", + "RayDaskCallback", + "local_ray_callbacks", + "unpack_ray_callbacks", +] diff --git a/python/ray/util/dask/callbacks.py b/python/ray/util/dask/callbacks.py new file mode 100644 --- /dev/null +++ b/python/ray/util/dask/callbacks.py @@ -0,0 +1,211 @@ +import contextlib +from collections import namedtuple + +from dask.callbacks import Callback + +# The names of the Ray-specific callbacks. These are the kwarg names that +# RayDaskCallback will accept on construction, and is considered the +# source-of-truth for what Ray-specific callbacks exist. +CBS = ( + "ray_presubmit", + "ray_postsubmit", + "ray_pretask", + "ray_posttask", + "ray_postsubmit_all", + "ray_finish", +) +# The Ray-specific callback method names for RayDaskCallback. +CB_FIELDS = tuple("_" + field for field in CBS) +# The Ray-specific callbacks that we do _not_ wish to drop from RayCallbacks +# if not given on a RayDaskCallback instance (will be filled with None +# instead). +CBS_DONT_DROP = {"ray_pretask", "ray_posttask"} + +# The Ray-specific callbacks for a single RayDaskCallback. +RayCallback = namedtuple("RayCallback", " ".join(CBS)) + +# The Ray-specific callbacks for one or more RayDaskCallbacks. +RayCallbacks = namedtuple("RayCallbacks", + " ".join([field + "_cbs" for field in CBS])) + + +class RayDaskCallback(Callback): + """ + Extends Dask's `Callback` class with Ray-specific hooks. When instantiating + or subclassing this class, both the normal Dask hooks (e.g. pretask, + posttask, etc.) and the Ray-specific hooks can be provided. + + See `dask.callbacks.Callback` for usage. + + Caveats: Any Dask-Ray scheduler must bring the Ray-specific callbacks into + context using the `local_ray_callbacks` context manager, since the built-in + `local_callbacks` context manager provided by Dask isn't aware of this + class. + """ + + # Set of active Ray-specific callbacks. + ray_active = set() + + def __init__(self, **kwargs): + """ + Ray-specific callbacks: + - def _ray_presubmit(task, key, deps): + Run before submitting a Ray task. If this callback returns a + non-`None` value, a Ray task will _not_ be created and this + value will be used as the would-be task's result value. + + Args: + task (tuple): A Dask task, where the first tuple item is + the task function, and the remaining tuple items are + the task arguments (either the actual argument values, + or Dask keys into the deps dictionary whose + corresponding values are the argument values). + key (str): The Dask graph key for the given task. + deps (dict): The dependencies of this task. + + Returns: + Either None, in which case a Ray task will be submitted, or + a non-None value, in which case a Ray task will not be + submitted and this return value will be used as the + would-be task result value. + + - def _ray_postsubmit(task, key, deps, object_ref): + Run after submitting a Ray task. + + Args: + task (tuple): A Dask task, where the first tuple item is + the task function, and the remaining tuple items are + the task arguments (either the actual argument values, + or Dask keys into the deps dictionary whose + corresponding values are the argument values). + key (str): The Dask graph key for the given task. + deps (dict): The dependencies of this task. + object_ref (ray.ObjectRef): The object reference for the + return value of the Ray task. + + - def _ray_pretask(key, object_refs): + Run before executing a Dask task within a Ray task. This + executes after the task has been submitted, within a Ray + worker. The return value of this task will be passed to the + _ray_posttask callback, if provided. + + Args: + key (str): The Dask graph key for the Dask task. + object_refs (List[ray.ObjectRef]): The object references + for the arguments of the Ray task. + + Returns: + A value that will be passed to the corresponding + _ray_posttask callback, if said callback is defined. + + - def _ray_posttask(key, result, pre_state): + Run after executing a Dask task within a Ray task. This + executes within a Ray worker. This callback receives the return + value of the _ray_pretask callback, if provided. + + Args: + key (str): The Dask graph key for the Dask task. + result (object): The task result value. + pre_state (object): The return value of the corresponding + _ray_pretask callback, if said callback is defined. + + - def _ray_postsubmit_all(object_refs, dsk): + Run after all Ray tasks have been submitted. + + Args: + object_refs (List[ray.ObjectRef]): The object references + for the output (leaf) Ray tasks of the task graph. + dsk (dict): The Dask graph. + + - def _ray_finish(result): + Run after all Ray tasks have finished executing and the final + result has been returned. + + Args: + result (object): The final result (output) of the Dask + computation, before any repackaging is done by + Dask collection-specific post-compute callbacks. + """ + for cb in CBS: + cb_func = kwargs.pop(cb, None) + if cb_func is not None: + setattr(self, "_" + cb, cb_func) + + super().__init__(**kwargs) + + @property + def _ray_callback(self): + return RayCallback( + *[getattr(self, field, None) for field in CB_FIELDS]) + + def __enter__(self): + self._ray_cm = add_ray_callbacks(self) + self._ray_cm.__enter__() + super().__enter__() + return self + + def __exit__(self, *args): + super().__exit__(*args) + self._ray_cm.__exit__(*args) + + def register(self): + type(self).ray_active.add(self._ray_callback) + super().register() + + def unregister(self): + type(self).ray_active.remove(self._ray_callback) + super().unregister() + + +class add_ray_callbacks: + def __init__(self, *callbacks): + self.callbacks = [normalize_ray_callback(c) for c in callbacks] + RayDaskCallback.ray_active.update(self.callbacks) + + def __enter__(self): + return self + + def __exit__(self, *args): + for c in self.callbacks: + RayDaskCallback.ray_active.discard(c) + + +def normalize_ray_callback(cb): + if isinstance(cb, RayDaskCallback): + return cb._ray_callback + elif isinstance(cb, RayCallback): + return cb + else: + raise TypeError( + "Callbacks must be either 'RayDaskCallback' or 'RayCallback' " + "namedtuple") + + +def unpack_ray_callbacks(cbs): + """Take an iterable of callbacks, return a list of each callback.""" + if cbs: + # Only drop callback methods that aren't in CBS_DONT_DROP. + return RayCallbacks(*( + [cb for cb in cbs_ if cb or CBS[idx] in CBS_DONT_DROP] or None + for idx, cbs_ in enumerate(zip(*cbs)))) + else: + return RayCallbacks(*([()] * len(CBS))) + + [email protected] +def local_ray_callbacks(callbacks=None): + """ + Allows Dask-Ray callbacks to work with nested schedulers. + + Callbacks will only be used by the first started scheduler they encounter. + This means that only the outermost scheduler will use global callbacks. + """ + global_callbacks = callbacks is None + if global_callbacks: + callbacks, RayDaskCallback.ray_active = (RayDaskCallback.ray_active, + set()) + try: + yield callbacks or () + finally: + if global_callbacks: + RayDaskCallback.ray_active = callbacks diff --git a/python/ray/util/dask/scheduler.py b/python/ray/util/dask/scheduler.py --- a/python/ray/util/dask/scheduler.py +++ b/python/ray/util/dask/scheduler.py @@ -10,6 +10,7 @@ from dask.system import CPU_COUNT from dask.threaded import pack_exception, _thread_get_id +from .callbacks import local_ray_callbacks, unpack_ray_callbacks from .common import unpack_object_refs main_thread = threading.current_thread() @@ -30,12 +31,14 @@ def ray_dask_get(dsk, keys, **kwargs): >>> dask.compute(obj, scheduler=ray_dask_get) - You can override the number of threads to use when submitting the - Ray tasks, or the threadpool used to submit Ray tasks: + You can override the currently active global Dask-Ray callbacks (e.g. + supplied via a context manager), the number of threads to use when + submitting the Ray tasks, or the threadpool used to submit Ray tasks: >>> dask.compute( obj, scheduler=ray_dask_get, + ray_callbacks=some_ray_dask_callbacks, num_workers=8, pool=some_cool_pool, ) @@ -44,6 +47,7 @@ def ray_dask_get(dsk, keys, **kwargs): dsk (Dict): Dask graph, represented as a task DAG dictionary. keys (List[str]): List of Dask graph keys whose values we wish to compute and return. + ray_callbacks (Optional[list[callable]]): Dask-Ray callbacks. num_workers (Optional[int]): The number of worker threads to use in the Ray task submission traversal of the Dask graph. pool (Optional[ThreadPool]): A multiprocessing threadpool to use to @@ -73,23 +77,48 @@ def ray_dask_get(dsk, keys, **kwargs): atexit.register(pool.close) pools[thread][num_workers] = pool - # NOTE: We hijack Dask's `get_async` function, injecting a different task - # executor. - object_refs = get_async( - _apply_async_wrapper(pool.apply_async, _rayify_task_wrapper), - len(pool._pool), - dsk, - keys, - get_id=_thread_get_id, - pack_exception=pack_exception, - **kwargs, - ) - # NOTE: We explicitly delete the Dask graph here so object references - # are garbage-collected before this function returns, i.e. before all Ray - # tasks are done. Otherwise, no intermediate objects will be cleaned up - # until all Ray tasks are done. - del dsk - result = ray_get_unpack(object_refs) + ray_callbacks = kwargs.pop("ray_callbacks", None) + + with local_ray_callbacks(ray_callbacks) as ray_callbacks: + # Unpack the Ray-specific callbacks. + ( + ray_presubmit_cbs, + ray_postsubmit_cbs, + ray_pretask_cbs, + ray_posttask_cbs, + ray_postsubmit_all_cbs, + ray_finish_cbs, + ) = unpack_ray_callbacks(ray_callbacks) + # NOTE: We hijack Dask's `get_async` function, injecting a different + # task executor. + object_refs = get_async( + _apply_async_wrapper( + pool.apply_async, + _rayify_task_wrapper, + ray_presubmit_cbs, + ray_postsubmit_cbs, + ray_pretask_cbs, + ray_posttask_cbs, + ), + len(pool._pool), + dsk, + keys, + get_id=_thread_get_id, + pack_exception=pack_exception, + **kwargs, + ) + if ray_postsubmit_all_cbs is not None: + for cb in ray_postsubmit_all_cbs: + cb(object_refs, dsk) + # NOTE: We explicitly delete the Dask graph here so object references + # are garbage-collected before this function returns, i.e. before all + # Ray tasks are done. Otherwise, no intermediate objects will be + # cleaned up until all Ray tasks are done. + del dsk + result = ray_get_unpack(object_refs) + if ray_finish_cbs is not None: + for cb in ray_finish_cbs: + cb(result) # cleanup pools associated with dead threads. with pools_lock: @@ -138,6 +167,10 @@ def _rayify_task_wrapper( loads, get_id, pack_exception, + ray_presubmit_cbs, + ray_postsubmit_cbs, + ray_pretask_cbs, + ray_posttask_cbs, ): """ The core Ray-Dask task execution wrapper, to be given to the thread pool's @@ -152,6 +185,10 @@ def _rayify_task_wrapper( loads (callable): A task_info deserializing function. get_id (callable): An ID generating function. pack_exception (callable): An exception serializing function. + ray_presubmit_cbs (callable): Pre-task submission callbacks. + ray_postsubmit_cbs (callable): Post-task submission callbacks. + ray_pretask_cbs (callable): Pre-task execution callbacks. + ray_posttask_cbs (callable): Post-task execution callbacks. Returns: A 3-tuple of the task's key, a literal or a Ray object reference for a @@ -159,7 +196,15 @@ def _rayify_task_wrapper( """ try: task, deps = loads(task_info) - result = _rayify_task(task, key, deps) + result = _rayify_task( + task, + key, + deps, + ray_presubmit_cbs, + ray_postsubmit_cbs, + ray_pretask_cbs, + ray_posttask_cbs, + ) id = get_id() result = dumps((result, id)) failed = False @@ -169,15 +214,27 @@ def _rayify_task_wrapper( return key, result, failed -def _rayify_task(task, key, deps): +def _rayify_task( + task, + key, + deps, + ray_presubmit_cbs, + ray_postsubmit_cbs, + ray_pretask_cbs, + ray_posttask_cbs, +): """ Rayifies the given task, submitting it as a Ray task to the Ray cluster. Args: - task: A Dask graph value, being either a literal, dependency key, Dask - task, or a list thereof. - key: The Dask graph key for the given task. - deps: The dependencies of this task. + task (tuple): A Dask graph value, being either a literal, dependency + key, Dask task, or a list thereof. + key (str): The Dask graph key for the given task. + deps (dict): The dependencies of this task. + ray_presubmit_cbs (callable): Pre-task submission callbacks. + ray_postsubmit_cbs (callable): Post-task submission callbacks. + ray_pretask_cbs (callable): Pre-task execution callbacks. + ray_posttask_cbs (callable): Post-task execution callbacks. Returns: A literal, a Ray object reference representing a submitted task, or a @@ -186,18 +243,46 @@ def _rayify_task(task, key, deps): if isinstance(task, list): # Recursively rayify this list. This will still bottom out at the first # actual task encountered, inlining any tasks in that task's arguments. - return [_rayify_task(t, deps) for t in task] + return [ + _rayify_task( + t, + key, + deps, + ray_presubmit_cbs, + ray_postsubmit_cbs, + ray_pretask_cbs, + ray_posttask_cbs, + ) for t in task + ] elif istask(task): # Unpacks and repacks Ray object references and submits the task to the # Ray cluster for execution. + if ray_presubmit_cbs is not None: + alternate_returns = [ + cb(task, key, deps) for cb in ray_presubmit_cbs + ] + for alternate_return in alternate_returns: + # We don't submit a Ray task if a presubmit callback returns + # a non-`None` value, instead we return said value. + # NOTE: This returns the first non-None presubmit callback + # return value. + if alternate_return is not None: + return alternate_return + func, args = task[0], task[1:] # If the function's arguments contain nested object references, we must # unpack said object references into a flat set of arguments so that # Ray properly tracks the object dependencies between Ray tasks. object_refs, repack = unpack_object_refs(args, deps) # Submit the task using a wrapper function. - return dask_task_wrapper.options(name=f"dask:{key!s}").remote( - func, repack, *object_refs) + object_ref = dask_task_wrapper.options(name=f"dask:{key!s}").remote( + func, repack, key, ray_pretask_cbs, ray_posttask_cbs, *object_refs) + + if ray_postsubmit_cbs is not None: + for cb in ray_postsubmit_cbs: + cb(task, key, deps, object_ref) + + return object_ref elif not ishashable(task): return task elif task in deps: @@ -207,7 +292,8 @@ def _rayify_task(task, key, deps): @ray.remote -def dask_task_wrapper(func, repack, *args): +def dask_task_wrapper(func, repack, key, ray_pretask_cbs, ray_posttask_cbs, + *args): """ A Ray remote function acting as a Dask task wrapper. This function will repackage the given flat `args` into its original data structures using @@ -219,6 +305,9 @@ def dask_task_wrapper(func, repack, *args): func (callable): The Dask task function to execute. repack (callable): A function that repackages the provided args into the original (possibly nested) Python objects. + key (str): The Dask key for this task. + ray_pretask_cbs (callable): Pre-task execution callbacks. + ray_posttask_cbs (callable): Post-task execution callback. *args (ObjectRef): Ray object references representing the Dask task's arguments. @@ -227,11 +316,21 @@ def dask_task_wrapper(func, repack, *args): dask_task_wrapper.remote() invocation will return a Ray object reference representing the Ray task's result. """ + if ray_pretask_cbs is not None: + pre_states = [ + cb(key, args) if cb is not None else None for cb in ray_pretask_cbs + ] repacked_args, repacked_deps = repack(args) # Recursively execute Dask-inlined tasks. actual_args = [_execute_task(a, repacked_deps) for a in repacked_args] # Execute the actual underlying Dask task. - return func(*actual_args) + result = func(*actual_args) + if ray_posttask_cbs is not None: + for cb, pre_state in zip(ray_posttask_cbs, pre_states): + if cb is not None: + cb(key, result, pre_state) + + return result def ray_get_unpack(object_refs): @@ -273,6 +372,15 @@ def ray_dask_get_sync(dsk, keys, **kwargs): >>> dask.compute(obj, scheduler=ray_dask_get_sync) + You can override the currently active global Dask-Ray callbacks (e.g. + supplied via a context manager): + + >>> dask.compute( + obj, + scheduler=ray_dask_get_sync, + ray_callbacks=some_ray_dask_callbacks, + ) + Args: dsk (Dict): Dask graph, represented as a task DAG dictionary. keys (List[str]): List of Dask graph keys whose values we wish to @@ -281,18 +389,46 @@ def ray_dask_get_sync(dsk, keys, **kwargs): Returns: Computed values corresponding to the provided keys. """ - # NOTE: We hijack Dask's `get_async` function, injecting a different task - # executor. - object_refs = get_async( - _apply_async_wrapper(apply_sync, _rayify_task_wrapper), - 1, - dsk, - keys, - **kwargs, - ) - # NOTE: We explicitly delete the Dask graph here so object references - # are garbage-collected before this function returns, i.e. before all Ray - # tasks are done. Otherwise, no intermediate objects will be cleaned up - # until all Ray tasks are done. - del dsk - return ray_get_unpack(object_refs) + + ray_callbacks = kwargs.pop("ray_callbacks", None) + + with local_ray_callbacks(ray_callbacks) as ray_callbacks: + # Unpack the Ray-specific callbacks. + ( + ray_presubmit_cbs, + ray_postsubmit_cbs, + ray_pretask_cbs, + ray_posttask_cbs, + ray_postsubmit_all_cbs, + ray_finish_cbs, + ) = unpack_ray_callbacks(ray_callbacks) + # NOTE: We hijack Dask's `get_async` function, injecting a different + # task executor. + object_refs = get_async( + _apply_async_wrapper( + apply_sync, + _rayify_task_wrapper, + ray_presubmit_cbs, + ray_postsubmit_cbs, + ray_pretask_cbs, + ray_posttask_cbs, + ), + 1, + dsk, + keys, + **kwargs, + ) + if ray_postsubmit_all_cbs is not None: + for cb in ray_postsubmit_all_cbs: + cb(object_refs, dsk) + # NOTE: We explicitly delete the Dask graph here so object references + # are garbage-collected before this function returns, i.e. before all + # Ray tasks are done. Otherwise, no intermediate objects will be + # cleaned up until all Ray tasks are done. + del dsk + result = ray_get_unpack(object_refs) + if ray_finish_cbs is not None: + for cb in ray_finish_cbs: + cb(result) + + return result
diff --git a/python/ray/tests/BUILD b/python/ray/tests/BUILD --- a/python/ray/tests/BUILD +++ b/python/ray/tests/BUILD @@ -79,6 +79,7 @@ py_test_module_list( "test_command_runner.py", "test_coordinator_server.py", "test_dask_scheduler.py", + "test_dask_callback.py", "test_debug_tools.py", "test_global_state.py", "test_job.py", diff --git a/python/ray/tests/test_dask_callback.py b/python/ray/tests/test_dask_callback.py new file mode 100644 --- /dev/null +++ b/python/ray/tests/test_dask_callback.py @@ -0,0 +1,234 @@ +import dask +import pytest + +import ray +from ray.util.dask import ray_dask_get, RayDaskCallback + + [email protected] +def add(x, y): + return x + y + + +def test_callback_active(): + """Test that callbacks are active within context""" + assert not RayDaskCallback.ray_active + + with RayDaskCallback(): + assert RayDaskCallback.ray_active + + assert not RayDaskCallback.ray_active + + +def test_presubmit_shortcircuit(ray_start_regular_shared): + """ + Test that presubmit return short-circuits task submission, and that task's + result is set to the presubmit return value. + """ + + class PresubmitShortcircuitCallback(RayDaskCallback): + def _ray_presubmit(self, task, key, deps): + return 0 + + def _ray_postsubmit(self, task, key, deps, object_ref): + pytest.fail("_ray_postsubmit shouldn't be called when " + "_ray_presubmit returns a value") + + with PresubmitShortcircuitCallback(): + z = add(2, 3) + result = z.compute(scheduler=ray_dask_get) + + assert result == 0 + + +def test_pretask_posttask_shared_state(ray_start_regular_shared): + """ + Test that pretask return value is passed to corresponding posttask + callback. + """ + + class PretaskPosttaskCallback(RayDaskCallback): + def _ray_pretask(self, key, object_refs): + return key + + def _ray_posttask(self, key, result, pre_state): + assert pre_state == key + + with PretaskPosttaskCallback(): + z = add(2, 3) + result = z.compute(scheduler=ray_dask_get) + + assert result == 5 + + +def test_postsubmit(ray_start_regular_shared): + """ + Test that postsubmit is called after each task. + """ + + class PostsubmitCallback(RayDaskCallback): + def __init__(self, postsubmit_actor): + self.postsubmit_actor = postsubmit_actor + + def _ray_postsubmit(self, task, key, deps, object_ref): + self.postsubmit_actor.postsubmit.remote(task, key, deps, + object_ref) + + @ray.remote + class PostsubmitActor: + def __init__(self): + self.postsubmit_counter = 0 + + def postsubmit(self, task, key, deps, object_ref): + self.postsubmit_counter += 1 + + def get_postsubmit_counter(self): + return self.postsubmit_counter + + postsubmit_actor = PostsubmitActor.remote() + with PostsubmitCallback(postsubmit_actor): + z = add(2, 3) + result = z.compute(scheduler=ray_dask_get) + + assert ray.get(postsubmit_actor.get_postsubmit_counter.remote()) == 1 + assert result == 5 + + +def test_postsubmit_all(ray_start_regular_shared): + """ + Test that postsubmit_all is called once. + """ + + class PostsubmitAllCallback(RayDaskCallback): + def __init__(self, postsubmit_all_actor): + self.postsubmit_all_actor = postsubmit_all_actor + + def _ray_postsubmit_all(self, object_refs, dsk): + self.postsubmit_all_actor.postsubmit_all.remote(object_refs, dsk) + + @ray.remote + class PostsubmitAllActor: + def __init__(self): + self.postsubmit_all_called = False + + def postsubmit_all(self, object_refs, dsk): + self.postsubmit_all_called = True + + def get_postsubmit_all_called(self): + return self.postsubmit_all_called + + postsubmit_all_actor = PostsubmitAllActor.remote() + with PostsubmitAllCallback(postsubmit_all_actor): + z = add(2, 3) + result = z.compute(scheduler=ray_dask_get) + + assert ray.get(postsubmit_all_actor.get_postsubmit_all_called.remote()) + assert result == 5 + + +def test_finish(ray_start_regular_shared): + """ + Test that finish callback is called once. + """ + + class FinishCallback(RayDaskCallback): + def __init__(self, finish_actor): + self.finish_actor = finish_actor + + def _ray_finish(self, result): + self.finish_actor.finish.remote(result) + + @ray.remote + class FinishActor: + def __init__(self): + self.finish_called = False + + def finish(self, result): + self.finish_called = True + + def get_finish_called(self): + return self.finish_called + + finish_actor = FinishActor.remote() + with FinishCallback(finish_actor): + z = add(2, 3) + result = z.compute(scheduler=ray_dask_get) + + assert ray.get(finish_actor.get_finish_called.remote()) + assert result == 5 + + +def test_multiple_callbacks(ray_start_regular_shared): + """ + Test that multiple callbacks are supported. + """ + + class PostsubmitCallback(RayDaskCallback): + def __init__(self, postsubmit_actor): + self.postsubmit_actor = postsubmit_actor + + def _ray_postsubmit(self, task, key, deps, object_ref): + self.postsubmit_actor.postsubmit.remote(task, key, deps, + object_ref) + + @ray.remote + class PostsubmitActor: + def __init__(self): + self.postsubmit_counter = 0 + + def postsubmit(self, task, key, deps, object_ref): + self.postsubmit_counter += 1 + + def get_postsubmit_counter(self): + return self.postsubmit_counter + + postsubmit_actor = PostsubmitActor.remote() + cb1 = PostsubmitCallback(postsubmit_actor) + cb2 = PostsubmitCallback(postsubmit_actor) + cb3 = PostsubmitCallback(postsubmit_actor) + with cb1, cb2, cb3: + z = add(2, 3) + result = z.compute(scheduler=ray_dask_get) + + assert ray.get(postsubmit_actor.get_postsubmit_counter.remote()) == 3 + assert result == 5 + + +def test_pretask_posttask_shared_state_multi(ray_start_regular_shared): + """ + Test that pretask return values are passed to the correct corresponding + posttask callbacks when multiple callbacks are given. + """ + + class PretaskPosttaskCallback(RayDaskCallback): + def __init__(self, suffix): + self.suffix = suffix + + def _ray_pretask(self, key, object_refs): + return key + self.suffix + + def _ray_posttask(self, key, result, pre_state): + assert pre_state == key + self.suffix + + class PretaskOnlyCallback(RayDaskCallback): + def _ray_pretask(self, key, object_refs): + return "baz" + + class PosttaskOnlyCallback(RayDaskCallback): + def _ray_posttask(self, key, result, pre_state): + assert pre_state is None + + cb1 = PretaskPosttaskCallback("foo") + cb2 = PretaskOnlyCallback() + cb3 = PosttaskOnlyCallback() + cb4 = PretaskPosttaskCallback("bar") + with cb1, cb2, cb3, cb4: + z = add(2, 3) + result = z.compute(scheduler=ray_dask_get) + + assert result == 5 + + +if __name__ == "__main__": + import sys + sys.exit(pytest.main(["-v", __file__]))
[dask-on-ray] Add support for scheduler plugin callbacks. <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### Feature Request Dask provides a [scheduler plugin callback abstraction](https://docs.dask.org/en/latest/diagnostics-local.html#custom-callbacks) that allows users to hook into the scheduler and each task's execution lifecycle, running user-defined callbacks at different stages. Having this plugin infrastructure would be very useful for implementing Dask-level progress reporting, diagnostics, caching, etc. ### API and Semantics The same as Dask's [existing scheduler plugin callback abstraction](https://docs.dask.org/en/latest/diagnostics-local.html#custom-callbacks): a class that can be instantiated with the callbacks given as constructor args, or can be subclassed, with the callbacks implemented as instance methods on the derived class. #### Desired Hooks 1. The 5 existing Dask callback hooks, used to hook into the Rayification lifecycle. 2. Ray presubmit - Run before submitting a Ray task. 3. Ray postsubmit - Run after submitting a Ray task. 4. Ray pretask - Run before executing a Dask task within a Ray task. 5. Ray posttask - Run after executing a Dask task within a Ray task. 6. Ray postsubmit all - Run after all Ray tasks for the Dask graph have been submitted. 7. Ray finish - Run after all Ray tasks have finished executing and a result has been returned. Each of these hooks will have the appropriate signature, e.g. Ray presubmit will be given the Dask task, the Dask key, and the task deps, while Ray pretask will be given the Dask key, the Dask task function, and the object reference arguments.
2020-09-03T01:51:10
ray-project/ray
10,531
ray-project__ray-10531
[ "9245" ]
2a7f56e42926d3b481eb27edf7bab0dbe5ab73b9
diff --git a/python/ray/tune/schedulers/hb_bohb.py b/python/ray/tune/schedulers/hb_bohb.py --- a/python/ray/tune/schedulers/hb_bohb.py +++ b/python/ray/tune/schedulers/hb_bohb.py @@ -93,7 +93,7 @@ def _unpause_trial(self, trial_runner, trial): trial_runner.trial_executor.unpause_trial(trial) trial_runner._search_alg.searcher.on_unpause(trial.trial_id) - def choose_trial_to_run(self, trial_runner): + def choose_trial_to_run(self, trial_runner, allow_recurse=True): """Fair scheduling within iteration by completion percentage. List of trials not used since all trials are tracked as state @@ -117,8 +117,17 @@ def choose_trial_to_run(self, trial_runner): for bracket in hyperband: if bracket and any(trial.status == Trial.PAUSED for trial in bracket.current_trials()): - # This will change the trial state and let the - # trial runner retry. + # This will change the trial state self._process_bracket(trial_runner, bracket) + + # If there are pending trials now, suggest one. + # This is because there might be both PENDING and + # PAUSED trials now, and PAUSED trials will raise + # an error before the trial runner tries again. + if allow_recurse and any( + trial.status == Trial.PENDING + for trial in bracket.current_trials()): + return self.choose_trial_to_run( + trial_runner, allow_recurse=False) # MAIN CHANGE HERE! return None diff --git a/python/ray/tune/schedulers/hyperband.py b/python/ray/tune/schedulers/hyperband.py --- a/python/ray/tune/schedulers/hyperband.py +++ b/python/ray/tune/schedulers/hyperband.py @@ -282,7 +282,8 @@ def debug_string(self): for i, band in enumerate(self._hyperbands): out += "\nRound #{}:".format(i) for bracket in band: - out += "\n {}".format(bracket) + if bracket: + out += "\n {}".format(bracket) return out def state(self):
diff --git a/python/ray/tune/tests/test_trial_scheduler.py b/python/ray/tune/tests/test_trial_scheduler.py --- a/python/ray/tune/tests/test_trial_scheduler.py +++ b/python/ray/tune/tests/test_trial_scheduler.py @@ -689,6 +689,30 @@ def result(score, ts): self.assertTrue("hyperband_info" in spy_result) self.assertEquals(spy_result["hyperband_info"]["budget"], 1) + def testPauseResumeChooseTrial(self): + def result(score, ts): + return {"episode_reward_mean": score, TRAINING_ITERATION: ts} + + sched = HyperBandForBOHB(max_t=10, reduction_factor=3, mode="min") + runner = _MockTrialRunner(sched) + runner._search_alg = MagicMock() + runner._search_alg.searcher = MagicMock() + trials = [Trial("__fake") for i in range(3)] + for t in trials: + runner.add_trial(t) + runner._launch_trial(t) + + all_results = [result(1, 5), result(2, 1), result(3, 5)] + for trial, trial_result in zip(trials, all_results): + decision = sched.on_trial_result(runner, trial, trial_result) + self.assertEqual(decision, TrialScheduler.PAUSE) + runner._pause_trial(trial) + + run_trial = sched.choose_trial_to_run(runner) + self.assertEqual(run_trial, trials[1]) + self.assertSequenceEqual([t.status for t in trials], + [Trial.PAUSED, Trial.PENDING, Trial.PAUSED]) + class _MockTrial(Trial): def __init__(self, i, config):
[tune] BOHB with DQN and MountainCar-v0 fails <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? I am using BOHB to optimize the hyperparameters for the DQN algorithm in order to solve the MountainCar-v0 problem. I alway run into the following issue (even if I use different values for num_samples): ```python == Status == Memory usage on this node: 7.0/15.6 GiB Using HyperBand: num_stopped=832 total_brackets=3 Round #0: None Bracket(Max Size (n)=2, Milestone (r)=1458, completed=100.0%): {RUNNING: 1, TERMINATED: 833} Bracket(Max Size (n)=324, Milestone (r)=8, completed=47.3%): {PAUSED: 166} Resources requested: 4/32 CPUs, 0/0 GPUs, 0.0/8.69 GiB heap, 0.0/2.98 GiB objects Result logdir: /home/dl-user/ray_results/MCv0_DQN_BOHB Number of trials: 1000 (166 PAUSED, 1 RUNNING, 833 TERMINATED) +-----------------------------+------------+----------------------+-------------------+-------------+--------------------+--------+------------------+--------+----------+ | Trial name | status | loc | batch_mode | lr | train_batch_size | iter | total time (s) | ts | reward | |-----------------------------+------------+----------------------+-------------------+-------------+--------------------+--------+------------------+--------+----------| | DQN_MountainCar-v0_0428be42 | PAUSED | | truncate_episodes | 1.99095e-05 | 408 | 2 | 25.6885 | 4032 | -200 | | DQN_MountainCar-v0_0428be45 | PAUSED | | truncate_episodes | 0.000382289 | 211 | 2 | 24.7536 | 5040 | -200 | | DQN_MountainCar-v0_0428be48 | PAUSED | | truncate_episodes | 0.000324929 | 233 | 2 | 25.5532 | 5040 | -200 | | DQN_MountainCar-v0_0747e5f2 | PAUSED | | truncate_episodes | 0.000114766 | 38 | 2 | 23.8492 | 7056 | -200 | | DQN_MountainCar-v0_0747e5f5 | PAUSED | | truncate_episodes | 9.1226e-05 | 200 | 2 | 24.2349 | 5040 | -200 | | DQN_MountainCar-v0_08218bf0 | PAUSED | | truncate_episodes | 0.000284028 | 69 | 2 | 25.3671 | 7056 | -200 | | DQN_MountainCar-v0_093c0b8c | PAUSED | | truncate_episodes | 0.00237606 | 114 | 2 | 23.3935 | 6048 | -200 | | DQN_MountainCar-v0_0a55eae6 | PAUSED | | truncate_episodes | 0.000417829 | 111 | 2 | 23.4849 | 6048 | -200 | | DQN_MountainCar-v0_0b307d56 | PAUSED | | truncate_episodes | 0.000196047 | 59 | 2 | 23.1338 | 7056 | -200 | | DQN_MountainCar-v0_0eedea91 | PAUSED | | truncate_episodes | 6.58278e-05 | 59 | 2 | 24.0254 | 7056 | -200 | | DQN_MountainCar-v0_1fcd888b | RUNNING | 172.16.160.219:47910 | truncate_episodes | 0.000237864 | 751 | 88 | 1638.34 | 199584 | -122.05 | | DQN_MountainCar-v0_0023f4f6 | TERMINATED | | truncate_episodes | 0.000255833 | 158 | 1 | 5.56779 | 1008 | -200 | | DQN_MountainCar-v0_0023f4f9 | TERMINATED | | complete_episodes | 0.000262904 | 156 | 1 | 5.43817 | 1200 | -200 | | DQN_MountainCar-v0_0023f4fc | TERMINATED | | complete_episodes | 0.0002605 | 260 | 1 | 5.33452 | 1200 | -200 | | DQN_MountainCar-v0_0108428e | TERMINATED | | truncate_episodes | 3.89327e-05 | 732 | 4 | 36.2218 | 5040 | -200 | | DQN_MountainCar-v0_01084291 | TERMINATED | | truncate_episodes | 2.39745e-05 | 714 | 4 | 36.2585 | 5040 | -200 | | DQN_MountainCar-v0_01084294 | TERMINATED | | truncate_episodes | 4.9252e-05 | 808 | 4 | 38.4182 | 5040 | -200 | | DQN_MountainCar-v0_01084297 | TERMINATED | | truncate_episodes | 7.42384e-05 | 804 | 4 | 38.0425 | 5040 | -200 | | DQN_MountainCar-v0_014223c0 | TERMINATED | | truncate_episodes | 0.0520328 | 71 | 1 | 6.21906 | 1008 | -200 | | DQN_MountainCar-v0_01939ac4 | TERMINATED | | complete_episodes | 8.34678e-05 | 124 | 1 | 5.37302 | 1200 | -200 | | DQN_MountainCar-v0_01a4cc45 | TERMINATED | | complete_episodes | 0.00973094 | 373 | 3 | 27.2147 | 24000 | -200 | +-----------------------------+------------+----------------------+-------------------+-------------+--------------------+--------+------------------+--------+----------+ ... 980 more trials not shown (156 PAUSED, 823 TERMINATED) Traceback (most recent call last): File "/home/dl-user/python-code/modularized_version_ray/ray_BOHB.py", line 123, in <module> verbose=1, File "/home/dl-user/.local/lib/python3.7/site-packages/ray/tune/tune.py", line 327, in run runner.step() File "/home/dl-user/.local/lib/python3.7/site-packages/ray/tune/trial_runner.py", line 342, in step self.trial_executor.on_no_available_trials(self) File "/home/dl-user/.local/lib/python3.7/site-packages/ray/tune/trial_executor.py", line 177, in on_no_available_trials raise TuneError("There are paused trials, but no more pending " ray.tune.error.TuneError: There are paused trials, but no more pending trials with sufficient resources. Process finished with exit code 1 ``` By the way, why is the first bracket ``none``? And also it seems that the whole process was pretty slow. Is there any systematic way to find out the bottlenecks? You guys are doing a great job with this framework! Thanks in advance for your help! *Ray version and other system information (Python version, TensorFlow version, OS):* OS: Ubuntu 16.04 Python Version: 3.7.8 Ray Version: 0.8.6 ### Reproduction (REQUIRED) ```python import ray from ray import tune from ray.tune.suggest.bohb import TuneBOHB from ray.tune.schedulers.hb_bohb import HyperBandForBOHB import ConfigSpace as CS ray.init() config_space = CS.ConfigurationSpace() config_space.add_hyperparameter(CS.UniformFloatHyperparameter("lr", lower=0.00001, upper=0.1, log=True)) config_space.add_hyperparameter(CS.UniformIntegerHyperparameter("train_batch_size", lower=32, upper=1024, log=False)) config_space.add_hyperparameter(CS.CategoricalHyperparameter("batch_mode", choices=['truncate_episodes', 'complete_episodes'])) bohb_search = TuneBOHB( space=config_space, bohb_config=None, max_concurrent=32, metric='episode_reward_mean', mode='max' ) bohb_hyperband = HyperBandForBOHB( time_attr='episodes_total', metric='episode_reward_mean', mode='max', max_t=2000, reduction_factor=3, ) config = { "env": "MountainCar-v0", "min_iter_time_s": 15, "num_gpus": 0, "num_workers": 1, "double_q": True, "n_step": 3, "target_network_update_freq": 1000, "buffer_size": 20000, # "prioritzed_replay": True, "learning_starts": 1000, "log_level": "ERROR" } analysis = tune.run( run_or_experiment="DQN", name='MCv0_DQN_BOHB', config=config, num_samples=1000, checkpoint_at_end=True, search_alg=bohb_search, scheduler=bohb_hyperband, verbose=1, ) ``` Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): If we cannot run your script, we cannot fix your issue. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
Hey - I have a very similar issue with BOHB, with very similar code to above. It seems to work when you use "training_iteration" but not when you use episodes or timesteps, with batch size as a hyperparameter. It seems to break when the number of update steps is not uniform across agents. Not sure if this helps someone resolve the issue. It would be great if it can be resolved!
2020-09-03T13:23:21
ray-project/ray
10,547
ray-project__ray-10547
[ "10503" ]
0c0b0d0a73fc7fed019b164336382ed6ee3c04ea
diff --git a/python/ray/experimental/dask/scheduler.py b/python/ray/experimental/dask/scheduler.py --- a/python/ray/experimental/dask/scheduler.py +++ b/python/ray/experimental/dask/scheduler.py @@ -196,7 +196,8 @@ def _rayify_task(task, key, deps): # Ray properly tracks the object dependencies between Ray tasks. object_refs, repack = unpack_object_refs(args, deps) # Submit the task using a wrapper function. - return dask_task_wrapper.remote(func, repack, *object_refs) + return dask_task_wrapper.options(name=f"dask:{key!s}").remote( + func, repack, *object_refs) elif not ishashable(task): return task elif task in deps:
[dask-on-ray] Set task name to Dask key at task submission time. <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### Describe your feature request To provide better dashboard, metrics, and debugging UX, we should set the task name to the Dask key. The Dask-Ray scheduler [uses a single remote function](https://github.com/ray-project/ray/blob/ef18893fb50c97c2c0d9d7b8b951455683f89635/python/ray/experimental/dask/scheduler.py#L208-L233) for executing all Dask tasks; Dask task functions are all dynamically composed (via partial application and the like), and using this single remote function allows us to cut down on function exporting overhead. However, this means that the default task name (derived from the function descriptor) is the same for every Dask task, which hurts dashboard and debugging UX considerably. By setting the task name to the Dask key at task submission time, we'll be able to attach meaningful semantics to each task that will be displayed in the dashboard and included in task spec log records. ### Blockers Blocked by https://github.com/ray-project/ray/issues/10371.
2020-09-03T19:29:46
ray-project/ray
10,563
ray-project__ray-10563
[ "10562" ]
f38dba09b2cd07a27d1e9659176c8a60972bfd44
diff --git a/python/ray/utils.py b/python/ray/utils.py --- a/python/ray/utils.py +++ b/python/ray/utils.py @@ -365,9 +365,10 @@ def resources_from_resource_arguments( elif default_num_gpus is not None: resources["GPU"] = default_num_gpus - memory = default_memory or runtime_memory - object_store_memory = (default_object_store_memory - or runtime_object_store_memory) + # Order of arguments matter for short circuiting. + memory = runtime_memory or default_memory + object_store_memory = (runtime_object_store_memory + or default_object_store_memory) if memory is not None: resources["memory"] = ray_constants.to_memory_units( memory, round_up=True) diff --git a/rllib/evaluation/worker_set.py b/rllib/evaluation/worker_set.py --- a/rllib/evaluation/worker_set.py +++ b/rllib/evaluation/worker_set.py @@ -98,9 +98,10 @@ def add_workers(self, num_workers: int) -> None: remote_args = { "num_cpus": self._remote_config["num_cpus_per_worker"], "num_gpus": self._remote_config["num_gpus_per_worker"], - "memory": self._remote_config["memory_per_worker"], - "object_store_memory": self._remote_config[ - "object_store_memory_per_worker"], + # memory=0 is an error, but memory=None means no limits. + "memory": self._remote_config["memory_per_worker"] or None, + "object_store_memory": self. + _remote_config["object_store_memory_per_worker"] or None, "resources": self._remote_config["custom_resources_per_worker"], } cls = RolloutWorker.as_remote(**remote_args).remote
diff --git a/python/ray/tests/test_basic.py b/python/ray/tests/test_basic.py --- a/python/ray/tests/test_basic.py +++ b/python/ray/tests/test_basic.py @@ -341,6 +341,30 @@ def test_function_descriptor(): assert d.get(python_descriptor2) == 123 +def test_ray_options(shutdown_only): + @ray.remote( + num_cpus=2, num_gpus=3, memory=150 * 2**20, resources={"custom1": 1}) + def foo(): + return ray.available_resources() + + ray.init(num_cpus=10, num_gpus=10, resources={"custom1": 2}) + + without_options = ray.get(foo.remote()) + with_options = ray.get( + foo.options( + num_cpus=3, + num_gpus=4, + memory=50 * 2**20, + resources={ + "custom1": 0.5 + }).remote()) + + to_check = ["CPU", "GPU", "memory", "custom1"] + for key in to_check: + assert without_options[key] != with_options[key] + assert without_options != with_options + + def test_nested_functions(ray_start_shared_local_modes): # Make sure that remote functions can use other values that are defined # after the remote function but before the first function invocation.
[Core] `ray.options(memory=)` not overriding default. <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? *Ray version and other system information (Python version, TensorFlow version, OS):* ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): ``` >>> @ray.remote(memory=2**30) ... def foo(): ... print(f"cluster memory: {ray.cluster_resources()['memory']}") ... print(f"available memory: {ray.available_resources()['memory']}") ... >>> ray.get(foo.remote()) (pid=41181) cluster memory: 103.0 (pid=41181) available memory: 82.0 >>> ray.get(foo.options(memory=2**31).remote()) (pid=41181) cluster memory: 103.0 (pid=41181) available memory: 82.0 ``` If we cannot run your script, we cannot fix your issue. - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-09-04T04:36:30
ray-project/ray
10,593
ray-project__ray-10593
[ "10492" ]
c4c0857107ee4c4b982667339e0e4c5d33d97e96
diff --git a/python/ray/job_config.py b/python/ray/job_config.py --- a/python/ray/job_config.py +++ b/python/ray/job_config.py @@ -15,7 +15,7 @@ class JobConfig: def __init__( self, worker_env=None, - num_java_workers_per_process=10, + num_java_workers_per_process=1, jvm_options=None, ): if worker_env is None:
diff --git a/java/test.sh b/java/test.sh --- a/java/test.sh +++ b/java/test.sh @@ -33,15 +33,18 @@ bazel build //java:gen_maven_deps echo "Build test jar." bazel build //java:all_tests_deploy.jar +# Enable multi-worker feature in Java test +TEST_ARGS=(-Dray.raylet.config.num_workers_per_process_java=10 -Dray.job.num-java-workers-per-process=10) + echo "Running tests under cluster mode." # TODO(hchen): Ideally, we should use the following bazel command to run Java tests. However, if there're skipped tests, # TestNG will exit with code 2. And bazel treats it as test failure. # bazel test //java:all_tests --action_env=ENABLE_MULTI_LANGUAGE_TESTS=1 --config=ci || cluster_exit_code=$? -ENABLE_MULTI_LANGUAGE_TESTS=1 run_testng java -cp "$ROOT_DIR"/../bazel-bin/java/all_tests_deploy.jar org.testng.TestNG -d /tmp/ray_java_test_output "$ROOT_DIR"/testng.xml +ENABLE_MULTI_LANGUAGE_TESTS=1 run_testng java -cp "$ROOT_DIR"/../bazel-bin/java/all_tests_deploy.jar "${TEST_ARGS[@]}" org.testng.TestNG -d /tmp/ray_java_test_output "$ROOT_DIR"/testng.xml echo "Running tests under single-process mode." # bazel test //java:all_tests --jvmopt="-Dray.run-mode=SINGLE_PROCESS" --config=ci || single_exit_code=$? -run_testng java -Dray.run-mode="SINGLE_PROCESS" -cp "$ROOT_DIR"/../bazel-bin/java/all_tests_deploy.jar org.testng.TestNG -d /tmp/ray_java_test_output "$ROOT_DIR"/testng.xml +run_testng java -Dray.run-mode="SINGLE_PROCESS" -cp "$ROOT_DIR"/../bazel-bin/java/all_tests_deploy.jar "${TEST_ARGS[@]}" org.testng.TestNG -d /tmp/ray_java_test_output "$ROOT_DIR"/testng.xml popd diff --git a/java/test/src/main/java/io/ray/test/ClassLoaderTest.java b/java/test/src/main/java/io/ray/test/ClassLoaderTest.java --- a/java/test/src/main/java/io/ray/test/ClassLoaderTest.java +++ b/java/test/src/main/java/io/ray/test/ClassLoaderTest.java @@ -40,7 +40,7 @@ public void tearDown() { @Test(groups = {"cluster"}) public void testClassLoaderInMultiThreading() throws Exception { - Assert.assertTrue(TestUtils.getRuntime().getRayConfig().numWorkersPerProcess > 1); + Assert.assertTrue(TestUtils.getNumWorkersPerProcess() > 1); final String jobResourcePath = resourcePath + "/" + Ray.getRuntimeContext().getCurrentJobId(); File jobResourceDir = new File(jobResourcePath); diff --git a/java/test/src/main/java/io/ray/test/ExitActorTest.java b/java/test/src/main/java/io/ray/test/ExitActorTest.java --- a/java/test/src/main/java/io/ray/test/ExitActorTest.java +++ b/java/test/src/main/java/io/ray/test/ExitActorTest.java @@ -73,7 +73,7 @@ public void testExitActor() throws IOException, InterruptedException { } public void testExitActorInMultiWorker() { - Assert.assertTrue(TestUtils.getRuntime().getRayConfig().numWorkersPerProcess > 1); + Assert.assertTrue(TestUtils.getNumWorkersPerProcess() > 1); ActorHandle<ExitingActor> actor1 = Ray.actor(ExitingActor::new) .setMaxRestarts(10000).remote(); int pid = actor1.task(ExitingActor::getPid).remote().get(); diff --git a/java/test/src/main/java/io/ray/test/FailureTest.java b/java/test/src/main/java/io/ray/test/FailureTest.java --- a/java/test/src/main/java/io/ray/test/FailureTest.java +++ b/java/test/src/main/java/io/ray/test/FailureTest.java @@ -21,17 +21,20 @@ public class FailureTest extends BaseTest { private static final String EXCEPTION_MESSAGE = "Oops"; + private String oldNumWorkersPerProcess; + @BeforeClass public void setUp() { // This is needed by `testGetThrowsQuicklyWhenFoundException`. // Set one worker per process. Otherwise, if `badFunc2` and `slowFunc` run in the same // process, `sleep` will delay `System.exit`. + oldNumWorkersPerProcess = System.getProperty("ray.raylet.config.num_workers_per_process_java"); System.setProperty("ray.raylet.config.num_workers_per_process_java", "1"); } @AfterClass public void tearDown() { - System.clearProperty("ray.raylet.config.num_workers_per_process_java"); + System.setProperty("ray.raylet.config.num_workers_per_process_java", oldNumWorkersPerProcess); } public static int badFunc() { diff --git a/java/test/src/main/java/io/ray/test/JobConfigTest.java b/java/test/src/main/java/io/ray/test/JobConfigTest.java --- a/java/test/src/main/java/io/ray/test/JobConfigTest.java +++ b/java/test/src/main/java/io/ray/test/JobConfigTest.java @@ -1,7 +1,6 @@ package io.ray.test; import io.ray.api.ActorHandle; -import io.ray.api.ObjectRef; import io.ray.api.Ray; import org.testng.Assert; import org.testng.annotations.AfterClass; @@ -11,9 +10,12 @@ @Test(groups = {"cluster"}) public class JobConfigTest extends BaseTest { + private String oldNumWorkersPerProcess; + @BeforeClass public void setupJobConfig() { System.setProperty("ray.raylet.config.enable_multi_tenancy", "true"); + oldNumWorkersPerProcess = System.getProperty("ray.job.num-java-workers-per-process"); System.setProperty("ray.job.num-java-workers-per-process", "3"); System.setProperty("ray.job.jvm-options.0", "-DX=999"); System.setProperty("ray.job.jvm-options.1", "-DY=998"); @@ -24,7 +26,7 @@ public void setupJobConfig() { @AfterClass public void tearDownJobConfig() { System.clearProperty("ray.raylet.config.enable_multi_tenancy"); - System.clearProperty("ray.job.num-java-workers-per-process"); + System.setProperty("ray.job.num-java-workers-per-process", oldNumWorkersPerProcess); System.clearProperty("ray.job.jvm-options.0"); System.clearProperty("ray.job.jvm-options.1"); System.clearProperty("ray.job.worker-env.foo1"); @@ -39,16 +41,8 @@ public static String getEnvVariable(String key) { return System.getenv(key); } - public static Integer getWorkersNum() { - return TestUtils.getRuntime().getRayConfig().numWorkersPerProcess; - } - public static class MyActor { - public Integer getWorkersNum() { - return TestUtils.getRuntime().getRayConfig().numWorkersPerProcess; - } - public String getJvmOptions(String propertyName) { return System.getProperty(propertyName); } @@ -68,9 +62,8 @@ public void testWorkerEnvVariable() { Assert.assertEquals("bar2", Ray.task(JobConfigTest::getEnvVariable, "foo2").remote().get()); } - public void testNumJavaWorkerPerProcess() { - ObjectRef<Integer> obj = Ray.task(JobConfigTest::getWorkersNum).remote(); - Assert.assertEquals(3, (int) obj.get()); + public void testNumJavaWorkersPerProcess() { + Assert.assertEquals(TestUtils.getNumWorkersPerProcess(), 3); } @@ -84,9 +77,5 @@ public void testInActor() { // test worker env variables Assert.assertEquals("bar1", Ray.task(MyActor::getEnvVariable, "foo1").remote().get()); Assert.assertEquals("bar2", Ray.task(MyActor::getEnvVariable, "foo2").remote().get()); - - // test workers number. - ObjectRef<Integer> obj2 = actor.task(MyActor::getWorkersNum).remote(); - Assert.assertEquals(3, (int) obj2.get()); } } diff --git a/java/test/src/main/java/io/ray/test/KillActorTest.java b/java/test/src/main/java/io/ray/test/KillActorTest.java --- a/java/test/src/main/java/io/ray/test/KillActorTest.java +++ b/java/test/src/main/java/io/ray/test/KillActorTest.java @@ -14,14 +14,17 @@ @Test(groups = {"cluster"}) public class KillActorTest extends BaseTest { + private String oldNumWorkersPerProcess; + @BeforeClass public void setUp() { + oldNumWorkersPerProcess = System.getProperty("ray.raylet.config.num_workers_per_process_java"); System.setProperty("ray.raylet.config.num_workers_per_process_java", "1"); } @AfterClass public void tearDown() { - System.clearProperty("ray.raylet.config.num_workers_per_process_java"); + System.setProperty("ray.raylet.config.num_workers_per_process_java", oldNumWorkersPerProcess); } public static class HangActor { diff --git a/java/test/src/main/java/io/ray/test/MultiThreadingTest.java b/java/test/src/main/java/io/ray/test/MultiThreadingTest.java --- a/java/test/src/main/java/io/ray/test/MultiThreadingTest.java +++ b/java/test/src/main/java/io/ray/test/MultiThreadingTest.java @@ -19,7 +19,7 @@ import org.testng.Assert; import org.testng.annotations.Test; -@Test +@Test(groups = {"cluster"}) public class MultiThreadingTest extends BaseTest { private static final Logger LOGGER = LoggerFactory.getLogger(MultiThreadingTest.class); @@ -221,11 +221,6 @@ static boolean testMissingWrapRunnable() throws InterruptedException { return true; } - @Test - public void testMissingWrapRunnableInDriver() throws InterruptedException { - testMissingWrapRunnable(); - } - @Test public void testMissingWrapRunnableInWorker() { Ray.task(MultiThreadingTest::testMissingWrapRunnable).remote().get(); diff --git a/java/test/src/main/java/io/ray/test/TestUtils.java b/java/test/src/main/java/io/ray/test/TestUtils.java --- a/java/test/src/main/java/io/ray/test/TestUtils.java +++ b/java/test/src/main/java/io/ray/test/TestUtils.java @@ -3,6 +3,7 @@ import com.google.common.base.Preconditions; import io.ray.api.ObjectRef; import io.ray.api.Ray; +import io.ray.runtime.AbstractRayRuntime; import io.ray.runtime.RayRuntimeInternal; import io.ray.runtime.RayRuntimeProxy; import io.ray.runtime.config.RunMode; @@ -83,8 +84,19 @@ public static RayRuntimeInternal getRuntime() { } public static RayRuntimeInternal getUnderlyingRuntime() { + if (Ray.internal() instanceof AbstractRayRuntime) { + return (RayRuntimeInternal) Ray.internal(); + } RayRuntimeProxy proxy = (RayRuntimeProxy) (java.lang.reflect.Proxy .getInvocationHandler(Ray.internal())); return proxy.getRuntimeObject(); } + + private static int getNumWorkersPerProcessRemoteFunction() { + return TestUtils.getRuntime().getRayConfig().numWorkersPerProcess; + } + + public static int getNumWorkersPerProcess() { + return Ray.task(TestUtils::getNumWorkersPerProcessRemoteFunction).remote().get(); + } }
Make the multi-worker feature for Java worker experimental Right now, the multi-worker feature for Java worker is enabled by default, but the `ActorHandle::kill()` API doesn't work well if multi-worker is enabled because it will kill the whole process instead of one worker in the process. To avoid complaints from Java users, we should disable the multi-worker feature by default, but we still enable it in unit test.
#7291
2020-09-05T00:36:46
ray-project/ray
10,672
ray-project__ray-10672
[ "10668" ]
d7c7aba99cc2e6383af605aaa4671f4266fad979
diff --git a/python/ray/scripts/scripts.py b/python/ray/scripts/scripts.py --- a/python/ray/scripts/scripts.py +++ b/python/ray/scripts/scripts.py @@ -622,7 +622,7 @@ def start(node_ip_address, redis_address, address, redis_port, port, " ray stop".format( redis_address, " --redis-password='" + redis_password + "'" if redis_password else "", - ", redis_password='" + redis_password + "'" + ", _redis_password='" + redis_password + "'" if redis_password else "")) else: # Start Ray on a non-head node. @@ -1472,7 +1472,7 @@ def memory(address, redis_password): if not address: address = services.find_redis_address_or_die() logger.info(f"Connecting to Ray instance at {address}.") - ray.init(address=address, redis_password=redis_password) + ray.init(address=address, _redis_password=redis_password) print(ray.internal.internal_api.memory_summary())
A lot of Ray commands fail in nightly build <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? This is because they keep passing the deleted `redis_password` keyword. For example, ``` $ ray memory 2020-09-09 05:24:50,248 INFO scripts.py:1474 -- Connecting to Ray instance at 172.31.56.46:6379. Traceback (most recent call last): File "/home/ubuntu/anaconda3/bin/ray", line 8, in <module> sys.exit(main()) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/ray/scripts/scripts.py", line 1602, in main return cli() File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/click/core.py", line 1259, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/home/ubuntu/anaconda3/lib/python3.6/site-packages/ray/scripts/scripts.py", line 1475, in memory ray.init(address=address, redis_password=redis_password) TypeError: init() got an unexpected keyword argument 'redis_password' ``` Also I feel confused about many prompts and docs still mentioning `redis_password` while it no longer exists in `ray.init` ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): If we cannot run your script, we cannot fix your issue. - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-09-09T06:13:47
ray-project/ray
10,680
ray-project__ray-10680
[ "10426" ]
799318d7d785881b7dc89aad817284683a4ea4f5
diff --git a/python/ray/tune/integration/wandb.py b/python/ray/tune/integration/wandb.py --- a/python/ray/tune/integration/wandb.py +++ b/python/ray/tune/integration/wandb.py @@ -9,6 +9,8 @@ from ray.tune.logger import Logger from ray.tune.utils import flatten_dict +import yaml + try: import wandb except ImportError: @@ -29,6 +31,12 @@ def _clean_log(obj): # Else try: pickle.dumps(obj) + yaml.dump( + obj, + Dumper=yaml.SafeDumper, + default_flow_style=False, + allow_unicode=True, + encoding="utf-8") return obj except Exception: # give up, similar to _SafeFallBackEncoder
[rllib] Weights & Biases logger cannot handle objects in configuration <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? The Weights & Biases logger cannot handle object references in RLlib configurations, for example in the callback API. ``` Process _WandbLoggingProcess-1: Traceback (most recent call last): File "/usr/lib/python3.7/multiprocessing/process.py", line 297, in _bootstrap self.run() File "[...]/ray/tune/integration/wandb.py", line 127, in run wandb.init(*self.args, **self.kwargs) File "[...]/wandb/__init__.py", line 1303, in init as_defaults=not allow_val_change) File "[...]/wandb/wandb_config.py", line 333, in _update self.persist() File "[...]/wandb/wandb_config.py", line 238, in persist conf_file.write(str(self)) File "[...]/wandb/wandb_config.py", line 374, in __str__ allow_unicode=True, encoding='utf-8') File "[...]/yaml/__init__.py", line 290, in dump return dump_all([data], stream, Dumper=Dumper, **kwds) File "[...]/yaml/__init__.py", line 278, in dump_all dumper.represent(data) File "[...]/yaml/representer.py", line 27, in represent node = self.represent_data(data) File "[...]/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "[...]/yaml/representer.py", line 207, in represent_dict return self.represent_mapping('tag:yaml.org,2002:map', data) File "[...]/yaml/representer.py", line 118, in represent_mapping node_value = self.represent_data(item_value) File "[...]/yaml/representer.py", line 48, in represent_data node = self.yaml_representers[data_types[0]](self, data) File "[...]/yaml/representer.py", line 207, in represent_dict return self.represent_mapping('tag:yaml.org,2002:map', data) File "[...]/yaml/representer.py", line 118, in represent_mapping node_value = self.represent_data(item_value) File "[...]/yaml/representer.py", line 58, in represent_data node = self.yaml_representers[None](self, data) File "[...]/yaml/representer.py", line 231, in represent_undefined raise RepresenterError("cannot represent an object", data) yaml.representer.RepresenterError: ('cannot represent an object', <class '__main__.MyCallbacks'>) ``` *Ray version and other system information (Python version, TensorFlow version, OS):* - Ray 0.8.7 - Ubuntu 18.04 - Python 3.7 ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): ``` from ray import tune from ray.rllib.agents.ppo import PPOTrainer from ray.rllib.agents.callbacks import DefaultCallbacks from ray.tune.integration.wandb import WandbLogger class MyCallbacks(DefaultCallbacks): def on_episode_end(self, worker, base_env, policies, episode, **kwargs): print("Episode ended") tune.run( PPOTrainer, checkpoint_freq=1, config={ "framework": "torch", "num_workers": 8, "num_gpus": 1, "env": "CartPole-v0", "callbacks": MyCallbacks, "logger_config": { "wandb": { "project": "test", "api_key_file": "./wandb_api_key_file", } } }, stop={ "training_iteration":10 }, loggers=[WandbLogger] ) ``` If we cannot run your script, we cannot fix your issue. - [X] I have verified my script runs in a clean environment and reproduces the issue. - [X] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
I noticed that the callback is already [explicitly filtered out in the result handling](https://github.com/ray-project/ray/blob/0aec4cbccb5862a7808f98b90696236da9e2a03e/python/ray/tune/integration/wandb.py#L159), but of course that doesn't help during initialization. Is there any reason why objects shouldn't be generally filtered out? Probably only numbers and strings make sense to log anyway? cc @krfricke Thanks. For the moment I submitted a [tiny PR](https://github.com/ray-project/ray/pull/10441) to fix the callback API specifically. I agree it makes sense to filter out invalid values, I'll get back to this soon. @krfricke I am having a similar problem trying to use APPO (PPO works fine). See https://pastebin.com/SdpNp0Pq Hey @janblumenkamp and @rhamnett this should be fixed via #10654. If this issue prevails, feel free to re-open! @krfricke @richardliaw Afraid that I am still seeing the same error when running an APPO tune job Hi @rhamnett, did you install the latest nightly ray release? Do you have a script to reproduce the error? @krfricke yes I am familiar with re-installing the nightly and can confirm that I have done that correctly. My script is lengthy and contains private intellectual property, so I am wondering if you can just simply adapt a PPO test you might have to APPO (should be trivial) and then let me know if you can re-produce? If after trying that, you can't reproduce, I will seek to dedicate some time to give you an example script.
2020-09-09T17:18:24
ray-project/ray
10,703
ray-project__ray-10703
[ "10692" ]
e72838c03d87c8eb10f7cb8df8804496b14121ab
diff --git a/python/ray/autoscaler/autoscaler.py b/python/ray/autoscaler/autoscaler.py --- a/python/ray/autoscaler/autoscaler.py +++ b/python/ray/autoscaler/autoscaler.py @@ -64,6 +64,9 @@ def __init__(self, process_runner=subprocess, update_interval_s=AUTOSCALER_UPDATE_INTERVAL_S): self.config_path = config_path + # Keep this before self.reset (self.provider needs to be created + # exactly once). + self.provider = None self.reset(errors_fatal=True) self.load_metrics = load_metrics @@ -250,6 +253,7 @@ def _update(self): self.should_update(node_id) for node_id in nodes): if node_id is not None: resources = self._node_resources(node_id) + logger.debug(f"{node_id}: Starting new thread runner.") T.append( threading.Thread( target=self.spawn_updater, @@ -295,9 +299,9 @@ def reset(self, errors_fatal=False): self.config = new_config self.runtime_hash = new_runtime_hash self.file_mounts_contents_hash = new_file_mounts_contents_hash - - self.provider = get_node_provider(self.config["provider"], - self.config["cluster_name"]) + if not self.provider: + self.provider = get_node_provider(self.config["provider"], + self.config["cluster_name"]) # Check whether we can enable the resource demand scheduler. if "available_node_types" in self.config: self.available_node_types = self.config["available_node_types"] @@ -462,6 +466,8 @@ def should_update(self, node_id): def spawn_updater(self, node_id, init_commands, ray_start_commands, node_resources, docker_config): + logger.info(f"Creating new (spawn_updater) updater thread for node" + f" {node_id}.") updater = NodeUpdaterThread( node_id=node_id, provider_config=self.config["provider"], @@ -492,6 +498,8 @@ def can_update(self, node_id): return False if self.num_failed_updates.get(node_id, 0) > 0: # TODO(ekl) retry? return False + logger.debug(f"{node_id} is not being updated and " + "passes config check (can_update=True).") return True def launch_new_node(self, count: int, node_type: Optional[str]) -> None: diff --git a/python/ray/autoscaler/commands.py b/python/ray/autoscaler/commands.py --- a/python/ray/autoscaler/commands.py +++ b/python/ray/autoscaler/commands.py @@ -455,7 +455,7 @@ def kill_node(config_file, yes, hard, override_cluster_name): def monitor_cluster(cluster_config_file, num_lines, override_cluster_name): """Tails the autoscaler logs of a Ray cluster.""" - cmd = "tail -n {} -f /tmp/ray/session_*/logs/monitor*".format(num_lines) + cmd = f"tail -n {num_lines} -f /tmp/ray/session_latest/logs/monitor*" exec_cluster( cluster_config_file, cmd=cmd, @@ -717,7 +717,7 @@ def get_or_create_head_node(config, logger, "get_or_create_head_node: " "Head node up-to-date, IP address is: {}", head_node_ip) - monitor_str = "tail -n 100 -f /tmp/ray/session_*/logs/monitor*" + monitor_str = "tail -n 100 -f /tmp/ray/session_latest/logs/monitor*" if override_cluster_name: modifiers = " --cluster-name={}".format( quote(override_cluster_name))
[autoscaler] monitor.py has a memory/thread leak ### What is the problem? With this cluster yaml, you can watch monitor.py via `htop` quickly eat up all of the available memory. It also seems to be creating hundreds of NodeUpdater threads. *Ray version and other system information (Python version, TensorFlow version, OS):* `master` latest wheels ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): ``` cluster_name: memoryleak min_workers: 2 max_workers: 2 initial_workers: 2 autoscaling_mode: default target_utilization_fraction: 0.8 idle_timeout_minutes: 60 provider: type: aws region: us-east-1 cache_stopped_nodes: false auth: ssh_user: ubuntu head_node: BlockDeviceMappings: - DeviceName: /dev/sda1 Ebs: VolumeSize: 300 DeleteOnTermination: true VolumeType: gp2 ImageId: latest_dlami InstanceType: t3.xlarge worker_nodes: BlockDeviceMappings: - DeviceName: /dev/sda1 Ebs: VolumeSize: 300 DeleteOnTermination: true VolumeType: gp2 ImageId: latest_dlami InstanceType: t3.xlarge file_mounts: {} cluster_synced_files: - ~/my-cluster-synced-file file_mounts_sync_continuously: true initialization_commands: [] setup_commands: - pip install --user -U https://s3-us-west-2.amazonaws.com/ray-wheels/latest/ray-0.9.0.dev0-cp37-cp37m-manylinux1_x86_64.whl head_setup_commands: - tmux new-session -d "python3 -m http.server" - pip install --user -U https://s3-us-west-2.amazonaws.com/ray-wheels/latest/ray-0.9.0.dev0-cp37-cp37m-manylinux1_x86_64.whl - sleep 5 - curl localhost:8000 -o ~/my-cluster-synced-file worker_setup_commands: [] ``` cc @ericl @wuisawesome
@wuisawesome could this be related to your new reloading code that creates a new node provider on config update? Just did some digging - we cannot create providers constantly, as is currently done in `autoscaler.reset()`. ``` ==> /tmp/ray/session_2020-09-10_02-30-21_292916_30660/logs/monitor.err <== File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/monitor.py", line 368, in <module> monitor.run() File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/monitor.py", line 313, in run self._run() File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/monitor.py", line 268, in _run self.autoscaler.update() File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 120, in update self.reset(errors_fatal=False) File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 302, in reset self.config["cluster_name"]) File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/autoscaler/node_provider.py", line 148, in get_node_provider return provider_cls(provider_config, cluster_name) File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/autoscaler/aws/node_provider.py", line 55, in __init__ traceback.print_stack() ==> /tmp/ray/session_2020-09-10_02-30-21_292916_30660/logs/monitor.out <== Creating a new NODE PROVIDER! ==> /tmp/ray/session_2020-09-10_02-30-21_292916_30660/logs/monitor.err <== File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/monitor.py", line 368, in <module> monitor.run() File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/monitor.py", line 313, in run self._run() File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/monitor.py", line 268, in _run self.autoscaler.update() File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 120, in update self.reset(errors_fatal=False) File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 302, in reset self.config["cluster_name"]) File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/autoscaler/node_provider.py", line 148, in get_node_provider return provider_cls(provider_config, cluster_name) File "/home/ubuntu/.local/lib/python3.7/site-packages/ray/autoscaler/aws/node_provider.py", line 55, in __init__ traceback.print_stack() ``` Seems like this is a necessary fix (or at least you need to clean up the existing provider): ``` @@ -295,9 +297,9 @@ class StandardAutoscaler: self.config = new_config self.runtime_hash = new_runtime_hash self.file_mounts_contents_hash = new_file_mounts_contents_hash - - self.provider = get_node_provider(self.config["provider"], - self.config["cluster_name"]) + if not self.provider: + self.provider = get_node_provider(self.config["provider"], + self.config["cluster_name"]) # Check whether we can enable the resource demand scheduler. if "available_node_types" in self.config: self.available_node_types = self.config["available_node_types"]
2020-09-10T05:08:52
ray-project/ray
10,705
ray-project__ray-10705
[ "10702" ]
eb025ea8cb27583e8ef6287f5654f23d1ab270ef
diff --git a/python/ray/autoscaler/updater.py b/python/ray/autoscaler/updater.py --- a/python/ray/autoscaler/updater.py +++ b/python/ray/autoscaler/updater.py @@ -78,17 +78,28 @@ def __init__(self, self.process_runner = process_runner self.node_id = node_id self.provider = provider + # Some node providers don't specify empty structures as + # defaults. Better to be defensive. + file_mounts = file_mounts or {} self.file_mounts = { remote: os.path.expanduser(local) for remote, local in file_mounts.items() } + self.initialization_commands = initialization_commands self.setup_commands = setup_commands self.ray_start_commands = ray_start_commands self.node_resources = node_resources self.runtime_hash = runtime_hash self.file_mounts_contents_hash = file_mounts_contents_hash - self.cluster_synced_files = cluster_synced_files + # TODO (Alex): This makes the assumption that $HOME on the head and + # worker nodes is the same. Also note that `cluster_synced_files` is + # set on the head -> worker updaters only (so `expanduser` is only run + # on the head node). + cluster_synced_files = cluster_synced_files or [] + self.cluster_synced_files = [ + os.path.expanduser(path) for path in cluster_synced_files + ] self.auth_config = auth_config self.is_head_node = is_head_node @@ -166,6 +177,8 @@ def sync_file_mounts(self, sync_cmd, step_numbers=(0, 2)): def do_sync(remote_path, local_path, allow_non_existing_paths=False): if allow_non_existing_paths and not os.path.exists(local_path): + cli_logger.print("sync: {} does not exist. Skipping.", + local_path) # Ignore missing source files. In the future we should support # the --delete-missing-args command to delete files that have # been removed @@ -199,17 +212,21 @@ def do_sync(remote_path, local_path, allow_non_existing_paths=False): _numbered=("[]", previous_steps + 1, total_steps)): for remote_path, local_path in self.file_mounts.items(): do_sync(remote_path, local_path) + previous_steps += 1 if self.cluster_synced_files: with cli_logger.group( "Processing worker file mounts", - _numbered=("[]", previous_steps + 2, total_steps)): + _numbered=("[]", previous_steps + 1, total_steps)): + cli_logger.print("synced files: {}", + str(self.cluster_synced_files)) for path in self.cluster_synced_files: do_sync(path, path, allow_non_existing_paths=True) + previous_steps += 1 else: cli_logger.print( "No worker file mounts to sync", - _numbered=("[]", previous_steps + 2, total_steps)) + _numbered=("[]", previous_steps + 1, total_steps)) def wait_ready(self, deadline): with cli_logger.group(
[autoscaler] cluster_synced_files doesn't support `~` ### What is the problem? `~/filename` is marked as "nonexistent" despite actually existing on the cluster. *Ray version and other system information (Python version, TensorFlow version, OS):* master ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): ``` sent 10,834 bytes received 637 bytes 22,942.00 bytes/sec total size is 684,320 speedup is 59.66 /home/ubuntu/.local/lib/python3.7/site-packages/ray/autoscaler/ from /home/ubuntu/.local/lib/python3.7/site-packages/ray/autoscaler/ [4/6] Processing worker file mounts synced files: ['~/my-cluster-synced-file'] Does ~/my-cluster-synced-file exist? False ``` ``` cluster_name: SYNC_TEST min_workers: 2 max_workers: 2 initial_workers: 2 autoscaling_mode: default target_utilization_fraction: 0.8 idle_timeout_minutes: 60 provider: type: aws region: us-east-1 cache_stopped_nodes: false auth: ssh_user: ubuntu head_node: BlockDeviceMappings: - DeviceName: /dev/sda1 Ebs: VolumeSize: 350 DeleteOnTermination: true VolumeType: gp2 ImageId: latest_dlami InstanceType: t3.xlarge worker_nodes: BlockDeviceMappings: - DeviceName: /dev/sda1 Ebs: VolumeSize: 300 DeleteOnTermination: true VolumeType: gp2 ImageId: latest_dlami InstanceType: t3.xlarge cluster_synced_files: - ~/my-cluster-synced-file file_mounts_sync_continuously: true initialization_commands: [] setup_commands: - pip install --user https://s3-us-west-2.amazonaws.com/ray-wheels/latest/ray-0.9.0.dev0-cp37-cp37m-manylinux1_x86_64.whl head_setup_commands: - tmux new-session -d "python3 -m http.server" - pip install --user https://s3-us-west-2.amazonaws.com/ray-wheels/latest/ray-0.9.0.dev0-cp37-cp37m-manylinux1_x86_64.whl - sleep 5 - curl localhost:8000 -o ~/my-cluster-synced-file worker_setup_commands: [] ``` If we cannot run your script, we cannot fix your issue. - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-09-10T05:28:00
ray-project/ray
10,706
ray-project__ray-10706
[ "10690" ]
7aa55ca82cfe9c1985b341dae81dfc9914617bf0
diff --git a/python/ray/autoscaler/autoscaler.py b/python/ray/autoscaler/autoscaler.py --- a/python/ray/autoscaler/autoscaler.py +++ b/python/ray/autoscaler/autoscaler.py @@ -430,6 +430,8 @@ def _get_node_type_specific_fields(self, node_id: str, return fields def _get_node_specific_docker_config(self, node_id): + if "docker" not in self.config: + return {} docker_config = copy.deepcopy(self.config.get("docker", {})) node_specific_docker = self._get_node_type_specific_fields( node_id, "docker")
diff --git a/python/ray/tests/test_resource_demand_scheduler.py b/python/ray/tests/test_resource_demand_scheduler.py --- a/python/ray/tests/test_resource_demand_scheduler.py +++ b/python/ray/tests/test_resource_demand_scheduler.py @@ -605,6 +605,32 @@ def testUpdateConfig(self): autoscaler.update() self.waitForNodes(0) + def testEmptyDocker(self): + config = MULTI_WORKER_CLUSTER.copy() + del config["docker"] + config["min_workers"] = 0 + config["max_workers"] = 10 + config_path = self.write_config(config) + self.provider = MockProvider() + runner = MockProcessRunner() + autoscaler = StandardAutoscaler( + config_path, + LoadMetrics(), + max_failures=0, + process_runner=runner, + update_interval_s=0) + assert len(self.provider.non_terminated_nodes({})) == 0 + autoscaler.update() + self.waitForNodes(0) + autoscaler.request_resources([{"CPU": 1}]) + autoscaler.update() + self.waitForNodes(1) + assert self.provider.mock_nodes[0].node_type == "m4.large" + autoscaler.request_resources([{"GPU": 8}]) + autoscaler.update() + self.waitForNodes(2) + assert self.provider.mock_nodes[1].node_type == "p2.8xlarge" + if __name__ == "__main__": import sys
[k8s][autoscaler] Possible command parsing issue, worker nodes quickly dying <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? There seem to be multiple bugs in the kuburnetes autoscaler. When running `ray`, commands appear to be parsed incorrectly, resulting in the failure to start the ray runtime on the head node. After manually entering the pod on the head node and starting the ray runtime, a worker node is briefly spawned and quickly dies. ### Version/System I'm working off of the latest master branch ray v0.9.0dev0, with Python3.7 on Ubuntu 20.04. I'm running on a preexisting Kubernetes cluster of two c5.2xlarge nodes on AWS. ### Reproduction Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): I use the provided example-full.yaml file with the only difference being the following configuration: ``` min_workers: 1 max_workers: 1 initial_workers: 1 ``` Started with the following command: `ray up python/ray/autoscaler/kubernetes/example-full.yaml` This produces the following output. The ray runtime fails to start. ``` (myenv) tcw@pop-os:~/projects/ray$ ray up example-full.yaml Commands running under a login shell can produce more output than special processing can handle. Thus, the output from subcommands will be logged as is. Consider using --use-normal-shells, if you tested your workflow and it is compatible. Cluster configuration valid Cluster: default 2020-09-09 20:16:16,175 INFO config.py:68 -- KubernetesNodeProvider: namespace 'ray' not found, attempting to create it 2020-09-09 20:16:16,213 INFO config.py:72 -- KubernetesNodeProvider: successfully created namespace 'ray' 2020-09-09 20:16:16,250 INFO config.py:97 -- KubernetesNodeProvider: autoscaler_service_account 'autoscaler' not found, attempting to create it 2020-09-09 20:16:16,291 INFO config.py:99 -- KubernetesNodeProvider: successfully created autoscaler_service_account 'autoscaler' 2020-09-09 20:16:16,401 INFO config.py:123 -- KubernetesNodeProvider: autoscaler_role 'autoscaler' not found, attempting to create it 2020-09-09 20:16:16,436 INFO config.py:125 -- KubernetesNodeProvider: successfully created autoscaler_role 'autoscaler' 2020-09-09 20:16:16,473 INFO config.py:156 -- KubernetesNodeProvider: autoscaler_role_binding 'autoscaler' not found, attempting to create it 2020-09-09 20:16:16,518 INFO config.py:158 -- KubernetesNodeProvider: successfully created autoscaler_role_binding 'autoscaler' 2020-09-09 20:16:16,554 INFO config.py:189 -- KubernetesNodeProvider: service 'ray-head' not found, attempting to create it 2020-09-09 20:16:16,613 INFO config.py:191 -- KubernetesNodeProvider: successfully created service 'ray-head' 2020-09-09 20:16:16,653 INFO config.py:189 -- KubernetesNodeProvider: service 'ray-workers' not found, attempting to create it 2020-09-09 20:16:16,702 INFO config.py:191 -- KubernetesNodeProvider: successfully created service 'ray-workers' This will create a new cluster [y/N]: y 2020-09-09 20:16:18,507 INFO commands.py:592 -- get_or_create_head_node: Launching new head node... 2020-09-09 20:16:18,507 INFO node_provider.py:87 -- KubernetesNodeProvider: calling create_namespaced_pod (count=1). 2020-09-09 20:16:18,606 INFO commands.py:631 -- get_or_create_head_node: Updating files on head node... 2020-09-09 20:16:18,609 INFO updater.py:97 -- NodeUpdater: ray-head-xn4b5: Updating to c1ad72dff590afc61824084810dc92021a0f3d39 2020-09-09 20:16:18,697 INFO updater.py:219 -- NodeUpdater: ray-head-xn4b5: Waiting for remote shell... 2020-09-09 20:16:18,729 INFO command_runner.py:223 -- NodeUpdater: ray-head-xn4b5: Running ['kubectl', '-n', 'ray', 'exec', '-it', 'ray-head-xn4b5', '--', 'bash', '--login', '-c', '-i', "'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && (uptime)'"] kubectl controls the Kubernetes cluster manager. Find more information at: https://kubernetes.io/docs/reference/kubectl/overview/ Basic Commands (Beginner): create Create a resource from a file or from stdin. expose Take a replication controller, service, deployment or pod and expose it as a new Kubernetes Service run Run a particular image on the cluster set Set specific features on objects Basic Commands (Intermediate): explain Documentation of resources get Display one or many resources edit Edit a resource on the server delete Delete resources by filenames, stdin, resources and names, or by resources and label selector Deploy Commands: rollout Manage the rollout of a resource scale Set a new size for a Deployment, ReplicaSet or Replication Controller autoscale Auto-scale a Deployment, ReplicaSet, or ReplicationController Cluster Management Commands: certificate Modify certificate resources. cluster-info Display cluster info top Display Resource (CPU/Memory/Storage) usage. cordon Mark node as unschedulable uncordon Mark node as schedulable drain Drain node in preparation for maintenance taint Update the taints on one or more nodes Troubleshooting and Debugging Commands: describe Show details of a specific resource or group of resources logs Print the logs for a container in a pod attach Attach to a running container exec Execute a command in a container port-forward Forward one or more local ports to a pod proxy Run a proxy to the Kubernetes API server cp Copy files and directories to and from containers. auth Inspect authorization Advanced Commands: diff Diff live version against would-be applied version apply Apply a configuration to a resource by filename or stdin patch Update field(s) of a resource using strategic merge patch replace Replace a resource by filename or stdin wait Experimental: Wait for a specific condition on one or many resources. convert Convert config files between different API versions kustomize Build a kustomization target from a directory or a remote url. Settings Commands: label Update the labels on a resource annotate Update the annotations on a resource completion Output shell completion code for the specified shell (bash or zsh) Other Commands: api-resources Print the supported API resources on the server api-versions Print the supported API versions on the server, in the form of "group/version" config Modify kubeconfig files plugin Provides utilities for interacting with plugins. version Print the client and server version information Usage: kubectl [flags] [options] Use "kubectl <command> --help" for more information about a given command. Use "kubectl options" for a list of global command-line options (applies to all commands). 2020-09-09 20:16:18,777 INFO log_timer.py:27 -- NodeUpdater: ray-head-xn4b5: Got remote shell [LogTimer=80ms] 2020-09-09 20:16:18,905 INFO command_runner.py:223 -- NodeUpdater: ray-head-xn4b5: Running ['kubectl', '-n', 'ray', 'exec', '-it', 'ray-head-xn4b5', '--', 'bash', '--login', '-c', '-i', "'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && (mkdir -p ~)'"] kubectl controls the Kubernetes cluster manager. Find more information at: https://kubernetes.io/docs/reference/kubectl/overview/ Basic Commands (Beginner): create Create a resource from a file or from stdin. expose Take a replication controller, service, deployment or pod and expose it as a new Kubernetes Service run Run a particular image on the cluster set Set specific features on objects Basic Commands (Intermediate): explain Documentation of resources get Display one or many resources edit Edit a resource on the server delete Delete resources by filenames, stdin, resources and names, or by resources and label selector Deploy Commands: rollout Manage the rollout of a resource scale Set a new size for a Deployment, ReplicaSet or Replication Controller autoscale Auto-scale a Deployment, ReplicaSet, or ReplicationController Cluster Management Commands: certificate Modify certificate resources. cluster-info Display cluster info top Display Resource (CPU/Memory/Storage) usage. cordon Mark node as unschedulable uncordon Mark node as schedulable drain Drain node in preparation for maintenance taint Update the taints on one or more nodes Troubleshooting and Debugging Commands: describe Show details of a specific resource or group of resources logs Print the logs for a container in a pod attach Attach to a running container exec Execute a command in a container port-forward Forward one or more local ports to a pod proxy Run a proxy to the Kubernetes API server cp Copy files and directories to and from containers. auth Inspect authorization Advanced Commands: diff Diff live version against would-be applied version apply Apply a configuration to a resource by filename or stdin patch Update field(s) of a resource using strategic merge patch replace Replace a resource by filename or stdin wait Experimental: Wait for a specific condition on one or many resources. convert Convert config files between different API versions kustomize Build a kustomization target from a directory or a remote url. Settings Commands: label Update the labels on a resource annotate Update the annotations on a resource completion Output shell completion code for the specified shell (bash or zsh) Other Commands: api-resources Print the supported API resources on the server api-versions Print the supported API versions on the server, in the form of "group/version" config Modify kubeconfig files plugin Provides utilities for interacting with plugins. version Print the client and server version information Usage: kubectl [flags] [options] Use "kubectl <command> --help" for more information about a given command. Use "kubectl options" for a list of global command-line options (applies to all commands). 2020-09-09 20:16:18,985 INFO updater.py:420 -- NodeUpdater: ray-head-xn4b5: Syncing /tmp/ray-bootstrap-ixliccej to ~/ray_bootstrap_config.yaml... error: unable to upgrade connection: container not found ("ray-node") rsync: connection unexpectedly closed (0 bytes received so far) [sender] rsync error: error in rsync protocol data stream (code 12) at io.c(235) [sender=3.1.3] 2020-09-09 20:16:19,670 WARNING command_runner.py:255 -- NodeUpdater: ray-head-xn4b5: rsync failed: 'Command '['/home/tcw/projects/mu0_minimal/myenv/lib/python3.7/site-packages/ray/autoscaler/kubernetes/kubectl-rsync.sh', '-avz', '/tmp/ray-bootstrap-ixliccej', 'ray-head-xn4b5@ray:/root/ray_bootstrap_config.yaml']' returned non-zero exit status 12.'. Falling back to 'kubectl cp' 2020-09-09 20:16:20,912 INFO log_timer.py:27 -- NodeUpdater: ray-head-xn4b5: Synced /tmp/ray-bootstrap-ixliccej to ~/ray_bootstrap_config.yaml [LogTimer=2006ms] 2020-09-09 20:16:21,012 INFO command_runner.py:223 -- NodeUpdater: ray-head-xn4b5: Running ['kubectl', '-n', 'ray', 'exec', '-it', 'ray-head-xn4b5', '--', 'bash', '--login', '-c', '-i', "'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && (ray stop)'"] kubectl controls the Kubernetes cluster manager. Find more information at: https://kubernetes.io/docs/reference/kubectl/overview/ Basic Commands (Beginner): create Create a resource from a file or from stdin. expose Take a replication controller, service, deployment or pod and expose it as a new Kubernetes Service run Run a particular image on the cluster set Set specific features on objects Basic Commands (Intermediate): explain Documentation of resources get Display one or many resources edit Edit a resource on the server delete Delete resources by filenames, stdin, resources and names, or by resources and label selector Deploy Commands: rollout Manage the rollout of a resource scale Set a new size for a Deployment, ReplicaSet or Replication Controller autoscale Auto-scale a Deployment, ReplicaSet, or ReplicationController Cluster Management Commands: certificate Modify certificate resources. cluster-info Display cluster info top Display Resource (CPU/Memory/Storage) usage. cordon Mark node as unschedulable uncordon Mark node as schedulable drain Drain node in preparation for maintenance taint Update the taints on one or more nodes Troubleshooting and Debugging Commands: describe Show details of a specific resource or group of resources logs Print the logs for a container in a pod attach Attach to a running container exec Execute a command in a container port-forward Forward one or more local ports to a pod proxy Run a proxy to the Kubernetes API server cp Copy files and directories to and from containers. auth Inspect authorization Advanced Commands: diff Diff live version against would-be applied version apply Apply a configuration to a resource by filename or stdin patch Update field(s) of a resource using strategic merge patch replace Replace a resource by filename or stdin wait Experimental: Wait for a specific condition on one or many resources. convert Convert config files between different API versions kustomize Build a kustomization target from a directory or a remote url. Settings Commands: label Update the labels on a resource annotate Update the annotations on a resource completion Output shell completion code for the specified shell (bash or zsh) Other Commands: api-resources Print the supported API resources on the server api-versions Print the supported API versions on the server, in the form of "group/version" config Modify kubeconfig files plugin Provides utilities for interacting with plugins. version Print the client and server version information Usage: kubectl [flags] [options] Use "kubectl <command> --help" for more information about a given command. Use "kubectl options" for a list of global command-line options (applies to all commands). 2020-09-09 20:16:21,057 INFO command_runner.py:223 -- NodeUpdater: ray-head-xn4b5: Running ['kubectl', '-n', 'ray', 'exec', '-it', 'ray-head-xn4b5', '--', 'bash', '--login', '-c', '-i', "'true && source ~/.bashrc && export OMP_NUM_THREADS=1 PYTHONWARNINGS=ignore && (ulimit -n 65536; ray start --head --num-cpus=$MY_CPU_REQUEST --port=6379 --object-manager-port=8076 --autoscaling-config=~/ray_bootstrap_config.yaml --webui-host 0.0.0.0)'"] kubectl controls the Kubernetes cluster manager. Find more information at: https://kubernetes.io/docs/reference/kubectl/overview/ Basic Commands (Beginner): create Create a resource from a file or from stdin. expose Take a replication controller, service, deployment or pod and expose it as a new Kubernetes Service run Run a particular image on the cluster set Set specific features on objects Basic Commands (Intermediate): explain Documentation of resources get Display one or many resources edit Edit a resource on the server delete Delete resources by filenames, stdin, resources and names, or by resources and label selector Deploy Commands: rollout Manage the rollout of a resource scale Set a new size for a Deployment, ReplicaSet or Replication Controller autoscale Auto-scale a Deployment, ReplicaSet, or ReplicationController Cluster Management Commands: certificate Modify certificate resources. cluster-info Display cluster info top Display Resource (CPU/Memory/Storage) usage. cordon Mark node as unschedulable uncordon Mark node as schedulable drain Drain node in preparation for maintenance taint Update the taints on one or more nodes Troubleshooting and Debugging Commands: describe Show details of a specific resource or group of resources logs Print the logs for a container in a pod attach Attach to a running container exec Execute a command in a container port-forward Forward one or more local ports to a pod proxy Run a proxy to the Kubernetes API server cp Copy files and directories to and from containers. auth Inspect authorization Advanced Commands: diff Diff live version against would-be applied version apply Apply a configuration to a resource by filename or stdin patch Update field(s) of a resource using strategic merge patch replace Replace a resource by filename or stdin wait Experimental: Wait for a specific condition on one or many resources. convert Convert config files between different API versions kustomize Build a kustomization target from a directory or a remote url. Settings Commands: label Update the labels on a resource annotate Update the annotations on a resource completion Output shell completion code for the specified shell (bash or zsh) Other Commands: api-resources Print the supported API resources on the server api-versions Print the supported API versions on the server, in the form of "group/version" config Modify kubeconfig files plugin Provides utilities for interacting with plugins. version Print the client and server version information Usage: kubectl [flags] [options] Use "kubectl <command> --help" for more information about a given command. Use "kubectl options" for a list of global command-line options (applies to all commands). 2020-09-09 20:16:21,096 INFO log_timer.py:27 -- NodeUpdater: ray-head-xn4b5: Ray start commands succeeded [LogTimer=84ms] 2020-09-09 20:16:21,096 INFO log_timer.py:27 -- NodeUpdater: ray-head-xn4b5: Applied config c1ad72dff590afc61824084810dc92021a0f3d39 [LogTimer=2486ms] 2020-09-09 20:16:21,254 INFO commands.py:718 -- get_or_create_head_node: Head node up-to-date, IP address is: 192.168.6.142 To monitor autoscaling activity, you can run: ray exec /home/tcw/projects/mu0_minimal/ray/example-full.yaml 'tail -n 100 -f /tmp/ray/session_*/logs/monitor*' To open a console on the cluster: ray attach /home/tcw/projects/mu0_minimal/ray/example-full.yaml To get a remote shell to the cluster manually, run: kubectl -n ray exec -it ray-head-xn4b5 bash ``` Other ray commands like `ray attach` and `ray down` likewise fail. Upon entering the head node and starting the ray runtime with ``` $ kubectl -n ray exec -it ray-head-xn4b5 bash (base) root@ray-head-xn4b5:/# ray start --head --num-cpus=3 --port=6379 --object-manager-port=8076 --autoscaling-config=~/ray_bootstrap_config.yaml --webui-host 0.0.0.0 ``` and getting logs with ``` (base) root@ray-head-xn4b5:/# tail -n 100 -f /tmp/ray/session_*/logs/monitor* ``` I received the following output: ``` ==> /tmp/ray/session_2020-09-09_17-20-39_779593_74/logs/monitor.err <== docker_config = self._get_node_specific_docker_config(node_id) File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 431, in _get_node_specific_docker_config node_id, "docker") File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 417, in _get_node_type_specific_fields fields = self.config[fields_key] KeyError: 'docker' 2020-09-09 17:21:11,100 INFO autoscaler.py:520 -- Cluster status: 1/1 target nodes (0 pending) - MostDelayedHeartbeats: {'192.168.6.142': 0.13077020645141602} - NodeIdleSeconds: Min=29 Mean=29 Max=29 - NumNodesConnected: 1 - NumNodesUsed: 0.0 - ResourceUsage: 0.0/3.0 CPU, 0.0 GiB/1.21 GiB memory, 0.0 GiB/0.42 GiB object_store_memory - TimeSinceLastHeartbeat: Min=0 Mean=0 Max=0 2020-09-09 17:21:11,112 INFO autoscaler.py:520 -- Cluster status: 1/1 target nodes (0 pending) - MostDelayedHeartbeats: {'192.168.6.142': 0.14269304275512695} - NodeIdleSeconds: Min=29 Mean=29 Max=29 - NumNodesConnected: 1 - NumNodesUsed: 0.0 - ResourceUsage: 0.0/3.0 CPU, 0.0 GiB/1.21 GiB memory, 0.0 GiB/0.42 GiB object_store_memory - TimeSinceLastHeartbeat: Min=0 Mean=0 Max=0 2020-09-09 17:21:11,128 ERROR autoscaler.py:123 -- StandardAutoscaler: Error during autoscaling. Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 121, in update self._update() File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 250, in _update self.should_update(node_id) for node_id in nodes): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 250, in <genexpr> self.should_update(node_id) for node_id in nodes): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 456, in should_update docker_config = self._get_node_specific_docker_config(node_id) File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 431, in _get_node_specific_docker_config node_id, "docker") File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 417, in _get_node_type_specific_fields fields = self.config[fields_key] KeyError: 'docker' 2020-09-09 17:21:11,129 CRITICAL autoscaler.py:130 -- StandardAutoscaler: Too many errors, abort. Error in monitor loop Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/monitor.py", line 313, in run self._run() File "/root/anaconda3/lib/python3.7/site-packages/ray/monitor.py", line 268, in _run self.autoscaler.update() File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 132, in update raise e File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 121, in update self._update() File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 250, in _update self.should_update(node_id) for node_id in nodes): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 250, in <genexpr> self.should_update(node_id) for node_id in nodes): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 456, in should_update docker_config = self._get_node_specific_docker_config(node_id) File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 431, in _get_node_specific_docker_config node_id, "docker") File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 417, in _get_node_type_specific_fields fields = self.config[fields_key] KeyError: 'docker' 2020-09-09 17:21:11,129 ERROR autoscaler.py:554 -- StandardAutoscaler: kill_workers triggered 2020-09-09 17:21:11,134 INFO node_provider.py:121 -- KubernetesNodeProvider: calling delete_namespaced_pod 2020-09-09 17:21:11,148 ERROR autoscaler.py:559 -- StandardAutoscaler: terminated 1 node(s) Error in sys.excepthook: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/worker.py", line 834, in custom_excepthook worker_id = global_worker.worker_id AttributeError: 'Worker' object has no attribute 'worker_id' Original exception was: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/monitor.py", line 368, in <module> monitor.run() File "/root/anaconda3/lib/python3.7/site-packages/ray/monitor.py", line 313, in run self._run() File "/root/anaconda3/lib/python3.7/site-packages/ray/monitor.py", line 268, in _run self.autoscaler.update() File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 132, in update raise e File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 121, in update self._update() File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 250, in _update self.should_update(node_id) for node_id in nodes): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 250, in <genexpr> self.should_update(node_id) for node_id in nodes): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 456, in should_update docker_config = self._get_node_specific_docker_config(node_id) File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 431, in _get_node_specific_docker_config node_id, "docker") File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 417, in _get_node_type_specific_fields fields = self.config[fields_key] KeyError: 'docker' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/monitor.py", line 380, in <module> redis_client, ray_constants.MONITOR_DIED_ERROR, message) File "/root/anaconda3/lib/python3.7/site-packages/ray/utils.py", line 128, in push_error_to_driver_through_redis pubsub_msg.SerializeAsString()) AttributeError: SerializeAsString ==> /tmp/ray/session_2020-09-09_17-20-39_779593_74/logs/monitor.out <== Destroying cluster. Confirm [y/N]: y [automatic, due to --yes] 1 random worker nodes will not be shut down. (due to --keep-min-workers) The head node will not be shut down. (due to --workers-only) No nodes remaining. ==> /tmp/ray/session_latest/logs/monitor.err <== docker_config = self._get_node_specific_docker_config(node_id) File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 431, in _get_node_specific_docker_config node_id, "docker") File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 417, in _get_node_type_specific_fields fields = self.config[fields_key] KeyError: 'docker' 2020-09-09 17:21:11,100 INFO autoscaler.py:520 -- Cluster status: 1/1 target nodes (0 pending) - MostDelayedHeartbeats: {'192.168.6.142': 0.13077020645141602} - NodeIdleSeconds: Min=29 Mean=29 Max=29 - NumNodesConnected: 1 - NumNodesUsed: 0.0 - ResourceUsage: 0.0/3.0 CPU, 0.0 GiB/1.21 GiB memory, 0.0 GiB/0.42 GiB object_store_memory - TimeSinceLastHeartbeat: Min=0 Mean=0 Max=0 2020-09-09 17:21:11,112 INFO autoscaler.py:520 -- Cluster status: 1/1 target nodes (0 pending) - MostDelayedHeartbeats: {'192.168.6.142': 0.14269304275512695} - NodeIdleSeconds: Min=29 Mean=29 Max=29 - NumNodesConnected: 1 - NumNodesUsed: 0.0 - ResourceUsage: 0.0/3.0 CPU, 0.0 GiB/1.21 GiB memory, 0.0 GiB/0.42 GiB object_store_memory - TimeSinceLastHeartbeat: Min=0 Mean=0 Max=0 2020-09-09 17:21:11,128 ERROR autoscaler.py:123 -- StandardAutoscaler: Error during autoscaling. Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 121, in update self._update() File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 250, in _update self.should_update(node_id) for node_id in nodes): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 250, in <genexpr> self.should_update(node_id) for node_id in nodes): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 456, in should_update docker_config = self._get_node_specific_docker_config(node_id) File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 431, in _get_node_specific_docker_config node_id, "docker") File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 417, in _get_node_type_specific_fields fields = self.config[fields_key] KeyError: 'docker' 2020-09-09 17:21:11,129 CRITICAL autoscaler.py:130 -- StandardAutoscaler: Too many errors, abort. Error in monitor loop Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/monitor.py", line 313, in run self._run() File "/root/anaconda3/lib/python3.7/site-packages/ray/monitor.py", line 268, in _run self.autoscaler.update() File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 132, in update raise e File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 121, in update self._update() File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 250, in _update self.should_update(node_id) for node_id in nodes): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 250, in <genexpr> self.should_update(node_id) for node_id in nodes): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 456, in should_update docker_config = self._get_node_specific_docker_config(node_id) File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 431, in _get_node_specific_docker_config node_id, "docker") File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 417, in _get_node_type_specific_fields fields = self.config[fields_key] KeyError: 'docker' 2020-09-09 17:21:11,129 ERROR autoscaler.py:554 -- StandardAutoscaler: kill_workers triggered 2020-09-09 17:21:11,134 INFO node_provider.py:121 -- KubernetesNodeProvider: calling delete_namespaced_pod 2020-09-09 17:21:11,148 ERROR autoscaler.py:559 -- StandardAutoscaler: terminated 1 node(s) Error in sys.excepthook: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/worker.py", line 834, in custom_excepthook worker_id = global_worker.worker_id AttributeError: 'Worker' object has no attribute 'worker_id' Original exception was: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/monitor.py", line 368, in <module> monitor.run() File "/root/anaconda3/lib/python3.7/site-packages/ray/monitor.py", line 313, in run self._run() File "/root/anaconda3/lib/python3.7/site-packages/ray/monitor.py", line 268, in _run self.autoscaler.update() File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 132, in update raise e File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 121, in update self._update() File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 250, in _update self.should_update(node_id) for node_id in nodes): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 250, in <genexpr> self.should_update(node_id) for node_id in nodes): File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 456, in should_update docker_config = self._get_node_specific_docker_config(node_id) File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 431, in _get_node_specific_docker_config node_id, "docker") File "/root/anaconda3/lib/python3.7/site-packages/ray/autoscaler/autoscaler.py", line 417, in _get_node_type_specific_fields fields = self.config[fields_key] KeyError: 'docker' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/root/anaconda3/lib/python3.7/site-packages/ray/monitor.py", line 380, in <module> redis_client, ray_constants.MONITOR_DIED_ERROR, message) File "/root/anaconda3/lib/python3.7/site-packages/ray/utils.py", line 128, in push_error_to_driver_through_redis pubsub_msg.SerializeAsString()) AttributeError: SerializeAsString ==> /tmp/ray/session_latest/logs/monitor.out <== Destroying cluster. Confirm [y/N]: y [automatic, due to --yes] 1 random worker nodes will not be shut down. (due to --keep-min-workers) The head node will not be shut down. (due to --workers-only) No nodes remaining. ``` If we cannot run your script, we cannot fix your issue. - [x] I have verified my script runs in a clean environment and reproduces the issue. - [x] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
@twillkens thanks a bunch for taking the time to open this issue! cc @ijrsvt @wuisawesome could you take a look at this? I've marked it as a release blocker for now.
2020-09-10T05:44:23
ray-project/ray
10,721
ray-project__ray-10721
[ "10720" ]
608bbb1ee7e5624f876a0e7d9882ffc4c27a039c
diff --git a/python/ray/autoscaler/commands.py b/python/ray/autoscaler/commands.py --- a/python/ray/autoscaler/commands.py +++ b/python/ray/autoscaler/commands.py @@ -959,10 +959,11 @@ def rsync(config_file: str, config = _bootstrap_config(config, no_config_cache=no_config_cache) is_file_mount = False - for remote_mount in config.get("file_mounts", {}).keys(): - if remote_mount in (source if down else target): - is_file_mount = True - break + if source and target: + for remote_mount in config.get("file_mounts", {}).keys(): + if (source if down else target).startswith(remote_mount): + is_file_mount = True + break provider = get_node_provider(config["provider"], config["cluster_name"]) try:
[autoscaler] Rsync breaks when `target` or `source` is None <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? When `rsync` is run with `source` and `target` set to `None` and `file_mounts` are included in the YAML, ray fails with ``` None type is not iterable ``` Introduced by #10633 *Ray version and other system information (Python version, TensorFlow version, OS):* ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): If we cannot run your script, we cannot fix your issue. - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-09-10T16:22:34
ray-project/ray
10,734
ray-project__ray-10734
[ "10735" ]
427ba3df47f826f5114c81a98ff29ba1fc8efd13
diff --git a/python/ray/autoscaler/command_runner.py b/python/ray/autoscaler/command_runner.py --- a/python/ray/autoscaler/command_runner.py +++ b/python/ray/autoscaler/command_runner.py @@ -706,6 +706,10 @@ def _docker_expand_user(self, string, any_char=False): return string def run_init(self, *, as_head, file_mounts): + BOOTSTRAP_MOUNTS = [ + "~/ray_bootstrap_config.yaml", "~/ray_bootstrap_key.pem" + ] + image = self.docker_config.get("image") image = self.docker_config.get( f"{'head' if as_head else 'worker'}_image", image) @@ -717,8 +721,14 @@ def run_init(self, *, as_head, file_mounts): self.run("docker pull {}".format(image), run_env="host") + # Bootstrap files cannot be bind mounted because docker opens the + # underlying inode. When the file is switched, docker becomes outdated. + cleaned_bind_mounts = file_mounts.copy() + for mnt in BOOTSTRAP_MOUNTS: + cleaned_bind_mounts.pop(mnt, None) + start_command = docker_start_cmds( - self.ssh_command_runner.ssh_user, image, file_mounts, + self.ssh_command_runner.ssh_user, image, cleaned_bind_mounts, self.container_name, self.docker_config.get("run_options", []) + self.docker_config.get( f"{'head' if as_head else 'worker'}_run_options", [])) @@ -743,7 +753,8 @@ def run_init(self, *, as_head, file_mounts): active_remote_mounts = [ mnt["Destination"] for mnt in active_mounts ] - for remote, local in file_mounts.items(): + # Ignore ray bootstrap files. + for remote, local in cleaned_bind_mounts.items(): remote = self._docker_expand_user(remote) if remote not in active_remote_mounts: cli_logger.error( @@ -753,4 +764,13 @@ def run_init(self, *, as_head, file_mounts): cli_logger.verbose( "Unable to check if file_mounts specified in the YAML " "differ from those on the running container.") + + # Explicitly copy in ray bootstrap files. + for mount in BOOTSTRAP_MOUNTS: + if mount in file_mounts: + self.ssh_command_runner.run( + "docker cp {src} {container}:{dst}".format( + src=os.path.join(DOCKER_MOUNT_PREFIX, mount), + container=self.container_name, + dst=self._docker_expand_user(mount))) self.initialized = True
diff --git a/python/ray/tests/test_autoscaler.py b/python/ray/tests/test_autoscaler.py --- a/python/ray/tests/test_autoscaler.py +++ b/python/ray/tests/test_autoscaler.py @@ -54,15 +54,17 @@ def check_call(self, cmd, *args, **kwargs): def check_output(self, cmd): self.check_call(cmd) return_string = "command-output" - key_to_delete = None - for pattern, pair in self.call_response.items(): + key_to_shrink = None + for pattern, response_list in self.call_response.items(): if pattern in str(cmd): - return_string = pair[0] - if pair[1] - 1 == 0: - key_to_delete = pattern + return_string = response_list[0] + key_to_shrink = pattern break - if key_to_delete: - del self.call_response[key_to_delete] + if key_to_shrink: + self.call_response[key_to_shrink] = self.call_response[ + key_to_shrink][1:] + if len(self.call_response[key_to_shrink]) == 0: + del self.call_response[key_to_shrink] return return_string.encode() @@ -108,8 +110,8 @@ def assert_not_has_call(self, ip, pattern): def clear_history(self): self.calls = [] - def respond_to_call(self, pattern, response, num_times=1): - self.call_response[pattern] = (response, num_times) + def respond_to_call(self, pattern, response_list): + self.call_response[pattern] = response_list class MockProvider(NodeProvider): @@ -400,6 +402,10 @@ def testGetOrCreateHeadNode(self): config_path = self.write_config(SMALL_CLUSTER) self.provider = MockProvider() runner = MockProcessRunner() + runner.respond_to_call("json .Mounts", ["[]"]) + # Two initial calls to docker cp, one before run, two final calls to cp + runner.respond_to_call(".State.Running", + ["false", "false", "false", "true", "true"]) get_or_create_head_node( SMALL_CLUSTER, config_path, @@ -414,6 +420,15 @@ def testGetOrCreateHeadNode(self): runner.assert_has_call("1.2.3.4", "head_setup_cmd") runner.assert_has_call("1.2.3.4", "start_ray_head") self.assertEqual(self.provider.mock_nodes[0].node_type, None) + runner.assert_has_call("1.2.3.4", pattern="docker run") + runner.assert_not_has_call( + "1.2.3.4", pattern="-v /tmp/ray_tmp_mount/~/ray_bootstrap_config") + runner.assert_has_call( + "1.2.3.4", + pattern="docker cp /tmp/ray_tmp_mount/~/ray_bootstrap_key.pem") + runner.assert_has_call( + "1.2.3.4", + pattern="docker cp /tmp/ray_tmp_mount/~/ray_bootstrap_config.yaml") def testScaleUp(self): config_path = self.write_config(SMALL_CLUSTER) diff --git a/python/ray/tests/test_command_runner.py b/python/ray/tests/test_command_runner.py --- a/python/ray/tests/test_command_runner.py +++ b/python/ray/tests/test_command_runner.py @@ -205,7 +205,7 @@ def test_docker_rsync(): remote_file = "/root/protected-file" remote_host_file = f"{DOCKER_MOUNT_PREFIX}{remote_file}" - process_runner.respond_to_call("docker inspect -f", "true") + process_runner.respond_to_call("docker inspect -f", ["true"]) cmd_runner.run_rsync_up( local_mount, remote_mount, options={"file_mount": True}) @@ -222,7 +222,7 @@ def test_docker_rsync(): process_runner.clear_history() ############################## - process_runner.respond_to_call("docker inspect -f", "true") + process_runner.respond_to_call("docker inspect -f", ["true"]) cmd_runner.run_rsync_up( local_file, remote_file, options={"file_mount": False})
[docker] Config updates do not propagate into the container <!--Please include [tune], [rllib], [autoscaler] etc. in the issue title if relevant--> ### What is the problem? Docker file mounts are based on the inode of the host file. When `rsync` replaces the host file, it creates a new inode, effectively making the mount useless. *Ray version and other system information (Python version, TensorFlow version, OS):* ### Reproduction (REQUIRED) 1. Run `ray up myscript.yaml` 2. Update worker nodes to include a larger number of `min/max/initial` `workers` 3. Rerun `ray up myscript.yaml` 4. Run `ray attach myscript.yaml` 5. Open up `~/ray_bootstrap_config.yaml` and note how the changed values _are not reflected there_ If we cannot run your script, we cannot fix your issue. - [ ] I have verified my script runs in a clean environment and reproduces the issue. - [ ] I have verified the issue also occurs with the [latest wheels](https://docs.ray.io/en/latest/installation.html).
2020-09-11T08:52:05
ray-project/ray
10,750
ray-project__ray-10750
[ "10747" ]
6662efd74b28faa176208523ed0bddc39bf7ec67
diff --git a/python/ray/services.py b/python/ray/services.py --- a/python/ray/services.py +++ b/python/ray/services.py @@ -6,6 +6,7 @@ import multiprocessing import os import random +import shutil import signal import socket import subprocess @@ -1314,15 +1315,8 @@ def start_raylet(redis_address, gcs_ip_address, gcs_port = redis_address.split(":") has_java_command = False - try: - java_proc = subprocess.run( - ["java", "-version"], - stdout=subprocess.PIPE, - stderr=subprocess.PIPE) - if java_proc.returncode == 0: - has_java_command = True - except OSError: - pass + if shutil.which("java") is not None: + has_java_command = True ray_java_installed = False try:
[java] Java usage is shown when importing from Python ### What is the problem? *Ray version and other system information (Python version, TensorFlow version, OS):* master ### Reproduction (REQUIRED) Please provide a script that can be run to reproduce the issue. The script should have **no external library dependencies** (i.e., use fake or mock data / environments): I get shown: <img width="505" alt="Screen Shot 2020-09-11 at 12 50 54 PM" src="https://user-images.githubusercontent.com/4529381/92967051-701ee500-f42d-11ea-877c-c6ddc3fde1ef.png"> when calling `ray.init()`. The Python library should have no dependency on Java.
That was easy to find: ``` git diff origin/releases/0.8.7..origin/master python/ray/services.py ``` I betcha dollars to donuts its afa0216280409ff9806640658605e24b64d81ea7, specifically https://github.com/ray-project/ray/commit/afa0216280409ff9806640658605e24b64d81ea7#diff-54ac27010a06993004bec4677c7e583eR1329-R1332 what you really want is to do the equivalent of a `which java` instead of attempting to actually run `java` -- if someone has errantly created a java binary of their own, that's their own fault. Indeed, that dialog is MacOS "helpfully" doing exactly this
2020-09-11T21:59:48
ray-project/ray
10,767
ray-project__ray-10767
[ "10763" ]
9795356ac083eee39b55c2f72428025fbeb6ba13
diff --git a/python/ray/autoscaler/commands.py b/python/ray/autoscaler/commands.py --- a/python/ray/autoscaler/commands.py +++ b/python/ray/autoscaler/commands.py @@ -730,9 +730,13 @@ def get_or_create_head_node(config, cf.bold(" ray exec {}{} {}"), config_file, modifiers, quote(monitor_str)) - cli_logger.print("Connect to a terminal on the cluster head") + cli_logger.print("Connect to a terminal on the cluster head:") cli_logger.print( cf.bold(" ray attach {}{}"), config_file, modifiers) + + remote_shell_str = updater.cmd_runner.remote_shell_command_str() + cli_logger.print("Get a remote shell to the cluster manually:") + cli_logger.print(" {}", remote_shell_str.strip()) finally: provider.cleanup()
diff --git a/python/ray/tests/test_cli.py b/python/ray/tests/test_cli.py --- a/python/ray/tests/test_cli.py +++ b/python/ray/tests/test_cli.py @@ -91,21 +91,23 @@ def _debug_check_line_by_line(result, expected_lines): i = 0 for out in output_lines: - print(out) - if i >= len(expected_lines): i += 1 print("!!!!!! Expected fewer lines") - print("\n".join(output_lines[i:])) + context = [f"CONTEXT: {line}" for line in output_lines[i - 3:i]] + print("\n".join(context)) + extra = [f"-- {line}" for line in output_lines[i:]] + print("\n".join(extra)) break exp = expected_lines[i] matched = re.fullmatch(exp + r" *", out) is not None if not matched: - print(f"!!!!!!! Expected (regex): {repr(exp)}") + print(f"!!! ERROR: Expected (regex): {repr(exp)}") + print(f"Got: {out}") i += 1 if i < len(expected_lines): - print("!!!!!!! Expected (regex):") + print("!!! ERROR: Expected extra lines (regex):") for line in expected_lines[i:]: print(repr(line)) diff --git a/python/ray/tests/test_cli_patterns/test_ray_up.txt b/python/ray/tests/test_cli_patterns/test_ray_up.txt --- a/python/ray/tests/test_cli_patterns/test_ray_up.txt +++ b/python/ray/tests/test_cli_patterns/test_ray_up.txt @@ -42,5 +42,7 @@ Acquiring an up-to-date head node Useful commands Monitor autoscaling with ray exec .+ 'tail -n 100 -f /tmp/ray/session_latest/logs/monitor\*' - Connect to a terminal on the cluster head + Connect to a terminal on the cluster head: ray attach .+ + Get a remote shell to the cluster manually: + ssh .+ diff --git a/python/ray/tests/test_cli_patterns/test_ray_up_record.txt b/python/ray/tests/test_cli_patterns/test_ray_up_record.txt --- a/python/ray/tests/test_cli_patterns/test_ray_up_record.txt +++ b/python/ray/tests/test_cli_patterns/test_ray_up_record.txt @@ -71,5 +71,7 @@ .+\.py.*Useful commands .+\.py.*Monitor autoscaling with .+\.py.* ray exec .+ 'tail -n 100 -f .+ -.+\.py.*Connect to a terminal on the cluster head +.+\.py.*Connect to a terminal on the cluster head: .+\.py.* ray attach .+ +.+\.py.*Get a remote shell to the cluster manually: +.+\.py.* ssh.+\.pem.+
[autoscaler] Usability regressions in new autoscaler output @richardliaw we've regressed the output here, it's a minor issue but super annoying. ``` Useful commands Monitor autoscaling with ray exec example-ml.yaml 'tail -n 100 -f /tmp/ray/session_latest/logs/monitor*' Connect to a terminal on the cluster head ray attach example-ml.yaml ``` It used to print the absolute path of the yaml, so you can copy paste into any terminal. Now, the user has to navigate to the directory YAML for the command to work.
Also, didn't we used to print the raw SSH command if you want to use a direct shell? If we don't print that, then users will have a hard time figuring out how to get direct access if they want more custom command execution. The first issue is fixed on master - https://github.com/ray-project/ray/commit/401195d9359b4d6641776f43706bc3ced414819e#diff-4fd45cbce68a36f9e4f5a688ac099228 If not please provide repro: ``` Useful commands Monitor autoscaling with ray exec /Users/rliaw/dev/cfgs/anyscale.yaml 'tail -n 100 -f /tmp/ray/session_latest/logs/monitor*' Connect to a terminal on the cluster head ray attach /Users/rliaw/dev/cfgs/anyscale.yaml ``` Thanks! Actually, how about the raw SSH command? the raw ssh command is not yet Hm, maybe it was that when you run "ray attach etc.", it used to print the raw SSH command being executed. Can we ensure this is printed out so the commands aren't so "magic"? Yeah - I'll add a line.
2020-09-14T00:06:24
ray-project/ray
10,775
ray-project__ray-10775
[ "10653" ]
829a2307dff07f8775aac2b766c5ffa56976e03a
diff --git a/rllib/agents/ddpg/ddpg_tf_policy.py b/rllib/agents/ddpg/ddpg_tf_policy.py --- a/rllib/agents/ddpg/ddpg_tf_policy.py +++ b/rllib/agents/ddpg/ddpg_tf_policy.py @@ -175,7 +175,6 @@ def ddpg_actor_critic_loss(policy, model, _, train_batch): if twin_q: td_error = q_t_selected - q_t_selected_target twin_td_error = twin_q_t_selected - q_t_selected_target - td_error = td_error + twin_td_error if use_huber: errors = huber_loss(td_error, huber_threshold) + \ huber_loss(twin_td_error, huber_threshold) diff --git a/rllib/agents/ddpg/ddpg_torch_policy.py b/rllib/agents/ddpg/ddpg_torch_policy.py --- a/rllib/agents/ddpg/ddpg_torch_policy.py +++ b/rllib/agents/ddpg/ddpg_torch_policy.py @@ -124,7 +124,6 @@ def ddpg_actor_critic_loss(policy, model, _, train_batch): if twin_q: td_error = q_t_selected - q_t_selected_target twin_td_error = twin_q_t_selected - q_t_selected_target - td_error = td_error + twin_td_error if use_huber: errors = huber_loss(td_error, huber_threshold) \ + huber_loss(twin_td_error, huber_threshold)
diff --git a/rllib/agents/ddpg/tests/test_ddpg.py b/rllib/agents/ddpg/tests/test_ddpg.py --- a/rllib/agents/ddpg/tests/test_ddpg.py +++ b/rllib/agents/ddpg/tests/test_ddpg.py @@ -495,7 +495,6 @@ def _ddpg_loss_helper(self, train_batch, weights, ks, fw, gamma, td_error = q_t_selected - q_t_selected_target twin_td_error = twin_q_t_selected - q_t_selected_target - td_error = td_error + twin_td_error errors = huber_loss(td_error, huber_threshold) + \ huber_loss(twin_td_error, huber_threshold)
[rllib] DDPG TWIN_Q LOSS ERROR The TD3 loss according to OpenAI Spinning up is: ![L1](https://spinningup.openai.com/en/latest/_images/math/7d5c18f49a242cc3eec554f717fe4f3bfc119bab.svg) ![L2](https://spinningup.openai.com/en/latest/_images/math/cd73726a8a3845ade467aed57714912f868f6b36.svg) They implement this as you would expect: ```` # MSE loss against Bellman backup loss_q1 = ((q1 - backup)**2).mean() loss_q2 = ((q2 - backup)**2).mean() loss_q = loss_q1 + loss_q2 ```` The rllib implementation is off and I ran into training issues when using it. Tensorflow: https://github.com/ray-project/ray/blob/39c598bab0929f007a49daa7262a7c9ee6492221/rllib/agents/ddpg/ddpg_tf_policy.py#L174-L184 Pytorch: https://github.com/ray-project/ray/blob/5851e893eee491e569fe1421314172bd5b37d583/rllib/agents/ddpg/ddpg_torch_policy.py#L124-L134 The error is on line 177 and 127 respectively where the td_error and twin_td_error are added together and then squared a couple lines later. Removing that line would bring the implementation in line with the spinningup reference implementation. This worked for me to improve training on my problem. They also do not multiply by the constant (0.5) but I would not expect that to make much of a difference either way.
Interesting, do you want to push a fix. Also cc @michaelzhiluo if you could take a look here.
2020-09-14T13:22:52
ray-project/ray
10,821
ray-project__ray-10821
[ "10809" ]
0d9d34a3c1d584ce875cc121677d2683802e0ea5
diff --git a/python/ray/utils.py b/python/ray/utils.py --- a/python/ray/utils.py +++ b/python/ray/utils.py @@ -499,22 +499,55 @@ def get_system_memory(): return psutil_memory_in_bytes -def _get_docker_cpus(): - # 1. Try using CFS Quota (https://bugs.openjdk.java.net/browse/JDK-8146115) - # 2. Try Nproc (CPU sets) - cpu_quota_file_name = "/sys/fs/cgroup/cpu/cpu.cfs_quota_us" - cpu_share_file_name = "/sys/fs/cgroup/cpu/cpu.cfs_period_us" - num_cpus = 0 +def _get_docker_cpus( + cpu_quota_file_name="/sys/fs/cgroup/cpu/cpu.cfs_quota_us", + cpu_share_file_name="/sys/fs/cgroup/cpu/cpu.cfs_period_us", + cpuset_file_name="/sys/fs/cgroup/cpuset/cpuset.cpus"): + # TODO (Alex): Don't implement this logic oursleves. + # Docker has 2 underyling ways of implementing CPU limits: + # https://docs.docker.com/config/containers/resource_constraints/#configure-the-default-cfs-scheduler + # 1. --cpuset-cpus 2. --cpus or --cpu-quota/--cpu-period (--cpu-shares is a + # soft limit so we don't worry about it). For Ray's purposes, if we use + # docker, the number of vCPUs on a machine is whichever is set (ties broken + # by smaller value). + + cpu_quota = None + # See: https://bugs.openjdk.java.net/browse/JDK-8146115 if os.path.exists(cpu_quota_file_name) and os.path.exists( cpu_quota_file_name): - with open(cpu_quota_file_name, "r") as f: - num_cpus = int(f.read()) - if num_cpus != -1: - with open(cpu_share_file_name, "r") as f: - num_cpus /= int(f.read()) - return num_cpus - - return int(subprocess.check_output("nproc")) + try: + with open(cpu_quota_file_name, "r") as quota_file, open( + cpu_share_file_name, "r") as period_file: + cpu_quota = float(quota_file.read()) / float( + period_file.read()) + except Exception as e: + logger.exception("Unexpected error calculating docker cpu quota.", + e) + if cpu_quota < 0: + cpu_quota = None + + cpuset_num = None + if os.path.exists(cpuset_file_name): + try: + with open(cpuset_file_name) as cpuset_file: + ranges_as_string = cpuset_file.read() + ranges = ranges_as_string.split(",") + cpu_ids = [] + for num_or_range in ranges: + if "-" in num_or_range: + start, end = num_or_range.split("-") + cpu_ids.extend(list(range(int(start), int(end) + 1))) + else: + cpu_ids.append(int(num_or_range)) + cpuset_num = len(cpu_ids) + except Exception as e: + logger.exception("Unexpected error calculating docker cpuset ids.", + e) + + if cpu_quota and cpuset_num: + return min(cpu_quota, cpuset_num) + else: + return cpu_quota or cpuset_num def get_num_cpus(): @@ -531,8 +564,7 @@ def get_num_cpus(): # Not easy to get cpu count in docker, see: # https://bugs.python.org/issue36054 docker_count = _get_docker_cpus() - if docker_count != cpu_count: - cpu_count = docker_count + if docker_count is not None and docker_count != cpu_count: if "RAY_DISABLE_DOCKER_CPU_WARNING" not in os.environ: logger.warning( "Detecting docker specified CPUs. In " @@ -543,6 +575,15 @@ def get_num_cpus(): "`RAY_USE_MULTIPROCESSING_CPU_COUNT=1` as an env var " "before starting Ray. Set the env var: " "`RAY_DISABLE_DOCKER_CPU_WARNING=1` to mute this warning.") + # TODO (Alex): We should probably add support for fractional cpus. + if int(docker_count) != float(docker_count): + logger.warning( + f"Ray currently does not support initializing Ray" + f"with fractional cpus. Your num_cpus will be " + f"truncated from {docker_count} to " + f"{int(docker_count)}.") + docker_count = int(docker_count) + cpu_count = docker_count except Exception: # `nproc` and cgroup are linux-only. If docker only works on linux
diff --git a/python/ray/tests/test_advanced_3.py b/python/ray/tests/test_advanced_3.py --- a/python/ray/tests/test_advanced_3.py +++ b/python/ray/tests/test_advanced_3.py @@ -4,6 +4,7 @@ import os import sys import socket +import tempfile import time import numpy as np @@ -736,6 +737,56 @@ def initialized(self): assert ray.available_resources()[resource_name] < quantity +def test_detect_docker_cpus(): + # No limits set + with tempfile.NamedTemporaryFile( + "w") as quota_file, tempfile.NamedTemporaryFile( + "w") as period_file, tempfile.NamedTemporaryFile( + "w") as cpuset_file: + quota_file.write("-1") + period_file.write("100000") + cpuset_file.write("0-63") + quota_file.flush() + period_file.flush() + cpuset_file.flush() + assert ray.utils._get_docker_cpus( + cpu_quota_file_name=quota_file.name, + cpu_share_file_name=period_file.name, + cpuset_file_name=cpuset_file.name) == 64 + + # No cpuset used + with tempfile.NamedTemporaryFile( + "w") as quota_file, tempfile.NamedTemporaryFile( + "w") as period_file, tempfile.NamedTemporaryFile( + "w") as cpuset_file: + quota_file.write("-1") + period_file.write("100000") + cpuset_file.write("0-10,20,50-63") + quota_file.flush() + period_file.flush() + cpuset_file.flush() + assert ray.utils._get_docker_cpus( + cpu_quota_file_name=quota_file.name, + cpu_share_file_name=period_file.name, + cpuset_file_name=cpuset_file.name) == 26 + + # Quota set + with tempfile.NamedTemporaryFile( + "w") as quota_file, tempfile.NamedTemporaryFile( + "w") as period_file, tempfile.NamedTemporaryFile( + "w") as cpuset_file: + quota_file.write("42") + period_file.write("100") + cpuset_file.write("0-63") + quota_file.flush() + period_file.flush() + cpuset_file.flush() + assert ray.utils._get_docker_cpus( + cpu_quota_file_name=quota_file.name, + cpu_share_file_name=period_file.name, + cpuset_file_name=cpuset_file.name) == 0.42 + + if __name__ == "__main__": import pytest sys.exit(pytest.main(["-v", __file__]))
[docker] / [core] Regression in detection of multi CPU count after docker fix ### What is the problem? The Ray snapshot 1.1.0 will detect only 1 CPU on certain multi-core systems **when not using docker** due to this change from https://github.com/ray-project/ray/pull/10507/files#diff-dd52f82c5ae34621334f725acaeacc81R502 ``` def _detect_docker_cpus() cpu_quota_file_name = "/sys/fs/cgroup/cpu/cpu.cfs_quota_us" cpu_share_file_name = "/sys/fs/cgroup/cpu/cpu.cfs_period_us" num_cpus = 0 if os.path.exists(cpu_quota_file_name) and os.path.exists( cpu_quota_file_name): with open(cpu_quota_file_name, "r") as f: num_cpus = int(f.read()) if num_cpus != -1: with open(cpu_share_file_name, "r") as f: num_cpus /= int(f.read()) return num_cpus ``` My current cgroup output is: ``` # cat /proc/self/cgroup 11:cpuset:/ 10:hugetlb:/ 9:perf_event:/ 8:cpu,cpuacct:/ 7:devices:/ 6:memory:/ 5:net_cls,net_prio:/ 4:blkio:/ 3:freezer:/ 2:pids:/ 1:name=systemd:/user.slice/user-0.slice/session-362.scope ``` Code was added to python/ray/utils.py which now requires an explicit environmental variable to be set to detect the number of CPUs. This is undesirable. The code is in the area of ``` if "RAY_USE_MULTIPROCESSING_CPU_COUNT" in os.environ:``` Please provide a better more robust way of detecting whether you are in docker and do not require a generic setup to explicitly set the env variable. I propose the following to detect docker: ``` def is_docker(): path = '/proc/self/cgroup' return ( os.path.exists('/.dockerenv') or os.path.isfile(path) and any('docker' in line for line in open(path)) ) ```
2020-09-16T02:34:44
ray-project/ray
10,840
ray-project__ray-10840
[ "10839" ]
9d0a9e73de7d8daf2f0056dd7f89786db240aa5d
diff --git a/python/ray/tune/__init__.py b/python/ray/tune/__init__.py --- a/python/ray/tune/__init__.py +++ b/python/ray/tune/__init__.py @@ -9,9 +9,9 @@ from ray.tune.trainable import Trainable from ray.tune.durable_trainable import DurableTrainable from ray.tune.suggest import grid_search -from ray.tune.session import (report, get_trial_dir, get_trial_name, - get_trial_id, make_checkpoint_dir, - save_checkpoint, checkpoint_dir) +from ray.tune.session import ( + report, get_trial_dir, get_trial_name, get_trial_id, make_checkpoint_dir, + save_checkpoint, checkpoint_dir, is_session_enabled) from ray.tune.progress_reporter import (ProgressReporter, CLIReporter, JupyterNotebookReporter) from ray.tune.sample import (function, sample_from, uniform, quniform, choice, @@ -28,6 +28,7 @@ "qrandint", "randn", "qrandn", "loguniform", "qloguniform", "ExperimentAnalysis", "Analysis", "CLIReporter", "JupyterNotebookReporter", "ProgressReporter", "report", "get_trial_dir", "get_trial_name", - "get_trial_id", "make_checkpoint_dir", "save_checkpoint", "checkpoint_dir", - "SyncConfig", "create_searcher", "create_scheduler" + "get_trial_id", "make_checkpoint_dir", "save_checkpoint", + "is_session_enabled", "checkpoint_dir", "SyncConfig", "create_searcher", + "create_scheduler" ] diff --git a/python/ray/tune/session.py b/python/ray/tune/session.py --- a/python/ray/tune/session.py +++ b/python/ray/tune/session.py @@ -7,6 +7,12 @@ _session = None +def is_session_enabled() -> bool: + """Returns True if running within an Tune process.""" + global _session + return _session is not None + + def get_session(): global _session if not _session:
diff --git a/python/ray/tune/tests/test_function_api.py b/python/ray/tune/tests/test_function_api.py --- a/python/ray/tune/tests/test_function_api.py +++ b/python/ray/tune/tests/test_function_api.py @@ -400,6 +400,18 @@ def train(config, checkpoint_dir=None): trial_dfs = list(analysis.trial_dataframes.values()) assert len(trial_dfs[0]["training_iteration"]) == 10 + def testEnabled(self): + def train(config, checkpoint_dir=None): + is_active = tune.is_session_enabled() + if is_active: + tune.report(active=is_active) + return is_active + + assert train({}) is False + analysis = tune.run(train) + t = analysis.trials[0] + assert t.last_result["active"] + def testBlankCheckpoint(self): def train(config, checkpoint_dir=None): restored = bool(checkpoint_dir)
[tune] How to test if I'm running inside a tune session? Is there an API to test if I'm running inside a tune session? I'd like to conditionally call `tune.report()` in my code. There are functions like `get_trial_dir`, `get_trial_name`, `get_trial_id`, that internally call `get_session()`. I guess I could use one of them see if they return `None` or not. But they also log a warning when they can't find a session which is not ideal.
2020-09-16T22:43:46
ray-project/ray
10,874
ray-project__ray-10874
[ "10869" ]
1b295a17cb3913c77568b214ab0ae8f964a2ee37
diff --git a/python/ray/autoscaler/resource_demand_scheduler.py b/python/ray/autoscaler/resource_demand_scheduler.py --- a/python/ray/autoscaler/resource_demand_scheduler.py +++ b/python/ray/autoscaler/resource_demand_scheduler.py @@ -56,6 +56,7 @@ def get_nodes_to_launch(self, nodes: List[NodeID], nodes: List of existing nodes in the cluster. pending_nodes: Summary of node types currently being launched. resource_demands: Vector of resource demands from the scheduler. + usage_by_ip: Mapping from ip to available resources. """ if resource_demands is None: @@ -113,7 +114,6 @@ def add_node(node_type, available_resources=None): tags = self.provider.node_tags(node_id) if TAG_RAY_USER_NODE_TYPE in tags: node_type = tags[TAG_RAY_USER_NODE_TYPE] - node_type_counts[node_type] += 1 ip = self.provider.internal_ip(node_id) available_resources = usage_by_ip.get(ip) add_node(node_type, available_resources)
diff --git a/python/ray/tests/test_resource_demand_scheduler.py b/python/ray/tests/test_resource_demand_scheduler.py --- a/python/ray/tests/test_resource_demand_scheduler.py +++ b/python/ray/tests/test_resource_demand_scheduler.py @@ -180,6 +180,27 @@ def test_get_nodes_to_launch_limits(): assert to_launch == [] +def test_calculate_node_resources(): + provider = MockProvider() + scheduler = ResourceDemandScheduler(provider, TYPES_A, 10) + + provider.create_node({}, {TAG_RAY_USER_NODE_TYPE: "p2.8xlarge"}, 2) + + nodes = provider.non_terminated_nodes({}) + + ips = provider.non_terminated_node_ips({}) + # 2 free p2.8xls + utilizations = {ip: {"GPU": 8} for ip in ips} + # 1 more on the way + pending_nodes = {"p2.8xlarge": 1} + # requires 4 p2.8xls (only 3 are in cluster/pending) + demands = [{"GPU": 8}] * (len(utilizations) + 2) + to_launch = scheduler.get_nodes_to_launch(nodes, pending_nodes, demands, + utilizations) + + assert to_launch == [("p2.8xlarge", 1)] + + class LoadMetricsTest(unittest.TestCase): def testResourceDemandVector(self): lm = LoadMetrics()
[autoscaler] Inconsistent reporting of worker node types On master, it seems I have a cluster with a cpu_4_ondemand head node and cpu_64_ondemand worker (68 CPUs in total), but somehow the debug string is reporting 2 workers of cpu_64_ondemand (which is wrong, since there is only 1 worker node, and the head node is cpu_4_ondemand). YAML: https://gist.github.com/ericl/f9c32b8bbbe4e1aed43899832730f70d (switching the Ray version to latest wheel) ``` - ResourceUsage: 0.0/68.0 CPU, 0.0/1.0 Custom1, 0.0 GiB/198.65 GiB memory, 0.0 GiB/59.19 GiB object_store_memory - TimeSinceLastHeartbeat: Min=0 Mean=0 Max=0 Worker node types: - cpu_64_ondemand: 2 ``` cc @wuisawesome
this is a bug in the debug string right? or are you noticing the wrong number of instances? ``` 2020-09-17 23:33:35,234 INFO resource_demand_scheduler.py:68 -- Cluster resources: [{'node:172.30.2.59': 1.0, 'CPU': 4.0, 'memory': 187.0, 'object_store_memory': 64.0}, {'CPU': 64.0, 'memory': 3598.0, 'node:172.30.2.114': 1.0, 'object_store_memory': 1064.0}] 2020-09-17 23:33:35,234 INFO resource_demand_scheduler.py:69 -- Node counts: defaultdict(<class 'int'>, {'cpu_4_ondemand': 2, 'cpu_64_ondemand': 2}) 2020-09-17 23:33:35,235 INFO resource_demand_scheduler.py:72 -- Resource demands: [{'accelerator_type:V100': 0.001, 'CPU': 1.0}] 2020-09-17 23:33:35,235 INFO resource_demand_scheduler.py:73 -- Unfulfilled demands: [{'accelerator_type:V100': 0.001, 'CPU': 1.0}] 2020-09-17 23:33:35,235 INFO resource_demand_scheduler.py:79 -- Node requests: [] ``` Here's some more inconsistency, with the node counts dict reporting 4 nodes instead of 2. I'm not sure if this is impacting autoscaling, but it seems likely if the node counts dict is reused for scheduling. ok i can reproduce the logging issue, doesn't seem to impact autoscaling behavior IIUC now the conclusion is it does impact the behavior, right? Yeah this will impact behavior if we need to instances of a type that is already up
2020-09-18T00:49:02
ray-project/ray
10,953
ray-project__ray-10953
[ "10923" ]
273015814df65ea1d187abb3d9afc5da46c145eb
diff --git a/python/ray/tune/tune.py b/python/ray/tune/tune.py --- a/python/ray/tune/tune.py +++ b/python/ray/tune/tune.py @@ -230,7 +230,7 @@ def run( max_failures (int): Try to recover a trial at least this many times. Ray will recover from the latest checkpoint if present. Setting to -1 will lead to infinite recovery retries. - Setting to 0 will disable retries. Defaults to 3. + Setting to 0 will disable retries. Defaults to 0. fail_fast (bool | str): Whether to fail upon the first error. If fail_fast='raise' provided, Tune will automatically raise the exception received by the Trainable. fail_fast='raise'
Issue on page /tune/api_docs/execution.html [tune] Documentation regarding `max_failures`: ``` max_failures (int): Try to recover a trial at least this many times. Ray will recover from the latest checkpoint if present. Setting to -1 will lead to infinite recovery retries. Setting to 0 will disable retries. Defaults to 3. ``` However, if you look into `tune.run`, the default is `max_failure=0`. I think that the doc is outdated, because the observed behavior is as if `max_failure=0` is true.
2020-09-22T14:26:40
ray-project/ray
10,979
ray-project__ray-10979
[ "10902" ]
a260e660168fe32a2b7e7084666bcc2e760f556b
diff --git a/python/ray/tune/trial_runner.py b/python/ray/tune/trial_runner.py --- a/python/ray/tune/trial_runner.py +++ b/python/ray/tune/trial_runner.py @@ -454,7 +454,10 @@ def _get_next_trial(self): """ trials_done = all(trial.is_finished() for trial in self._trials) wait_for_trial = trials_done and not self._search_alg.is_finished() - self._update_trial_queue(blocking=wait_for_trial) + # Only fetch a new trial if we have no pending trial + if not any(trial.status == Trial.PENDING for trial in self._trials) \ + or wait_for_trial: + self._update_trial_queue(blocking=wait_for_trial) with warn_if_slow("choose_trial_to_run"): trial = self._scheduler_alg.choose_trial_to_run(self) logger.debug("Running trial {}".format(trial))
diff --git a/python/ray/tune/tests/test_trial_runner.py b/python/ray/tune/tests/test_trial_runner.py --- a/python/ray/tune/tests/test_trial_runner.py +++ b/python/ray/tune/tests/test_trial_runner.py @@ -288,6 +288,44 @@ def on_trial_result(self, trial_runner, trial, result): self.assertEqual(trials[0].status, Trial.RUNNING) self.assertEqual(runner.trial_executor._committed_resources.cpu, 2) + def testQueueFilling(self): + ray.init(num_cpus=4) + + def f1(config): + for i in range(10): + yield i + + tune.register_trainable("f1", f1) + + search_alg = BasicVariantGenerator() + search_alg.add_configurations({ + "foo": { + "run": "f1", + "num_samples": 100, + "config": { + "a": tune.sample_from(lambda spec: 5.0 / 7), + "b": tune.sample_from(lambda spec: "long" * 40) + }, + "resources_per_trial": { + "cpu": 2 + } + } + }) + + runner = TrialRunner(search_alg=search_alg) + + runner.step() + runner.step() + runner.step() + self.assertEqual(len(runner._trials), 3) + + runner.step() + self.assertEqual(len(runner._trials), 3) + + self.assertEqual(runner._trials[0].status, Trial.RUNNING) + self.assertEqual(runner._trials[1].status, Trial.RUNNING) + self.assertEqual(runner._trials[2].status, Trial.PENDING) + if __name__ == "__main__": import pytest
[tune] The search algorithms generate more trials than necessary ### What is the problem? I was under the impression that #10802 should avoid producing too many trials before resources are ready. However, on master, it seems like many trials are generated (especially when using search algorithms) before resources are ready, and this happens far into the experiment. cc @krfricke
2020-09-23T15:36:06