{ "name": "root", "gauges": { "Huggy.Policy.Entropy.mean": { "value": 1.4042747020721436, "min": 1.4042747020721436, "max": 1.4274088144302368, "count": 40 }, "Huggy.Policy.Entropy.sum": { "value": 71106.8515625, "min": 69009.8125, "max": 77716.421875, "count": 40 }, "Huggy.Environment.EpisodeLength.mean": { "value": 84.7551724137931, "min": 77.12401883830455, "max": 382.79545454545456, "count": 40 }, "Huggy.Environment.EpisodeLength.sum": { "value": 49158.0, "min": 48872.0, "max": 50529.0, "count": 40 }, "Huggy.Step.mean": { "value": 1999904.0, "min": 49911.0, "max": 1999904.0, "count": 40 }, "Huggy.Step.sum": { "value": 1999904.0, "min": 49911.0, "max": 1999904.0, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.mean": { "value": 2.4087376594543457, "min": 0.01227047573775053, "max": 2.51291823387146, "count": 40 }, "Huggy.Policy.ExtrinsicValueEstimate.sum": { "value": 1397.06787109375, "min": 1.6074323654174805, "max": 1575.545166015625, "count": 40 }, "Huggy.Environment.CumulativeReward.mean": { "value": 3.744817124358539, "min": 1.9632456996513687, "max": 4.025326363768494, "count": 40 }, "Huggy.Environment.CumulativeReward.sum": { "value": 2171.9939321279526, "min": 257.1851866543293, "max": 2458.47745770216, "count": 40 }, "Huggy.Policy.ExtrinsicReward.mean": { "value": 3.744817124358539, "min": 1.9632456996513687, "max": 4.025326363768494, "count": 40 }, "Huggy.Policy.ExtrinsicReward.sum": { "value": 2171.9939321279526, "min": 257.1851866543293, "max": 2458.47745770216, "count": 40 }, "Huggy.Losses.PolicyLoss.mean": { "value": 0.01760850117861992, "min": 0.010704247847509882, "max": 0.02134924580216951, "count": 40 }, "Huggy.Losses.PolicyLoss.sum": { "value": 0.052825503535859755, "min": 0.021408495695019764, "max": 0.05674871765622812, "count": 40 }, "Huggy.Losses.ValueLoss.mean": { "value": 0.05397070199251175, "min": 0.021507739617178837, "max": 0.06417246218770742, "count": 40 }, "Huggy.Losses.ValueLoss.sum": { "value": 0.16191210597753525, "min": 0.043015479234357674, "max": 0.1807886337240537, "count": 40 }, "Huggy.Policy.LearningRate.mean": { "value": 3.4686988438e-06, "min": 3.4686988438e-06, "max": 0.000295304326565225, "count": 40 }, "Huggy.Policy.LearningRate.sum": { "value": 1.04060965314e-05, "min": 1.04060965314e-05, "max": 0.0008440807686397499, "count": 40 }, "Huggy.Policy.Epsilon.mean": { "value": 0.10115620000000002, "min": 0.10115620000000002, "max": 0.198434775, "count": 40 }, "Huggy.Policy.Epsilon.sum": { "value": 0.30346860000000003, "min": 0.20749910000000005, "max": 0.58136025, "count": 40 }, "Huggy.Policy.Beta.mean": { "value": 6.769437999999999e-05, "min": 6.769437999999999e-05, "max": 0.004921895272500001, "count": 40 }, "Huggy.Policy.Beta.sum": { "value": 0.00020308314, "min": 0.00020308314, "max": 0.014069876475000003, "count": 40 }, "Huggy.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 }, "Huggy.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 40 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1685585518", "python_version": "3.10.11 (main, Apr 5 2023, 14:15:10) [GCC 9.4.0]", "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics", "mlagents_version": "0.31.0.dev0", "mlagents_envs_version": "0.31.0.dev0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.11.0+cu102", "numpy_version": "1.21.2", "end_time_seconds": "1685587930" }, "total": 2412.8882849849997, "count": 1, "self": 0.3832756110000446, "children": { "run_training.setup": { "total": 0.04082583400003159, "count": 1, "self": 0.04082583400003159 }, "TrainerController.start_learning": { "total": 2412.46418354, "count": 1, "self": 4.459569020992603, "children": { "TrainerController._reset_env": { "total": 3.810170492999987, "count": 1, "self": 3.810170492999987 }, "TrainerController.advance": { "total": 2404.0734931030074, "count": 232517, "self": 4.570713754051212, "children": { "env_step": { "total": 1875.2262227860278, "count": 232517, "self": 1587.066748465922, "children": { "SubprocessEnvManager._take_step": { "total": 285.28032259309293, "count": 232517, "self": 16.791005039125594, "children": { "TorchPolicy.evaluate": { "total": 268.48931755396734, "count": 222998, "self": 268.48931755396734 } } }, "workers": { "total": 2.879151727012811, "count": 232517, "self": 0.0, "children": { "worker_root": { "total": 2404.69376454195, "count": 232517, "is_parallel": true, "self": 1106.4628520809365, "children": { "run_training.setup": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "steps_from_proto": { "total": 0.0008878190000132236, "count": 1, "is_parallel": true, "self": 0.00025564400004896015, "children": { "_process_rank_one_or_two_observation": { "total": 0.0006321749999642634, "count": 2, "is_parallel": true, "self": 0.0006321749999642634 } } }, "UnityEnvironment.step": { "total": 0.02921345299995437, "count": 1, "is_parallel": true, "self": 0.00032536099985236433, "children": { "UnityEnvironment._generate_step_input": { "total": 0.00022860700005367107, "count": 1, "is_parallel": true, "self": 0.00022860700005367107 }, "communicator.exchange": { "total": 0.02795144000003802, "count": 1, "is_parallel": true, "self": 0.02795144000003802 }, "steps_from_proto": { "total": 0.0007080450000103156, "count": 1, "is_parallel": true, "self": 0.00020765200002870188, "children": { "_process_rank_one_or_two_observation": { "total": 0.0005003929999816137, "count": 2, "is_parallel": true, "self": 0.0005003929999816137 } } } } } } }, "UnityEnvironment.step": { "total": 1298.2309124610133, "count": 232516, "is_parallel": true, "self": 39.517795996066525, "children": { "UnityEnvironment._generate_step_input": { "total": 78.70435628295121, "count": 232516, "is_parallel": true, "self": 78.70435628295121 }, "communicator.exchange": { "total": 1084.5762785920558, "count": 232516, "is_parallel": true, "self": 1084.5762785920558 }, "steps_from_proto": { "total": 95.43248158993981, "count": 232516, "is_parallel": true, "self": 34.79765381285597, "children": { "_process_rank_one_or_two_observation": { "total": 60.63482777708384, "count": 465032, "is_parallel": true, "self": 60.63482777708384 } } } } } } } } } } }, "trainer_advance": { "total": 524.2765565629286, "count": 232517, "self": 6.781448870954591, "children": { "process_trajectory": { "total": 134.3942335749755, "count": 232517, "self": 133.13231407297553, "children": { "RLTrainer._checkpoint": { "total": 1.2619195019999552, "count": 10, "self": 1.2619195019999552 } } }, "_update_policy": { "total": 383.10087411699857, "count": 97, "self": 323.61444694500096, "children": { "TorchPPOOptimizer.update": { "total": 59.486427171997605, "count": 2910, "self": 59.486427171997605 } } } } } } }, "trainer_threads": { "total": 1.0859998837986495e-06, "count": 1, "self": 1.0859998837986495e-06 }, "TrainerController._save_models": { "total": 0.12094983699989825, "count": 1, "self": 0.0024701379998077755, "children": { "RLTrainer._checkpoint": { "total": 0.11847969900009048, "count": 1, "self": 0.11847969900009048 } } } } } } }