{ "name": "root", "gauges": { "Agent.Policy.Entropy.mean": { "value": 1.375830054283142, "min": 1.3740845918655396, "max": 1.4282022714614868, "count": 200 }, "Agent.Policy.Entropy.sum": { "value": 7949.5458984375, "min": 6871.9453125, "max": 9895.36328125, "count": 200 }, "Agent.DroneBasedReforestation.TreeDropCount.mean": { "value": 0.9866666666666667, "min": 0.3888888888888889, "max": 0.9922480620155039, "count": 200 }, "Agent.DroneBasedReforestation.TreeDropCount.sum": { "value": 74.0, "min": 7.0, "max": 128.0, "count": 200 }, "Agent.DroneBasedReforestation.RechargeEnergyCount.mean": { "value": 10.306666666666667, "min": 5.125, "max": 44.833333333333336, "count": 200 }, "Agent.DroneBasedReforestation.RechargeEnergyCount.sum": { "value": 773.0, "min": 115.0, "max": 1452.0, "count": 200 }, "Agent.DroneBasedReforestation.SaveLocationCount.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.SaveLocationCount.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.DroneBasedReforestation.OutofEnergyCount.mean": { "value": 0.013333333333333334, "min": 0.007751937984496124, "max": 0.4444444444444444, "count": 200 }, "Agent.DroneBasedReforestation.OutofEnergyCount.sum": { "value": 1.0, "min": 1.0, "max": 11.0, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.mean": { "value": 60.95712257385254, "min": 14.001348132178897, "max": 69.74257485071819, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceUntilTreeDrop.sum": { "value": 4571.78419303894, "min": 294.02831077575684, "max": 6425.419198989868, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeTreeDropReward.mean": { "value": 4.051989175875982, "min": 0.6367781010490877, "max": 7.112424261737288, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeTreeDropReward.sum": { "value": 303.8991881906986, "min": 17.193008728325367, "max": 693.824112713337, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceReward.mean": { "value": 1.7606877692540486, "min": 0.18164153695106505, "max": 1.9754605204970748, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistanceReward.sum": { "value": 132.05158269405365, "min": 4.625169456005096, "max": 183.901890873909, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.mean": { "value": 0.17606877515713373, "min": 0.018164153645435967, "max": 0.19754605216008644, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeNormalizedDistanceUntilTreeDrop.sum": { "value": 13.20515813678503, "min": 0.4625169411301613, "max": 18.390189103782177, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.mean": { "value": 65.71742180029551, "min": 22.95957660675049, "max": 75.41852946030467, "count": 200 }, "Agent.DroneBasedReforestation.CumulativeDistancetoExistingTrees.sum": { "value": 4928.806635022163, "min": 413.2723789215088, "max": 8868.855376243591, "count": 200 }, "Agent.Environment.LessonNumber.difficulty.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.difficulty.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.task.mean": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.LessonNumber.task.sum": { "value": 0.0, "min": 0.0, "max": 0.0, "count": 200 }, "Agent.Environment.EpisodeLength.mean": { "value": 76.12, "min": 45.325, "max": 335.0, "count": 200 }, "Agent.Environment.EpisodeLength.sum": { "value": 5709.0, "min": 4875.0, "max": 7080.0, "count": 200 }, "Agent.Step.mean": { "value": 1199974.0, "min": 5673.0, "max": 1199974.0, "count": 200 }, "Agent.Step.sum": { "value": 1199974.0, "min": 5673.0, "max": 1199974.0, "count": 200 }, "Agent.Policy.CuriosityValueEstimate.mean": { "value": 0.2106102555990219, "min": 0.013147092424333096, "max": 1.0176823139190674, "count": 200 }, "Agent.Policy.CuriosityValueEstimate.sum": { "value": 15.795768737792969, "min": 0.3944127857685089, "max": 37.65424346923828, "count": 200 }, "Agent.Policy.ExtrinsicValueEstimate.mean": { "value": 49.56050109863281, "min": 0.05967150628566742, "max": 55.748924255371094, "count": 200 }, "Agent.Policy.ExtrinsicValueEstimate.sum": { "value": 3717.03759765625, "min": 1.4917876720428467, "max": 6791.39453125, "count": 200 }, "Agent.Environment.CumulativeReward.mean": { "value": 105.32323019901911, "min": 42.939094327978395, "max": 108.22267144726169, "count": 200 }, "Agent.Environment.CumulativeReward.sum": { "value": 7899.242264926434, "min": 815.8427922315896, "max": 14075.67839050293, "count": 200 }, "Agent.Policy.CuriosityReward.mean": { "value": 0.16023189174632232, "min": 0.0, "max": 6.216257442127574, "count": 200 }, "Agent.Policy.CuriosityReward.sum": { "value": 12.017391880974174, "min": 0.0, "max": 205.13649559020996, "count": 200 }, "Agent.Policy.ExtrinsicReward.mean": { "value": 94.79090963840484, "min": 38.64518851414323, "max": 97.40040655289927, "count": 200 }, "Agent.Policy.ExtrinsicReward.sum": { "value": 7109.3182228803635, "min": 734.2585817687213, "max": 12668.110847473145, "count": 200 }, "Agent.IsTraining.mean": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 200 }, "Agent.IsTraining.sum": { "value": 1.0, "min": 1.0, "max": 1.0, "count": 200 }, "Agent.Losses.PolicyLoss.mean": { "value": 0.019896318728569895, "min": 0.014201972305912662, "max": 0.03352782945148647, "count": 138 }, "Agent.Losses.PolicyLoss.sum": { "value": 0.019896318728569895, "min": 0.014201972305912662, "max": 0.03352782945148647, "count": 138 }, "Agent.Losses.ValueLoss.mean": { "value": 79.89770062764485, "min": 24.139037211736042, "max": 153.07112185160318, "count": 138 }, "Agent.Losses.ValueLoss.sum": { "value": 79.89770062764485, "min": 24.139037211736042, "max": 153.07112185160318, "count": 138 }, "Agent.Policy.LearningRate.mean": { "value": 2.0678493107499904e-06, "min": 2.0678493107499904e-06, "max": 0.00029776350074549996, "count": 138 }, "Agent.Policy.LearningRate.sum": { "value": 2.0678493107499904e-06, "min": 2.0678493107499904e-06, "max": 0.00029776350074549996, "count": 138 }, "Agent.Policy.Epsilon.mean": { "value": 0.10068925000000001, "min": 0.10068925000000001, "max": 0.1992545, "count": 138 }, "Agent.Policy.Epsilon.sum": { "value": 0.10068925000000001, "min": 0.10068925000000001, "max": 0.1992545, "count": 138 }, "Agent.Policy.Beta.mean": { "value": 4.4393574999999844e-05, "min": 4.4393574999999844e-05, "max": 0.00496279955, "count": 138 }, "Agent.Policy.Beta.sum": { "value": 4.4393574999999844e-05, "min": 4.4393574999999844e-05, "max": 0.00496279955, "count": 138 }, "Agent.Losses.CuriosityForwardLoss.mean": { "value": 0.021680166323979694, "min": 0.017716502382730443, "max": 0.6002316027879715, "count": 138 }, "Agent.Losses.CuriosityForwardLoss.sum": { "value": 0.021680166323979694, "min": 0.017716502382730443, "max": 0.6002316027879715, "count": 138 }, "Agent.Losses.CuriosityInverseLoss.mean": { "value": 2.3713037073612213, "min": 2.2995146910349527, "max": 3.353024572134018, "count": 138 }, "Agent.Losses.CuriosityInverseLoss.sum": { "value": 2.3713037073612213, "min": 2.2995146910349527, "max": 3.353024572134018, "count": 138 } }, "metadata": { "timer_format_version": "0.1.0", "start_time_seconds": "1717309787", "python_version": "3.9.18 (main, Sep 11 2023, 14:09:26) [MSC v.1916 64 bit (AMD64)]", "command_line_arguments": "C:\\Users\\pdsie\\anaconda3\\envs\\mlagents20\\Scripts\\mlagents-learn c:/users/pdsie/documents/hivex/src/hivex/training/baseline/ml_agents/configs/mlagents/tmp/train/DroneBasedReforestation_difficulty_6_task_3_run_id_1_train.yaml --run-id=DroneBasedReforestation/train/DroneBasedReforestation_difficulty_6_task_3_run_id_1_train --base-port 5007", "mlagents_version": "0.30.0", "mlagents_envs_version": "0.30.0", "communication_protocol_version": "1.5.0", "pytorch_version": "1.7.1+cu110", "numpy_version": "1.21.0", "end_time_seconds": "1717313768" }, "total": 3980.9708423, "count": 1, "self": 0.5520710000005238, "children": { "run_training.setup": { "total": 0.0553342, "count": 1, "self": 0.0553342 }, "TrainerController.start_learning": { "total": 3980.3634371, "count": 1, "self": 5.711303599787243, "children": { "TrainerController._reset_env": { "total": 2.0899473000000004, "count": 1, "self": 2.0899473000000004 }, "TrainerController.advance": { "total": 3972.388613200212, "count": 403359, "self": 5.34119149997241, "children": { "env_step": { "total": 3967.0474217002397, "count": 403359, "self": 1936.8579967001137, "children": { "SubprocessEnvManager._take_step": { "total": 2026.5863883000686, "count": 403359, "self": 11.043841099993188, "children": { "TorchPolicy.evaluate": { "total": 2015.5425472000754, "count": 400008, "self": 2015.5425472000754 } } }, "workers": { "total": 3.603036700057487, "count": 403359, "self": 0.0, "children": { "worker_root": { "total": 3973.1631220001045, "count": 403359, "is_parallel": true, "self": 2243.7691715001138, "children": { "steps_from_proto": { "total": 0.005795299999999948, "count": 1, "is_parallel": true, "self": 9.909999999968555e-05, "children": { "_process_maybe_compressed_observation": { "total": 0.005652900000000072, "count": 2, "is_parallel": true, "self": 3.310000000000812e-05, "children": { "_observation_to_np_array": { "total": 0.005619800000000064, "count": 3, "is_parallel": true, "self": 2.5699999999906353e-05, "children": { "process_pixels": { "total": 0.005594100000000157, "count": 3, "is_parallel": true, "self": 0.0002377000000002294, "children": { "image_decompress": { "total": 0.005356399999999928, "count": 3, "is_parallel": true, "self": 0.005356399999999928 } } } } } } }, "_process_rank_one_or_two_observation": { "total": 4.330000000019041e-05, "count": 2, "is_parallel": true, "self": 4.330000000019041e-05 } } }, "UnityEnvironment.step": { "total": 1729.3881551999907, "count": 403359, "is_parallel": true, "self": 17.73496050013364, "children": { "UnityEnvironment._generate_step_input": { "total": 20.708846699980814, "count": 403359, "is_parallel": true, "self": 20.708846699980814 }, "communicator.exchange": { "total": 1543.1375892999351, "count": 403359, "is_parallel": true, "self": 1543.1375892999351 }, "steps_from_proto": { "total": 147.80675869994116, "count": 403359, "is_parallel": true, "self": 28.920287400085414, "children": { "_process_maybe_compressed_observation": { "total": 106.67767319989967, "count": 806718, "is_parallel": true, "self": 8.187807899532814, "children": { "_observation_to_np_array": { "total": 98.48986530036686, "count": 1210599, "is_parallel": true, "self": 8.1511440003883, "children": { "process_pixels": { "total": 90.33872129997856, "count": 1210599, "is_parallel": true, "self": 41.692075799878126, "children": { "image_decompress": { "total": 48.64664550010043, "count": 1210599, "is_parallel": true, "self": 48.64664550010043 } } } } } } }, "_process_rank_one_or_two_observation": { "total": 12.208798099956063, "count": 806718, "is_parallel": true, "self": 12.208798099956063 } } } } } } } } } } } } }, "trainer_threads": { "total": 3.3600000278966036e-05, "count": 1, "self": 3.3600000278966036e-05, "children": { "thread_root": { "total": 0.0, "count": 0, "is_parallel": true, "self": 0.0, "children": { "trainer_advance": { "total": 3976.1845601000186, "count": 193662, "is_parallel": true, "self": 4.92975520000482, "children": { "process_trajectory": { "total": 3238.143357700014, "count": 193662, "is_parallel": true, "self": 3237.7439340000137, "children": { "RLTrainer._checkpoint": { "total": 0.399423699999943, "count": 2, "is_parallel": true, "self": 0.399423699999943 } } }, "_update_policy": { "total": 733.1114472, "count": 138, "is_parallel": true, "self": 490.4996294999991, "children": { "TorchPPOOptimizer.update": { "total": 242.61181770000093, "count": 3366, "is_parallel": true, "self": 242.61181770000093 } } } } } } } } }, "TrainerController._save_models": { "total": 0.17353940000020884, "count": 1, "self": 0.01145020000058139, "children": { "RLTrainer._checkpoint": { "total": 0.16208919999962745, "count": 1, "self": 0.16208919999962745 } } } } } } }