summaryrefslogtreecommitdiff
path: root/examples/pybullet/gym/pybullet_envs/minitaur/agents/scripts/configs.py
diff options
context:
space:
mode:
Diffstat (limited to 'examples/pybullet/gym/pybullet_envs/minitaur/agents/scripts/configs.py')
-rw-r--r--examples/pybullet/gym/pybullet_envs/minitaur/agents/scripts/configs.py6
1 files changed, 1 insertions, 5 deletions
diff --git a/examples/pybullet/gym/pybullet_envs/minitaur/agents/scripts/configs.py b/examples/pybullet/gym/pybullet_envs/minitaur/agents/scripts/configs.py
index 669f61827..1ea65948b 100644
--- a/examples/pybullet/gym/pybullet_envs/minitaur/agents/scripts/configs.py
+++ b/examples/pybullet/gym/pybullet_envs/minitaur/agents/scripts/configs.py
@@ -11,7 +11,6 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
-
"""Example configurations using the PPO algorithm."""
from __future__ import absolute_import
@@ -33,10 +32,7 @@ def default():
use_gpu = False
# Network
network = networks.ForwardGaussianPolicy
- weight_summaries = dict(
- all=r'.*',
- policy=r'.*/policy/.*',
- value=r'.*/value/.*')
+ weight_summaries = dict(all=r'.*', policy=r'.*/policy/.*', value=r'.*/value/.*')
policy_layers = 200, 100
value_layers = 200, 100
init_mean_factor = 0.05