From 3bf3f1af439058438ef783207a8d60b136624957 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Mario=20Sa=CC=88nger?= <mario.saenger@student.hu-berlin.de> Date: Tue, 26 Jun 2018 13:57:58 +0200 Subject: [PATCH] Minor changes to seq2seq attention --- code_jurica/seq2seq_attention.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/code_jurica/seq2seq_attention.py b/code_jurica/seq2seq_attention.py index 4a87398..67c81e4 100644 --- a/code_jurica/seq2seq_attention.py +++ b/code_jurica/seq2seq_attention.py @@ -34,6 +34,7 @@ tf.set_random_seed(1234) # Don't pre-allocate memory; allocate as-needed config.gpu_options.allow_growth = True config.gpu_options.allocator_type = 'BFC' +config.gpu_options.visible_device_list = "2" sess = tf.Session(graph=tf.get_default_graph(), config=config) K.set_session(sess) @@ -55,7 +56,7 @@ with open('models/icd10_mappings.p', 'rb') as handle: callbacks_list = [ EarlyStopping( monitor='val_loss', - patience=5, + patience=50, # min_delta=0.001 ), ModelCheckpoint( -- GitLab