From 3bf3f1af439058438ef783207a8d60b136624957 Mon Sep 17 00:00:00 2001
From: =?UTF-8?q?Mario=20Sa=CC=88nger?= <mario.saenger@student.hu-berlin.de>
Date: Tue, 26 Jun 2018 13:57:58 +0200
Subject: [PATCH] Minor changes to seq2seq attention

---
 code_jurica/seq2seq_attention.py | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)

diff --git a/code_jurica/seq2seq_attention.py b/code_jurica/seq2seq_attention.py
index 4a87398..67c81e4 100644
--- a/code_jurica/seq2seq_attention.py
+++ b/code_jurica/seq2seq_attention.py
@@ -34,6 +34,7 @@ tf.set_random_seed(1234)
 # Don't pre-allocate memory; allocate as-needed
 config.gpu_options.allow_growth = True
 config.gpu_options.allocator_type = 'BFC'
+config.gpu_options.visible_device_list = "2"
 sess = tf.Session(graph=tf.get_default_graph(), config=config)
 K.set_session(sess)
 
@@ -55,7 +56,7 @@ with open('models/icd10_mappings.p', 'rb') as handle:
 callbacks_list = [
     EarlyStopping(
         monitor='val_loss',
-        patience=5,
+        patience=50,
         # min_delta=0.001
     ),
     ModelCheckpoint(
-- 
GitLab