Skip to content
Snippets Groups Projects
Commit 31a4d1d1 authored by Mario Sänger's avatar Mario Sänger
Browse files

Add encoder-decoder illustration

parent 80e9f392
No related merge requests found
File added
File added
......@@ -134,9 +134,10 @@ The goal of the model is to reassemble the dictionary symptom name from the
certificate line.
For this we adopt the encoder-decoder architecture proposed in
\cite{sutskever_sequence_2014}. As encoder we utilize a forward LSTM model,
which takes the single words of a certificate line as inputs and scans the line
from left to right. Each token will be represented using pre-trained FastText
\cite{sutskever_sequence_2014}. Figure \ref{fig:encoder_decoder} illustrates the
architecture of the model. As encoder we utilize a forward LSTM model, which
takes the single words of a certificate line as inputs and scans the line from
left to right. Each token will be represented using pre-trained FastText
word embeddings. Word embedding models represent words using a real-valued
vector and caputure syntactic and semantic similiarities between them. FastText
embeddings take sub-word information into account during training whereby the
......@@ -168,15 +169,19 @@ To enable our model to attend to different parts of a disease name we add an
extra attention layer \cite{raffel_feed-forward_2015} to the model. We train the
model using the provided ICD-10 dictionaries from all three languages.
During development we also experimented with character-level RNNs, but
couldn't achieve any approvements.
During development we also experimented with character-level RNNs for better
ICD-10 classification, however couldn't achieve any performance approvements.
\begin{figure}
\includegraphics[width=\textwidth]{Input.pdf}
\caption{A figure caption is always placed below the illustration.
Please note that short captions are centered, while long ones are
justified by the macro package automatically.}
\label{fig1}
\includegraphics[width=\textwidth,trim={0 17cm 0 3cm},clip=true]{encoder-decoder-model.pdf}
\caption{Illustration of the neural encoder-decoder model for symptom
extraction. The encoder processes a death certificate line token-wise from left
to right. The final state of the encoder forms a semantic representation of the
line and serves as initial input for the decoding process. The decoder will be
trained to predict the symptom name word by word. All input tokens will be
represented using the concatenation of the FastText embeddings of all three
languages.}
\label{fig:encoder_decoder}
\end{figure}
\section{Experiments and Results}
......
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment