Questions tagged [multihead-attention]

18 questions
0
votes
0 answers

How to add a multihead attention layer to a CNN-LSTM model?

I'm trying to make a hybrid binary text classification model using a multi-head attention mechanism with CNN-LSTM. However, I'm facing an issue when trying to pass the values obtained from CNN-LSTM to the attention layer. This was what I…
0
votes
0 answers

Transformer-Model for hierarchical text

I have a NLP task/idea in mind, where the input and output text is structured purely hierarchically, like multi-level bulletpoint lists or a table of content. The question is: Is there any resaetch for this particular type of text for transformer…
-1
votes
0 answers

Calculating Sentence Level Attention

How do I quantify the attention between input and output sentences in a sequence-to-sequence language modelling scenario [translation or summarization]? For instance, consider these input and output statements, i.e., document is the input, and…
1
2