Questions tagged [multihead-attention]
18 questions
0
votes
0 answers
How to add a multihead attention layer to a CNN-LSTM model?
I'm trying to make a hybrid binary text classification model using a multi-head attention mechanism with CNN-LSTM. However, I'm facing an issue when trying to pass the values obtained from CNN-LSTM to the attention layer.
This was what I…

Harsha Vardhan
- 11
- 2
0
votes
0 answers
Transformer-Model for hierarchical text
I have a NLP task/idea in mind, where the input and output text is structured purely hierarchically, like multi-level bulletpoint lists or a table of content.
The question is: Is there any resaetch for this particular type of text for transformer…

Patrick
- 1
- 3
-1
votes
0 answers
Calculating Sentence Level Attention
How do I quantify the attention between input and output sentences in a sequence-to-sequence language modelling scenario [translation or summarization]?
For instance, consider these input and output statements, i.e., document is the input, and…