Bi-directional attention, a characteristic of encoder models, is the result of applying self-attention in both forward and backward directions, enabling each word to be influenced by all other words in the sentence.
Bi-directional attention, a characteristic of encoder models, is the result of applying self-attention in both forward and backward directions, enabling each word to be influenced by all other words in the sentence.