Questions
BU.520.710.T1.SP25 Quiz 4
Single choice
Attention Mechanism improves word embedding by:
Options
A.Replacing traditional word embeddings like Word2Vec and GloVe with fixed vector representations
B.Increasing the number of parameters in a neural network without improving contextual understanding
C.Ignoring the sequential nature of text and treating words as independent entities.
D.Dynamically adjusting word embeddings by weighting the relevance of surrounding words in context
View Explanation
Verified Answer
Please login to view
Step-by-Step Analysis
Examining the concept of attention mechanisms in word embedding requires comparing how each option describes the role of attention in contextual representations.
Option 1 argues that attention replaces traditional fixed embeddings like Word2Vec and GloVe with fixed vector representations. In reality, attention mechanisms do not replace fixed embeddings; they operate......Login to view full explanationLog in for full answers
We've collected over 50,000 authentic exam questions and detailed explanations from around the globe. Log in now and get instant access to the answers!
Similar Questions
On scaled dot-product attention and training stability of a transformer: I Without scaling by 𝐷 𝑘 , the variance of the dot product 𝑞 𝑛 ⊤ 𝑘 𝑚 grows with dimensionality, producing large logits that can saturate the softmax. II Scaling by 𝐷 𝑘 primarily solves exploding-gradient problems inside the value projection 𝑉 . III The softmax-normalized matrix S o f t m a x ( 𝑄 𝐾 ⊤ ) is applied row-wise, making each row represent how strongly a query attends to all keys. IV Scaled dot-product attention computes A t t e n t i o n ( 𝑄 , 𝐾 , 𝑉 ) = S o f t m a x ! ( 𝑄 𝐾 ⊤ 𝐷 𝑘 ) 𝑉 , and the resulting matrix always has the same dimension as 𝑉 .
Which innovation is at the core of the transformer architecture and enables modeling long-range dependencies effectively?
As defined in Attention is All You Need, what is the size of the cross-attention matrix between the encoder and decoder given the following English to Spanish translation: I am very handsome -> Soy muy guapo Please assume the following: d_k = d_q = 64 d_v = 32 Please ignore the <SOS> and <EOS> tokens. cross-attention means Attention(Q, K, V) NOTE: Please round to the nearest integer. [Fill in the blank] rows[Fill in the blank] columns
Which of the following attention models uses a subset of the input to derive the output, and can not be trained directly with gradient methods?
More Practical Tools for Students Powered by AI Study Helper
Making Your Study Simpler
Join us and instantly unlock extensive past papers & exclusive solutions to get a head start on your studies!