Questions
Questions

CS-7643-O01, OAN, OSZ Quiz #4: Module 3

Multiple fill-in-the-blank

As defined in Attention is All You Need, what is the size of the cross-attention matrix between the encoder and decoder given the following English to Spanish translation: I am very handsome -> Soy muy guapo Please assume the following: d_k = d_q = 64 d_v = 32 Please ignore the <SOS> and <EOS> tokens. cross-attention means Attention(Q, K, V) NOTE: Please round to the nearest integer. [Fill in the blank] rows[Fill in the blank] columns

View Explanation

View Explanation

Verified Answer
Please login to view
Step-by-Step Analysis
To determine the size of the cross-attention matrix, we need to recall how cross-attention is defined in Transformer architectures. The cross-attention matrix is formed by Attention(Q, K, V), where Q comes from the decoder (the current target sequence being generated) and K and V come from the encoder (the source sequence). The dimensions of this mat......Login to view full explanation

Log in for full answers

We've collected over 50,000 authentic exam questions and detailed explanations from around the globe. Log in now and get instant access to the answers!

Similar Questions

More Practical Tools for Students Powered by AI Study Helper

Join us and instantly unlock extensive past papers & exclusive solutions to get a head start on your studies!