Course Assignments
Recurrent Neural Networks:Sequence Models(Deep Learning Specialization)Answeres:2025
Question 1 Which expression refers to the s-th word in the r-th training example? ❌ x^(s)<r> ❌ x<r>(s) ✅ x(r)<s> ❌ x<s>(r) Explanation: The usual notation is x^{(r)}_{<s>}, i.e. the r-th training example (superscript or parentheses) and the s-th word within that example (angle-bracket subscript). Option x(r)<s> matches that ordering: example index first, then time/word… <a href="https://codeshala.io/platform/coursera/course/sequence-modelsdeep-learning-specialization/assignment/recurrent-neural-networkssequence-modelsdeep-learning-specializationansweres2025/" rel="bookmark"><span class="screen-reader-text">Recurrent Neural Networks:Sequence Models(Deep Learning Specialization)Answeres:2025</span></a>
Natural Language Processing & Word Embeddings:Sequence Models(Deep Learning Specialization)Answeres:2025
Question 1 True/False: Embedding vectors could be 60,000 dimensional to capture full variation. ✅ False ❌ True Explanation: While you could make embeddings as large as the vocabulary, that’s unnecessary and wasteful. Good embeddings are low-dimensional (e.g., 50–1,000) and capture semantics via learned structure — increasing dimension to the vocab size mostly increases parameters and… <a href="https://codeshala.io/platform/coursera/course/sequence-modelsdeep-learning-specialization/assignment/natural-language-processing-word-embeddingssequence-modelsdeep-learning-specializationansweres2025/" rel="bookmark"><span class="screen-reader-text">Natural Language Processing & Word Embeddings:Sequence Models(Deep Learning Specialization)Answeres:2025</span></a>
Sequence Models & Attention Mechanism:Sequence Models(Deep Learning Specialization)Answeres:2025
Question 1 Suppose this encoder–decoder model for MT.“This model is a ‘conditional language model’ in the sense that the encoder portion (green) is modeling the probability of the input sentence xxx.” True/False. ✅ False ❌ True Explanation: The encoder encodes the input sentence xxx into a representation (hidden states); it does not model P(x)P(x)P(x). The… <a href="https://codeshala.io/platform/coursera/course/sequence-modelsdeep-learning-specialization/assignment/sequence-models-attention-mechanismsequence-modelsdeep-learning-specializationansweres2025/" rel="bookmark"><span class="screen-reader-text">Sequence Models & Attention Mechanism:Sequence Models(Deep Learning Specialization)Answeres:2025</span></a>
Transformers:Sequence Models(Deep Learning Specialization)Answeres:2025
Question 1 A Transformer Network, like RNNs, GRUs, and LSTMs, can process information one word at a time (sequentially). ✅ False ❌ True Explanation:Transformers process all tokens in parallel using self-attention. Unlike RNNs, which are sequential by design, Transformers handle sequences non-sequentially, enabling faster and more efficient training. Question 2 Transformer Network methodology is taken… <a href="https://codeshala.io/platform/coursera/course/sequence-modelsdeep-learning-specialization/assignment/transformerssequence-modelsdeep-learning-specializationansweres2025/" rel="bookmark"><span class="screen-reader-text">Transformers:Sequence Models(Deep Learning Specialization)Answeres:2025</span></a>