In 2017, the Google machine translation team proposed the Transformer in their paperAttention is All You Need. The Transformer consists of an encoder and a(n) --------. (Fill in the blank.)
Among image preprocessing techniques, gamma correction is a common non-linear brightness adjustment method. Which of the following statements are true about the application and features of gamma correction?
The natural language processing field usually uses distributed semantic representation to represent words. Each word is no longer a completely orthogonal 0-1 vector, but a point in a multi-dimensional real number space, which is specifically represented as a real number vector.
Vision transformer (ViT) performs well in image classification tasks. Which of the following is the main advantage of ViT?
The objective of -------- is to extract and classify named entities in a text into pre-defined classes such as names, organizations, locations, time expressions, monetary values, and percentages. (Enter the abbreviation.)
Which of the following statements about the functions of the encoder and decoder is true?
Which of the following methods are useful when tackling overfitting?
Which of the following has never been used as a method in the history of NLP?
In the image recognition algorithm, the structure design of the convolutional layer has a great impact on its performance. Which of the following statements are true about the structure and mechanism of the convolutional layer? (Transposed convolution is not considered.)
Which of the following statements are true about the differences between using convolutional neural networks (CNNs) in text tasks and image tasks?
PDF + Testing Engine
|
---|
$66 |
Testing Engine
|
---|
$50 |
PDF (Q&A)
|
---|
$42 |
Huawei Free Exams |
---|
![]() |