Aptitude
Reasoning
Verbal
Technical
AI
Exams
Search
AI
/
Computer Vision
/
Exercises
/
Exercise 8
Computer Vision
hard
0/15 answered
Q1
What is the implicit positional encoding hypothesis for Vision Transformers?
A
Positional encodings are always required for good performance
B
Implicit encodings work only for square images
C
ViTs cannot learn any positional information without explicit encodings
D
ViTs may learn implicit positional information through the boundary patterns of patch embeddings rather than explicit positional encodings
← Prev
1 / 15
Next →
← Exercise 7
Exercise 8 of 8
All Exercises →