[스노피 AI] Vision Transformer 쉽게 이해하기 - 6.Multi-Head Attention in Vision Transformers
Multi-Head Attention in Vision Transformers
1. Introduction to Multi-Head Attention
2. Concept of Multi-Head Attention
3. Multi-Head Attention Mechanism
4. Mathematical Formulation
5. Advantages of Multi-Head Attention
6. Multi-Head Attention in Vision Transformers
7. Example Calculation
2개의 attention head와 3개의 Token input sequence가 있는 예를 살펴보겠습니다.
7.1. Input Embeddings
7.2. Linear Projections
사진 설명을 입력하세요.
7.3. Attention Calculation
7.4. Concatenation and Final Projection
Concat(Attention1, Attention2)
Output=Concat(Attention1, Attention2)WO
8. Challenges and Considerations
9. Summary
댓글 영역