반응형
250x250
Notice
Recent Posts
Recent Comments
Link
일 | 월 | 화 | 수 | 목 | 금 | 토 |
---|---|---|---|---|---|---|
1 | 2 | 3 | 4 | 5 | ||
6 | 7 | 8 | 9 | 10 | 11 | 12 |
13 | 14 | 15 | 16 | 17 | 18 | 19 |
20 | 21 | 22 | 23 | 24 | 25 | 26 |
27 | 28 | 29 | 30 |
Tags
- blip-2
- evaluating object hallucination in large vision-language models
- 논문 리뷰
- vlm 환각이란
- 1차 미분 마스크
- polling-based object probing evaluation
- vlm 환각
- vlm hallucination
- 객체 검출
- 엔트로피란
- clip adapter
- dinov2: learning robust visual features without supervision
- clip
- object detection
- dinov2 논문 리뷰
- dinov2: learning robust visual features without supervision 논문 리뷰
- 원격 학습 안끊기게
- 딥러닝 엔트로피
- mobilenetv1
- dinov2: learning robust visual features without supervision 논문
- Object detection article
- vlm
- 이미지 필터링
- evaluating object hallucination in large vision-language models paper
- 딥러닝 목적함수
- 기계학습
- 논문 요약
- 에지 검출
- evaluating object hallucination in large vision-language models 논문
- vlm hallucination paper
Archives
- Today
- Total
목록vit 논문 리뷰 (1)
My Vision, Computer Vision

An Image is Worth 16x16 Words: Transformers for Image Recognition at ScaleWhile the Transformer architecture has become the de-facto standard for natural language processing tasks, its applications to computer vision remain limited. In vision, attention is either applied in conjunction with convolutional networks, or used to reparxiv.orgAbstractTransformer가 사실상 NLP 분야의 표준이 되었지만 Computer vision에 ..
Paper
2024. 8. 27. 19:38