Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Llava] Add Llava to transformers #27662

Merged
merged 166 commits into from
Dec 7, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
166 commits
Select commit Hold shift + click to select a range
d95b4c2
add model like
younesbelkada Nov 22, 2023
add5ed6
logits match
younesbelkada Nov 22, 2023
066a60e
minor fixes
younesbelkada Nov 22, 2023
e074c36
fixes
younesbelkada Nov 22, 2023
ef1b497
up
younesbelkada Nov 22, 2023
ebf33ce
up
younesbelkada Nov 22, 2023
0454cc6
add todo
younesbelkada Nov 22, 2023
9ed7c5b
llava processor
younesbelkada Nov 22, 2023
0ae0a26
keep the processor simple
younesbelkada Nov 22, 2023
56afb58
add conversion script
younesbelkada Nov 22, 2023
2630f05
fixup
younesbelkada Nov 22, 2023
25a4e30
fix copies
younesbelkada Nov 22, 2023
3b51d6f
up
younesbelkada Nov 22, 2023
3c0f1a1
add to index
younesbelkada Nov 22, 2023
9edca36
Merge remote-tracking branch 'upstream/main' into add-llava-final
younesbelkada Nov 24, 2023
af230f3
fix config + logits
younesbelkada Nov 24, 2023
eab82be
fix
younesbelkada Nov 24, 2023
d1d1dce
refactor
younesbelkada Nov 24, 2023
7858485
more refactor
younesbelkada Nov 24, 2023
c271132
more refactor
younesbelkada Nov 24, 2023
b699eda
fix copies
younesbelkada Nov 24, 2023
44a9681
add authors
younesbelkada Nov 24, 2023
eb6f87c
v1 tests
younesbelkada Nov 24, 2023
b29b635
add `LlavaProcessor` in init
younesbelkada Nov 24, 2023
c4a750a
remove unneeded import
younesbelkada Nov 24, 2023
43c2024
up
younesbelkada Nov 24, 2023
d1579bc
up
younesbelkada Nov 24, 2023
db0c237
docs
younesbelkada Nov 24, 2023
e53272d
up
younesbelkada Nov 24, 2023
ec522f4
fix CI
younesbelkada Nov 24, 2023
c59aa18
fix CI
younesbelkada Nov 24, 2023
3a31a92
add attention mask in test
younesbelkada Nov 24, 2023
ff3ae4e
make fixup
younesbelkada Nov 24, 2023
84f6fa6
remove the vision model
ArthurZucker Nov 28, 2023
bbb1a6a
that' s the dirty way to do it
ArthurZucker Nov 28, 2023
550a124
Merge branch 'main' of github.com:huggingface/transformers into add-l…
ArthurZucker Nov 28, 2023
191ecf6
nits
ArthurZucker Nov 28, 2023
2255a86
nits
ArthurZucker Nov 28, 2023
6928570
updates
ArthurZucker Nov 30, 2023
f87f27e
add more tests
ArthurZucker Nov 30, 2023
83320b0
add input tests
ArthurZucker Nov 30, 2023
550a17f
fixup
ArthurZucker Nov 30, 2023
2a358e1
more styling
ArthurZucker Nov 30, 2023
67c1759
nits
ArthurZucker Nov 30, 2023
c436779
updates amd cleanup
ArthurZucker Nov 30, 2023
e67ece8
fixup the generation expected results
ArthurZucker Nov 30, 2023
dc34e1b
fix the testing script
ArthurZucker Nov 30, 2023
587b8e6
some cleanup and simplification which does not work yet but almost th…
ArthurZucker Nov 30, 2023
33b7dc8
make correct dispatch operations
ArthurZucker Dec 3, 2023
ebec096
vectorize works for batch of images and text
ArthurZucker Dec 3, 2023
a2e5b22
last todos
ArthurZucker Dec 3, 2023
5cd83cd
nits
ArthurZucker Dec 4, 2023
46ccfcd
update test and modeling code
ArthurZucker Dec 4, 2023
a305e21
remove useless function for now
ArthurZucker Dec 4, 2023
310984f
fix few issues
younesbelkada Dec 4, 2023
4aa44fc
fix generation
younesbelkada Dec 4, 2023
c8c3906
some nits
ArthurZucker Dec 4, 2023
b988898
add bakllava
younesbelkada Dec 4, 2023
3fb71ff
nits
ArthurZucker Dec 4, 2023
db6c309
Merge branch 'add-llava-final' of https://github.com/younesbelkada/tr…
ArthurZucker Dec 4, 2023
66f1b5d
remove duplicated code
ArthurZucker Dec 4, 2023
480441c
finis merge
ArthurZucker Dec 4, 2023
9f0a0bb
cleanup
ArthurZucker Dec 4, 2023
7c09eb4
missed this line
ArthurZucker Dec 4, 2023
5af1886
fill the todos
ArthurZucker Dec 4, 2023
a1ff5ce
add left padding offset
ArthurZucker Dec 4, 2023
0db1d75
add left and rignt padding logic
ArthurZucker Dec 4, 2023
a8b64d0
bool to properly index
ArthurZucker Dec 4, 2023
6df6f2e
make sure
ArthurZucker Dec 4, 2023
ab9bac9
more cleanups
ArthurZucker Dec 4, 2023
60051e4
batch is fixed :wink:
ArthurZucker Dec 5, 2023
e331d3b
add correct device for tensor creation
ArthurZucker Dec 5, 2023
9ac4ecf
fix some dtype missmatch
ArthurZucker Dec 5, 2023
86bd66c
ruff
ArthurZucker Dec 5, 2023
a28ec06
update conversion script
younesbelkada Dec 5, 2023
9336ca6
Update src/transformers/__init__.py
younesbelkada Dec 5, 2023
dc33cb4
fa 2 support + fix conversion script
younesbelkada Dec 5, 2023
bed1782
more
younesbelkada Dec 5, 2023
8c9d44e
correct reshaping
ArthurZucker Dec 5, 2023
9b09b62
fix test dict
ArthurZucker Dec 5, 2023
89d8834
fix copies by ignoring
ArthurZucker Dec 5, 2023
245f07f
fix nit
younesbelkada Dec 5, 2023
306b0fd
skip clip vision model
ArthurZucker Dec 5, 2023
d39d732
Merge branch 'add-llava-final' of https://github.com/younesbelkada/tr…
ArthurZucker Dec 5, 2023
bba8cff
fixup
ArthurZucker Dec 5, 2023
14b1628
fixup
ArthurZucker Dec 5, 2023
24ccf04
LlavaForVisionText2Text -> LlavaForCausalLM
ArthurZucker Dec 5, 2023
f9ee924
update
ArthurZucker Dec 5, 2023
b1386ee
fix
younesbelkada Dec 5, 2023
72275fe
Merge branch 'add-llava-final' of https://github.com/younesbelkada/tr…
younesbelkada Dec 5, 2023
9276212
raise correct errors
ArthurZucker Dec 5, 2023
0876a96
Merge branch 'add-llava-final' of https://github.com/younesbelkada/tr…
ArthurZucker Dec 5, 2023
a5ce98f
fix
younesbelkada Dec 5, 2023
6bde4bd
Merge branch 'add-llava-final' of https://github.com/younesbelkada/tr…
younesbelkada Dec 5, 2023
3fe11d4
docs
younesbelkada Dec 5, 2023
8baeeb5
nuke for now
ArthurZucker Dec 5, 2023
ce2f1e5
Merge branch 'add-llava-final' of https://github.com/younesbelkada/tr…
ArthurZucker Dec 5, 2023
96569c7
nits here and there
ArthurZucker Dec 5, 2023
fcfb2c3
fixup
ArthurZucker Dec 5, 2023
5aeca7d
fix remaining tests
ArthurZucker Dec 5, 2023
433cf04
update LlavaForConditionalGeneration instead of CausalLM
ArthurZucker Dec 5, 2023
0e35a44
fixups
ArthurZucker Dec 5, 2023
e098a5a
pipeline support
ArthurZucker Dec 5, 2023
ef1a1a2
slow and piepline tests
ArthurZucker Dec 5, 2023
cdf9bbe
supports batch
ArthurZucker Dec 5, 2023
3769e91
nits
ArthurZucker Dec 5, 2023
9ca289e
Merge branch 'main' of github.com:huggingface/transformers into add-l…
ArthurZucker Dec 5, 2023
c205929
cleanup
ArthurZucker Dec 5, 2023
a87e7d1
fix first integration tests
ArthurZucker Dec 5, 2023
f4f8756
add pad token where needed
ArthurZucker Dec 5, 2023
37fe79a
correct etsts
ArthurZucker Dec 5, 2023
32c3b32
fixups
ArthurZucker Dec 5, 2023
3bbca52
update pipeline testr
ArthurZucker Dec 5, 2023
1099f3b
fix quality
younesbelkada Dec 5, 2023
7046d84
nits
ArthurZucker Dec 5, 2023
bb34021
Merge branch 'add-llava-final' of https://github.com/younesbelkada/tr…
ArthurZucker Dec 5, 2023
85aa150
revert unneeded change
younesbelkada Dec 5, 2023
9a9bb50
nit
younesbelkada Dec 5, 2023
a0f16ed
Merge branch 'add-llava-final' of https://github.com/younesbelkada/tr…
younesbelkada Dec 5, 2023
4fc5dab
use BatchFeature
ArthurZucker Dec 5, 2023
136b737
Merge branch 'add-llava-final' of https://github.com/younesbelkada/tr…
ArthurZucker Dec 5, 2023
e0d9e70
from ...feature_extraction_utils import BatchFeature
ArthurZucker Dec 5, 2023
c27ef69
nits
ArthurZucker Dec 5, 2023
b065941
nits
ArthurZucker Dec 5, 2023
967a1e9
properly update
ArthurZucker Dec 5, 2023
076b7fc
more f*** nits
ArthurZucker Dec 5, 2023
6d1c94a
fix copies
ArthurZucker Dec 5, 2023
c5a76b7
comment
ArthurZucker Dec 5, 2023
8eae5ea
keep slow test slow
ArthurZucker Dec 5, 2023
d677e64
Update src/transformers/models/llava/processing_llava.py
younesbelkada Dec 5, 2023
4eb1729
add piepline example
younesbelkada Dec 5, 2023
ebe5d12
add pixel values in docstrign
younesbelkada Dec 5, 2023
0ab7a79
update pr doctest
younesbelkada Dec 5, 2023
5f07c95
fix
younesbelkada Dec 5, 2023
8c92e84
fix slow tests
younesbelkada Dec 5, 2023
cf062cc
remove hack
younesbelkada Dec 5, 2023
6e75d47
fixup
younesbelkada Dec 5, 2023
58d9ef4
small note
younesbelkada Dec 5, 2023
d3dfc08
forward contrib credits from PR25789
shauray8 Dec 5, 2023
a4a47b0
forward contrib credits from original implementation and work
haotian-liu Dec 5, 2023
00c8559
add arthur
younesbelkada Dec 5, 2023
62006e9
Update src/transformers/models/llava/processing_llava.py
younesbelkada Dec 5, 2023
c254603
update docstring
younesbelkada Dec 5, 2023
b11c307
nit
younesbelkada Dec 5, 2023
9d50368
move to not doctested because of timeout issues
younesbelkada Dec 5, 2023
0a21c52
fixup
younesbelkada Dec 5, 2023
4457339
add description
younesbelkada Dec 5, 2023
82645e7
more
younesbelkada Dec 5, 2023
9e503a5
fix-copies
younesbelkada Dec 5, 2023
286b27b
Merge branch 'main' into add-llava-final
younesbelkada Dec 5, 2023
1779ce2
fix docs
younesbelkada Dec 5, 2023
6c1e63a
Merge branch 'add-llava-final' of https://github.com/younesbelkada/tr…
younesbelkada Dec 5, 2023
4eb55a5
add beam search
younesbelkada Dec 5, 2023
f0b3610
Merge branch 'add-llava-final' of https://github.com/younesbelkada/tr…
younesbelkada Dec 5, 2023
c49f9bd
add more comments
younesbelkada Dec 6, 2023
348e7b6
add typehints on processor
younesbelkada Dec 6, 2023
e3732a9
add speedup plot
younesbelkada Dec 6, 2023
8a301e4
update slow tests and docs
younesbelkada Dec 6, 2023
8ad85b3
push test
younesbelkada Dec 6, 2023
57a816f
push batched test
younesbelkada Dec 6, 2023
661ff8d
fix batched generation with different number of images
younesbelkada Dec 6, 2023
d809222
remove benchmark due to a bug
younesbelkada Dec 6, 2023
0f7629a
fix test
younesbelkada Dec 6, 2023
6678803
Merge branch 'main' into add-llava-final
younesbelkada Dec 7, 2023
3c39ba1
fix copies
younesbelkada Dec 7, 2023
a7cc780
add gcolab demo
younesbelkada Dec 7, 2023
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -397,6 +397,7 @@ Current number of checkpoints: ![](https://img.shields.io/endpoint?url=https://h
1. **[LiLT](https://huggingface.co/docs/transformers/model_doc/lilt)** (from South China University of Technology) released with the paper [LiLT: A Simple yet Effective Language-Independent Layout Transformer for Structured Document Understanding](https://arxiv.org/abs/2202.13669) by Jiapeng Wang, Lianwen Jin, Kai Ding.
1. **[LLaMA](https://huggingface.co/docs/transformers/model_doc/llama)** (from The FAIR team of Meta AI) released with the paper [LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971) by Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand Joulin, Edouard Grave, Guillaume Lample.
1. **[Llama2](https://huggingface.co/docs/transformers/model_doc/llama2)** (from The FAIR team of Meta AI) released with the paper [Llama2: Open Foundation and Fine-Tuned Chat Models](https://ai.meta.com/research/publications/llama-2-open-foundation-and-fine-tuned-chat-models/) by Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, Cristian Canton Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez Madian Khabsa, Isabel Kloumann, Artem Korenev, Punit Singh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushka rMishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, Eric Michael Smith, Ranjan Subramanian, Xiaoqing EllenTan, Binh Tang, Ross Taylor, Adina Williams, Jian Xiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, Thomas Scialom.
1. **[Llava](https://huggingface.co/docs/transformers/main/model_doc/llava)** (from Microsoft Research & University of Wisconsin-Madison) released with the paper [Improved Baselines with Visual Instruction Tuning](https://arxiv.org/pdf/2310.03744) by Haotian Liu, Chunyuan Li, Yuheng Li and Yong Jae Lee.
1. **[Longformer](https://huggingface.co/docs/transformers/model_doc/longformer)** (from AllenAI) released with the paper [Longformer: The Long-Document Transformer](https://arxiv.org/abs/2004.05150) by Iz Beltagy, Matthew E. Peters, Arman Cohan.
1. **[LongT5](https://huggingface.co/docs/transformers/model_doc/longt5)** (from Google AI) released with the paper [LongT5: Efficient Text-To-Text Transformer for Long Sequences](https://arxiv.org/abs/2112.07916) by Mandy Guo, Joshua Ainslie, David Uthus, Santiago Ontanon, Jianmo Ni, Yun-Hsuan Sung, Yinfei Yang.
1. **[LUKE](https://huggingface.co/docs/transformers/model_doc/luke)** (from Studio Ousia) released with the paper [LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention](https://arxiv.org/abs/2010.01057) by Ikuya Yamada, Akari Asai, Hiroyuki Shindo, Hideaki Takeda, Yuji Matsumoto.
Expand Down
1 change: 1 addition & 0 deletions README_es.md
Original file line number Diff line number Diff line change
Expand Up @@ -372,6 +372,7 @@ Número actual de puntos de control: ![](https://img.shields.io/endpoint?url=htt
1. **[LiLT](https://huggingface.co/docs/transformers/model_doc/lilt)** (from South China University of Technology) released with the paper [LiLT: A Simple yet Effective Language-Independent Layout Transformer for Structured Document Understanding](https://arxiv.org/abs/2202.13669) by Jiapeng Wang, Lianwen Jin, Kai Ding.
1. **[LLaMA](https://huggingface.co/docs/transformers/model_doc/llama)** (from The FAIR team of Meta AI) released with the paper [LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971) by Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand Joulin, Edouard Grave, Guillaume Lample.
1. **[Llama2](https://huggingface.co/docs/transformers/model_doc/llama2)** (from The FAIR team of Meta AI) released with the paper [Llama2: Open Foundation and Fine-Tuned Chat Models](https://ai.meta.com/research/publications/llama-2-open-foundation-and-fine-tuned-chat-models/XXX) by Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, Cristian Canton Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez Madian Khabsa, Isabel Kloumann, Artem Korenev, Punit Singh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushka rMishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, Eric Michael Smith, Ranjan Subramanian, Xiaoqing EllenTan, Binh Tang, Ross Taylor, Adina Williams, Jian Xiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, Thomas Scialom..
1. **[Llava](https://huggingface.co/docs/transformers/main/model_doc/llava)** (from Microsoft Research & University of Wisconsin-Madison) released with the paper [Improved Baselines with Visual Instruction Tuning](https://arxiv.org/pdf/2310.03744) by Haotian Liu, Chunyuan Li, Yuheng Li and Yong Jae Lee.
1. **[Longformer](https://huggingface.co/docs/transformers/model_doc/longformer)** (from AllenAI) released with the paper [Longformer: The Long-Document Transformer](https://arxiv.org/abs/2004.05150) by Iz Beltagy, Matthew E. Peters, Arman Cohan.
1. **[LongT5](https://huggingface.co/docs/transformers/model_doc/longt5)** (from Google AI) released with the paper [LongT5: Efficient Text-To-Text Transformer for Long Sequences](https://arxiv.org/abs/2112.07916) by Mandy Guo, Joshua Ainslie, David Uthus, Santiago Ontanon, Jianmo Ni, Yun-Hsuan Sung, Yinfei Yang.
1. **[LUKE](https://huggingface.co/docs/transformers/model_doc/luke)** (from Studio Ousia) released with the paper [LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention](https://arxiv.org/abs/2010.01057) by Ikuya Yamada, Akari Asai, Hiroyuki Shindo, Hideaki Takeda, Yuji Matsumoto.
Expand Down
1 change: 1 addition & 0 deletions README_hd.md
Original file line number Diff line number Diff line change
Expand Up @@ -346,6 +346,7 @@ conda install -c huggingface transformers
1. **[LiLT](https://huggingface.co/docs/transformers/model_doc/lilt)** (दक्षिण चीन प्रौद्योगिकी विश्वविद्यालय से) साथ में कागज [LiLT: एक सरल लेकिन प्रभावी भाषा-स्वतंत्र लेआउट ट्रांसफार्मर संरचित दस्तावेज़ समझ के लिए](https://arxiv.org/abs/2202.13669) जियापेंग वांग, लियानवेन जिन, काई डिंग द्वारा पोस्ट किया गया।
1. **[LLaMA](https://huggingface.co/docs/transformers/model_doc/llama)** (The FAIR team of Meta AI से) Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand Joulin, Edouard Grave, Guillaume Lample. द्वाराअनुसंधान पत्र [LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971) के साथ जारी किया गया
1. **[Llama2](https://huggingface.co/docs/transformers/model_doc/llama2)** (The FAIR team of Meta AI से) Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, Cristian Canton Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez Madian Khabsa, Isabel Kloumann, Artem Korenev, Punit Singh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushka rMishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, Eric Michael Smith, Ranjan Subramanian, Xiaoqing EllenTan, Binh Tang, Ross Taylor, Adina Williams, Jian Xiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, Thomas Scialom.. द्वाराअनुसंधान पत्र [Llama2: Open Foundation and Fine-Tuned Chat Models](https://ai.meta.com/research/publications/llama-2-open-foundation-and-fine-tuned-chat-models/XXX) के साथ जारी किया गया
1. **[Llava](https://huggingface.co/docs/transformers/main/model_doc/llava)** (Microsoft Research & University of Wisconsin-Madison से) Haotian Liu, Chunyuan Li, Yuheng Li and Yong Jae Lee. द्वाराअनुसंधान पत्र [Improved Baselines with Visual Instruction Tuning](https://arxiv.org/pdf/2310.03744) के साथ जारी किया गया
1. **[Longformer](https://huggingface.co/docs/transformers/model_doc/longformer)** (from AllenAI) released with the paper [Longformer: The Long-Document Transformer](https://arxiv.org/abs/2004.05150) by Iz Beltagy, Matthew E. Peters, Arman Cohan.
1. **[LongT5](https://huggingface.co/docs/transformers/model_doc/longt5)** (मैंडी गुओ, जोशुआ आइंस्ली, डेविड यूथस, सैंटियागो ओंटानन, जियानमो नि, यूं-हुआन सुंग, यिनफेई यांग द्वारा पोस्ट किया गया।
1. **[LUKE](https://huggingface.co/docs/transformers/model_doc/luke)** (स्टूडियो औसिया से) साथ में पेपर [LUKE: डीप कॉन्टेक्स्टुअलाइज्ड एंटिटी रिप्रेजेंटेशन विद एंटिटी-अवेयर सेल्फ-अटेंशन](https ://arxiv.org/abs/2010.01057) Ikuya Yamada, Akari Asai, Hiroyuki Shindo, Hideaki Takeda, Yuji Matsumoto द्वारा।
Expand Down
1 change: 1 addition & 0 deletions README_ja.md
Original file line number Diff line number Diff line change
Expand Up @@ -406,6 +406,7 @@ Flax、PyTorch、TensorFlowをcondaでインストールする方法は、それ
1. **[LiLT](https://huggingface.co/docs/transformers/model_doc/lilt)** (South China University of Technology から) Jiapeng Wang, Lianwen Jin, Kai Ding から公開された研究論文: [LiLT: A Simple yet Effective Language-Independent Layout Transformer for Structured Document Understanding](https://arxiv.org/abs/2202.13669)
1. **[LLaMA](https://huggingface.co/docs/transformers/model_doc/llama)** (The FAIR team of Meta AI から) Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand Joulin, Edouard Grave, Guillaume Lample. から公開された研究論文 [LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)
1. **[Llama2](https://huggingface.co/docs/transformers/model_doc/llama2)** (The FAIR team of Meta AI から) Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, Cristian Canton Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez Madian Khabsa, Isabel Kloumann, Artem Korenev, Punit Singh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushka rMishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, Eric Michael Smith, Ranjan Subramanian, Xiaoqing EllenTan, Binh Tang, Ross Taylor, Adina Williams, Jian Xiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, Thomas Scialom.. から公開された研究論文 [Llama2: Open Foundation and Fine-Tuned Chat Models](https://ai.meta.com/research/publications/llama-2-open-foundation-and-fine-tuned-chat-models/XXX)
1. **[Llava](https://huggingface.co/docs/transformers/main/model_doc/llava)** (Microsoft Research & University of Wisconsin-Madison から) Haotian Liu, Chunyuan Li, Yuheng Li and Yong Jae Lee. から公開された研究論文 [Improved Baselines with Visual Instruction Tuning](https://arxiv.org/pdf/2310.03744)
1. **[Longformer](https://huggingface.co/docs/transformers/model_doc/longformer)** (AllenAI から) Iz Beltagy, Matthew E. Peters, Arman Cohan から公開された研究論文: [Longformer: The Long-Document Transformer](https://arxiv.org/abs/2004.05150)
1. **[LongT5](https://huggingface.co/docs/transformers/model_doc/longt5)** (Google AI から) Mandy Guo, Joshua Ainslie, David Uthus, Santiago Ontanon, Jianmo Ni, Yun-Hsuan Sung, Yinfei Yang から公開された研究論文: [LongT5: Efficient Text-To-Text Transformer for Long Sequences](https://arxiv.org/abs/2112.07916)
1. **[LUKE](https://huggingface.co/docs/transformers/model_doc/luke)** (Studio Ousia から) Ikuya Yamada, Akari Asai, Hiroyuki Shindo, Hideaki Takeda, Yuji Matsumoto から公開された研究論文: [LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention](https://arxiv.org/abs/2010.01057)
Expand Down
1 change: 1 addition & 0 deletions README_ko.md
Original file line number Diff line number Diff line change
Expand Up @@ -321,6 +321,7 @@ Flax, PyTorch, TensorFlow 설치 페이지에서 이들을 conda로 설치하는
1. **[LiLT](https://huggingface.co/docs/transformers/model_doc/lilt)** (South China University of Technology 에서) Jiapeng Wang, Lianwen Jin, Kai Ding 의 [LiLT: A Simple yet Effective Language-Independent Layout Transformer for Structured Document Understanding](https://arxiv.org/abs/2202.13669) 논문과 함께 발표했습니다.
1. **[LLaMA](https://huggingface.co/docs/transformers/model_doc/llama)** (The FAIR team of Meta AI 에서 제공)은 Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand Joulin, Edouard Grave, Guillaume Lample.의 [LLaMA: Open and Efficient Foundation Language Models](https://arxiv.org/abs/2302.13971)논문과 함께 발표했습니다.
1. **[Llama2](https://huggingface.co/docs/transformers/model_doc/llama2)** (The FAIR team of Meta AI 에서 제공)은 Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, Cristian Canton Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez Madian Khabsa, Isabel Kloumann, Artem Korenev, Punit Singh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushka rMishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, Eric Michael Smith, Ranjan Subramanian, Xiaoqing EllenTan, Binh Tang, Ross Taylor, Adina Williams, Jian Xiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, Thomas Scialom..의 [Llama2: Open Foundation and Fine-Tuned Chat Models](https://ai.meta.com/research/publications/llama-2-open-foundation-and-fine-tuned-chat-models/XXX)논문과 함께 발표했습니다.
1. **[Llava](https://huggingface.co/docs/transformers/main/model_doc/llava)** (Microsoft Research & University of Wisconsin-Madison 에서 제공)은 Haotian Liu, Chunyuan Li, Yuheng Li and Yong Jae Lee.의 [Improved Baselines with Visual Instruction Tuning](https://arxiv.org/pdf/2310.03744)논문과 함께 발표했습니다.
1. **[Longformer](https://huggingface.co/docs/transformers/model_doc/longformer)** (AllenAI 에서) Iz Beltagy, Matthew E. Peters, Arman Cohan 의 [Longformer: The Long-Document Transformer](https://arxiv.org/abs/2004.05150) 논문과 함께 발표했습니다.
1. **[LongT5](https://huggingface.co/docs/transformers/model_doc/longt5)** (Google AI 에서) Mandy Guo, Joshua Ainslie, David Uthus, Santiago Ontanon, Jianmo Ni, Yun-Hsuan Sung, Yinfei Yang 의 [LongT5: Efficient Text-To-Text Transformer for Long Sequences](https://arxiv.org/abs/2112.07916) 논문과 함께 발표했습니다.
1. **[LUKE](https://huggingface.co/docs/transformers/model_doc/luke)** (Studio Ousia 에서) Ikuya Yamada, Akari Asai, Hiroyuki Shindo, Hideaki Takeda, Yuji Matsumoto 의 [LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention](https://arxiv.org/abs/2010.01057) 논문과 함께 발표했습니다.
Expand Down
Loading
Loading