@arankomatsuzaki
Retrieval meets Long Context Large Language Models Llama with 32K context using retrieval-augmentation at generation outperforms finetuned LLM with 32K context via positional interpolation, while taking much less computation https://t.co/VHWwpyagJD https://t.co/TvW8ZxwYmV