diff --git a/docs/_posts/bunyamin-polat/2023-05-29-fingen_flant5_finetuned_fiqa_en.md b/docs/_posts/bunyamin-polat/2023-05-29-fingen_flant5_finetuned_fiqa_en.md index 73fa1c1fcd..45eb17cb13 100644 --- a/docs/_posts/bunyamin-polat/2023-05-29-fingen_flant5_finetuned_fiqa_en.md +++ b/docs/_posts/bunyamin-polat/2023-05-29-fingen_flant5_finetuned_fiqa_en.md @@ -19,7 +19,29 @@ use_language_switcher: "Python-Scala-Java" ## Description -The `fingen_flant5_finetuned_fiqa` model is the Text Generation model that has been fine-tuned on FLAN-T5 using FIQA dataset. FLAN-T5 is a state-of-the-art language model developed by Facebook AI that utilizes the T5 architecture for text-generation tasks. +The `fingen_flant5_finetuned_fiqa` model is the Text Generation model that has been fine-tuned on FLAN-T5 using FIQA dataset. FLAN-T5 is a state-of-the-art language model developed by Google AI that utilizes the T5 architecture for text-generation tasks. + +References: + +```bibtex +@article{flant5_paper, + title={Scaling instruction-finetuned language models}, + author={Chung, Hyung Won and Hou, Le and Longpre, Shayne and Zoph, Barret and Tay, Yi and Fedus, William and Li, Eric and Wang, Xuezhi and Dehghani, Mostafa and Brahma, Siddhartha and others}, + journal={arXiv preprint arXiv:2210.11416}, + year={2022} +} + +@article{t5_paper, + title={Exploring the limits of transfer learning with a unified text-to-text transformer}, + author={Raffel, Colin and Shazeer, Noam and Roberts, Adam and Lee, Katherine and Narang, Sharan and Matena, Michael and Zhou, Yanqi and Li, Wei and Liu, Peter J}, + journal={The Journal of Machine Learning Research}, + volume={21}, + number={1}, + pages={5485--5551}, + year={2020}, + publisher={JMLRORG} +} +``` {:.btn-box}