Skip to content

Commit

Permalink
Fix references
Browse files Browse the repository at this point in the history
  • Loading branch information
boorlakov committed Oct 4, 2023
1 parent c2f986e commit 8d22a34
Show file tree
Hide file tree
Showing 2 changed files with 9 additions and 9 deletions.
18 changes: 9 additions & 9 deletions bibliography/references.tex
Original file line number Diff line number Diff line change
Expand Up @@ -3,37 +3,37 @@
\bibitem{human-wer}
\textit{Lippmann R. P.} Speech recognition by machines and humans [Текст] // Speech communication. – 1997. – Т. 22. – №. 1. – С. 1-15.
\bibitem{whisper}
\textit{Radford A. et al.} Robust speech recognition via large-scale weak supervision [Текст] // arXiv preprint arXiv:2212.04356. – 2022.
\textit{Radford A. et al.} Robust speech recognition via large-scale weak supervision [Электронный ресурс]. -- URL: \url{https://arxiv.org/pdf/2212.04356.pdf} (дата обр. 11.06.2023)
\bibitem{state-of-gpt}
\textit{Karpathy A.} State of GPT [Электронный ресурс]. -- URL: \url{https://karpathy.ai/stateofgpt.pdf} (дата обр. 11.06.2023)
\bibitem{transformer-paper}
\textit{Vaswani A. et al.} Attention is all you need [Текст] // Advances in neural information processing systems. – 2017. – Т. 30.
\bibitem{backprop-theory}
\textit{Hecht-Nielsen R.} Theory of the backpropagation neural network [Текст] // Neural networks for perception. – Academic Press, 1992. – С. 65-93.
\bibitem{optimizers-paper}
\textit{Ruder S.} An overview of gradient descent optimization algorithms [Текст] // arXiv preprint arXiv:1609.04747. – 2016.
\textit{Ruder S.} An overview of gradient descent optimization algorithms [Электронный ресурс]. -- URL: \url{https://arxiv.org/pdf/1609.04747.pdf} (дата обр. 11.06.2023)
\bibitem{adafactor-paper}
\textit{Shazeer N., Stern M.} Adafactor: Adaptive learning rates with sublinear memory cost [Текст] // International Conference on Machine Learning. – PMLR, 2018. – С. 4596-4604.
\bibitem{word2vec-paper}
\textit{Mikolov T. et al.} Efficient estimation of word representations in vector space [Текст] // arXiv preprint arXiv:1301.3781. – 2013.
\textit{Mikolov T. et al.} Efficient estimation of word representations in vector space [Электронный ресурс]. -- URL: \url{https://arxiv.org/pdf/1301.3781.pdf} (дата обр. 11.06.2023)
\bibitem{bpe-paper}
\textit{Sennrich R., Haddow B., Birch A.} Neural machine translation of rare words with subword units [Текст] // arXiv preprint arXiv:1508.07909. – 2015.
\textit{Sennrich R., Haddow B., Birch A.} Neural machine translation of rare words with subword units [Электронный ресурс]. -- URL: \url{https://arxiv.org/pdf/1508.07909.pdf} (дата обр. 11.06.2023)
\bibitem{gpt-paper}
\textit{Radford A. et al.} Improving language understanding by generative pre-training. – 2018.
\textit{Radford A. et al.} Improving language understanding by generative pre-training [Электронный ресурс]. -- URL: \url{https://www.cs.ubc.ca/~amuham01/LING530/papers/radford2018improving.pdf} (дата обр. 11.06.2023)
\bibitem{t5-paper}
\textit{Raffel C. et al.} Exploring the limits of transfer learning with a unified text-to-text transformer [Текст] // The Journal of Machine Learning Research. – 2020. – Т. 21. – №. 1. – С. 5485-5551.
\bibitem{bert-paper}
\textit{Devlin J. et al.} Bert: Pre-training of deep bidirectional transformers for language understanding [Текст] // arXiv preprint arXiv:1810.04805. – 2018.
\textit{Devlin J. et al.} Bert: Pre-training of deep bidirectional transformers for language understanding [Электронный ресурс]. -- URL: \url{https://arxiv.org/pdf/1810.04805.pdf} (дата обр. 11.06.2023)
\bibitem{chatgpt-docs}
Документация ChatGPT [Электронный ресурс]. -- URL: \url{https://openai.com/blog/chatgpt} (дата обр. 06.06.2023)
\bibitem{llama-paper}
\textit{Touvron H. et al.} Llama: Open and efficient foundation language models [Текст] // arXiv preprint arXiv:2302.13971. – 2023.
\textit{Touvron H. et al.} Llama: Open and efficient foundation language models [Электронный ресурс]. -- URL: \url{https://arxiv.org/pdf/2302.13971.pdf}
\bibitem{alpaca-docs}
Документация Alpaca [Электронный ресурс]. -- URL: \url{https://crfm.stanford.edu/2023/03/13/alpaca.html} (дата обр. 06.06.2023)
\bibitem{flan-paper}
\textit{Chung H. W. et al.} Scaling instruction-finetuned language models [Текст] // arXiv preprint arXiv:2210.11416. – 2022.
\textit{Chung H. W. et al.} Scaling instruction-finetuned language models [Электронный ресурс]. -- URL: \url{https://arxiv.org/pdf/2210.11416.pdf} (дата обр. 11.06.2023)
\bibitem{sentencepiece-paper}
\textit{Kudo T., Richardson J.} Sentencepiece: A simple and language independent subword tokenizer and detokenizer for neural text processing [Текст] // arXiv preprint arXiv:1808.06226. – 2018.
\textit{Kudo T., Richardson J.} Sentencepiece: A simple and language independent subword tokenizer and detokenizer for neural text processing [Электронный ресурс]. -- URL: \url{https://arxiv.org/pdf/1808.06226.pdf} (дата обр. 11.06.2023)
\bibitem{python-lang-site}
Официальный сайт языка программирования Python [Электронный ресурс]. -- URL: \url{https://www.python.org/} (дата обр. 16.06.2023)
\bibitem{cpp-docs}
Expand Down
Binary file modified thesis.pdf
Binary file not shown.

0 comments on commit 8d22a34

Please sign in to comment.