anism: in addition to retrieving the best documents,
RAG generates answers -instead of retrieving the best
span of text- and complements the process by evalu-
ating relevant words within the retrieved documents
related to the given question. Then, all individual an-
swer predictions are aggregated into a final prediction,
a method which they found to enable the backpropa-
gation of the error signals in the output to the retrieval
mechanism, and with it potentially improving the per-
formance on end-to-end systems such as the one de-
scribed in this research. It would be therefore interest-
ing to gauge the performance of a RAG architecture,
compared to the architectures depicted in this paper.
Finally, we acknowledge that the methodology de-
picted in this research work could be applied in dif-
ferent industry sectors, but it does provide a roadmap
for researchers and practitioners in artificial intelli-
gence for higher education to implement open domain
question-answering systems. In this regard, this con-
ference is the right venue to promote a discussion on
these topics.
ACKNOWLEDGEMENTS
The authors would like to thank Marist College IT de-
partment for their sponsorship and support. We would
also like to thank the Data Science and Analytics
team for their collaboration throughout this research
project. A special mention goes to Maria Kapogian-
nis, Ethan Aug, Michael Volchko and Jack Spagna
who worked on the creation of the set of questions
and answers used to fine-tune the models.
REFERENCES
Apache Software Foundation (2021). Lucene. https://
lucene.apache.org.
Chen, D., Fisch, A., Weston, J., and Bordes, A. (2017).
Reading wikipedia to answer open-domain questions.
Clark, K., Luong, M.-T., Le, Q. V., and Manning, C. D.
(2020). Electra: Pre-training text encoders as discrim-
inators rather than generators.
Devlin, J., Chang, M.-W., Lee, K., and Toutanova, K.
(2019). BERT: Pre-training of Deep Bidirectional
Transformers for Language Understanding. eprint:
1810.04805.
Elastic (2021). Elasticsearch. https://github.com/elastic/
elasticsearch.
Erhan, D., Bengio, Y., Courville, A., Manzagol, P.-A., Vin-
cent, P., and Bengio, S. (2010). Why does unsuper-
vised pre-training help deep learning? Journal of Ma-
chine Learning Research, 11(19):625–660.
Goodfellow, I., Bengio, Y., and Courville, A. (2016).
Deep Learning, page 526. MIT Press. http://www.
deeplearningbook.org.
Green, B. F., Wolf, A. K., Chomsky, C. L., and Laughery,
K. (1961). Baseball: an automatic question-answerer.
In IRE-AIEE-ACM ’61 (Western).
Johnson, J., Douze, M., and J
´
egou, H. (2017). Billion-
scale similarity search with gpus. arXiv preprint
arXiv:1702.08734.
Jurafsky, D. and Martin, J. H. (2021). Speech and Language
Processing: An Introduction to Natural Language
Processing, Computational Linguistics, and Speech
Recognition. 3rd, draft edition. https://web.stanford.
edu/
∼
jurafsky/slp3/.
Karpukhin, V., Oguz, B., Min, S., Lewis, P., Wu, L.,
Edunov, S., Chen, D., and Yih, W.-t. (2020). Dense
passage retrieval for open-domain question answer-
ing. In Proceedings of the 2020 Conference on
Empirical Methods in Natural Language Processing
(EMNLP), pages 6769–6781, Online. Association for
Computational Linguistics.
Lewis, P., Perez, E., Piktus, A., Petroni, F., Karpukhin,
V., Goyal, N., K
¨
uttler, H., Lewis, M., tau
Yih, W., Rockt
¨
aschel, T., Riedel, S., and Kiela,
D. (2021). Retrieval-augmented generation for
knowledge-intensive nlp tasks.
Lin, J., Ma, X., Lin, S.-C., Yang, J.-H., Pradeep, R., and
Nogueira, R. (2021). Pyserini: An easy-to-use python
toolkit to support replicable ir research with sparse
and dense representations.
Liu, Y., Ott, M., Goyal, N., Du, J., Joshi, M., Chen, D.,
Levy, O., Lewis, M., Zettlemoyer, L., and Stoyanov,
V. (2019). Roberta: A robustly optimized bert pre-
training approach.
Pietsch, Soni, Chan, M
¨
oller, and Kosti
´
c (2021). Haystack
(version 0.5.0). https://github.com/deepset-ai/
haystack/.
Rajpurkar, P., Jia, R., and Liang, P. (2018). Know what you
don’t know: Unanswerable questions for squad.
Robertson, S. and Zaragoza, H. (2009). The probabilistic
relevance framework: Bm25 and beyond. Founda-
tions and Trends in Information Retrieval, 3:333–389.
Salton, G. (1971). The SMART Retrieval Sys-
tem—Experiments in Automatic Document Process-
ing. Prentice-Hall, Inc., USA.
Seo, M., Kembhavi, A., Farhadi, A., and Hajishirzi, H.
(2018). Bidirectional attention flow for machine com-
prehension.
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J.,
Jones, L., Gomez, A. N., Kaiser, L., and Polosukhin,
I. (2017). Attention Is All You Need. eprint:
1706.03762.
Wang, S. and Jiang, J. (2016). Machine comprehension us-
ing match-lstm and answer pointer.
Wang, W., Wei, F., Dong, L., Bao, H., Yang, N., and Zhou,
M. (2020). Minilm: Deep self-attention distillation for
task-agnostic compression of pre-trained transform-
ers.
Wolf, T., Debut, L., Sanh, V., Chaumond, J., Delangue, C.,
Moi, A., Cistac, P., Rault, T., Louf, R., Funtowicz,
CSEDU 2022 - 14th International Conference on Computer Supported Education
74