DEVELOPMENT OF A QUESTION ANSWERING CHATBOT FOR BLOCKCHAIN DOMAIN
DOI:
https://doi.org/10.37943/15XNDZ6667Keywords:
Chatbot, LLM, LangChain, RAG, NLP, ChatGPTAbstract
Large Language Models (LLMs), such as ChatGPT, have transformed the field of natural language processing with their capacity for language comprehension and generation of human-like, fluent responses for many downstream tasks. Despite their impressive capabilities, they often fall short in domain-specific and knowledge-intensive domains due to a lack of access to relevant data. Moreover, most state-of-art LLMs lack transparency as they are often accessible only through APIs. Furthermore, their application in critical real-world scenarios is hindered by their proclivity to produce hallucinated information and inability to leverage external knowledge sources. To address these limitations, we propose an innovative system that enhances LLMs by integrating them with an external knowledge management module. The system allows LLMs to utilize data stored in vector databases, providing them with relevant information for their responses. Additionally, it enables them to retrieve information from the Internet, further broadening their knowledge base. The research approach circumvents the need to retrain LLMs, which can be a resource-intensive process. Instead, it focuses on making more efficient use of existing models. Preliminary results indicate that the system holds promise for improving the performance of LLMs in domain-specific and knowledge-intensive tasks. By equipping LLMs with real-time access to external data, it is possible to harness their language generation capabilities more effectively, without the need to continually strive for larger models.
References
Okasov, B. (2022, September 30). Dlja cifrovizacii vseh gosuslug v Kazahstane ispol'zujut opyt Kaspi.kz [The experience of Kaspi.kz is used to digitalize all public services in Kazakhstan]. KTK. https://www.ktk.kz/ru/news/video/2022/09/30/223965/
Alaklabi, S., & Kang, K. (2021). Perceptions towards cryptocurrency adoption: A case of Saudi Arabian citizens. Journal of Electronic Banking Systems, 1–17. https://doi.org/10.5171/2021.110411
Mensah, I. K., & Mwakapesa, D. S. (2022). The drivers of the behavioral adoption intention of bitcoin payment from the perspective of Chinese citizens. Security and Communication Networks, 2022, 1–17. https://doi.org/10.1155/2022/7373658
King, M. R. (2022). The future of AI in medicine: A perspective from a chatbot. Annals of Biomedical Engineering, 51(2), 291–295. https://doi.org/10.1007/s10439-022-03121-w
Carayannopoulos, S. (2018). Using chatbots to aid transition. International Journal of Information and Learning Technology, 35(2), 118–129. https://doi:10.1108/ijilt-10-2017-0097
Dale, R. (2016). The return of the chatbots. Natural Language Engineering, 22(05), 811–817. https://doi:10.1017/s1351324916000243
Zemčík, M. T. (2019). A brief history of chatbots. https://www.semanticscholar.org/paper/A-Brief-History-of-Chatbots-Zem%C4%8D%C3%ADk/b72c89500dd57f1a4ceadb97f3dbf5015948a5e7
Shum, H., He, X., & Li, D. (2018). From Eliza to XiaoIce: challenges and opportunities with social chatbots. Frontiers of Information Technology & Electronic Engineering, 19(1), 10–26. https://doi.org/10.1631/fitee.1700826
Klopfenstein, L. C., Delpriori, S., Malatini, S., & Bogliolo, A. (2017). The rise of Bots. Proceedings of the 2017 Conference on Designing Interactive Systems. https://doi.org/10.1145/3064663.3064672
Zemčík, T. (2019). A brief history of chatbots. DEStech Transactions on Computer Science and Engineering, aicae. https://doi.org/10.12783/dtcse/aicae2019/31439
Marietto, M. D. G. B., De Aguiar, R. V., De Oliveira Barbosa, G., Botelho, W. T., Pimentel, E. P., França, R. D. S., & Da Silva, V. L. (2013). Artificial Intelligence Markup Language: A brief tutorial. International Journal of Computer Science & Engineering Survey, 4(3), 1–20. https://doi.org/10.5121/ijcses.2013.4301
Molnar, G., & Szuts, Z. (2018). The role of Chatbots in formal education. 2018 IEEE 16th International Symposium on Intelligent Systems and Informatics (SISY). https://doi.org/10.1109/sisy.2018.8524609
Scopus preview - Scopus - Welcome to Scopus. (n.d.). https://www.scopus.com/term/analyzer.uri?sort=plf-f&src=s&sid=de60450aae1e6d2749d1e0a05c7dacdc&sot=a&sdt=a&sl=22&s=TITLE-ABS-KEY%28chatbot%29&origin=resultslist&count=10&analyzeResults=Analyze+results
Zhou, C., Li, Q., Li, C., Yu, J., Liu, Y., Wang, G., Zhang, K., Ji, C., Yan, Q., He, L., Peng, H., Li, J., Wu, J., Liu, Z., Xie, P., Xiong, C., Pei, J., Yu, P. S., & Sun, L. (2023, May 1). A comprehensive survey on pretrained foundation models: A history from Bert to chatgpt. arXiv.org. https://arxiv.org/abs/2302.09419
Y. Fu, H. Peng, and T. Khot. (2022). How does GPT Obtain its Ability? Tracing Emergent Abilities of Language Models to their Sources. Notion. https://yaofu.notion.site/How-does-GPT-Obtain-its-Ability-Tracing-Emergent-Abilities-of-Language-Models-to-their-Sources-b9a57ac0fcf74f30a1ab9e3e36fa1dc1
Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, L., & Polosukhin, I. (2023, August 2). Attention is all you need. arXiv.org. https://arxiv.org/abs/1706.03762
Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019, May 24). Bert: Pre-training of deep bidirectional Transformers for language understanding. arXiv.org. https://arxiv.org/abs/1810.04805
OpenAI. (2022, November 30). Introducing ChatGPT. OpenAI. https://openai.com/blog/chatgpt#OpenAI
Liu, X., Zheng, Y., Du, Z., Ding, M., Qian, Y., Yang, Z., & Tang, J. (2021). GPT understands, too. arXiv (Cornell University). https://doi.org/10.48550/arxiv.2103.10385
Alshater, M. (2022). Exploring the role of artificial intelligence in enhancing academic performance: A case study of chatgpt. SSRN Electronic Journal. https://doi.org/10.2139/ssrn.4312358
Biswas, S. (2022). Role of CHATGPT in computer programming. Mesopotamian Journal of Computer Science, 20–28. https://doi.org/10.58496/mjcsc/2022/004
Jeblick, K., Schachtner, B., Dexl, J., Mittermeier, A., Stüber, A. T., Topalis, J., Weber, T., Wesp, P., Sabel, B., Ricke, J., & Ingrisch, M. (2022, December 30). Chatgpt makes medicine easy to swallow: An exploratory case study on simplified radiology reports. arXiv.org. https://arxiv.org/abs/2212.14882
Bang, Y. (2023, February 8). A multitask, multilingual, multimodal evaluation of ChatGPT on reasoning, hallucination, and interactivity. arXiv.org. https://arxiv.org/abs/2302.04023
OpenAI. (2023, March 15). GPT-4 Technical Report. arXiv.org. https://arxiv.org/abs/2303.08774
Natalie. (2023, October). What is ChatGPT? OpenAI Help Center. https://help.openai.com/en/articles/6783457-what-is-chatgpt
Zhao, W. X., Zhou, K., Li, J., Tang, T., Wang, X., Hou, Y., Min, Y., Zhang, B., Zhang, J., Dong, Z., Du, Y., Yang, C., Chen, Y., Chen, Z., Jiang, J., Ren, R., Li, Y., Tang, X., Liu, Z., … Wen, J.-R. (2023, September 11). A survey of large language models. arXiv.org. https://arxiv.org/abs/2303.18223
Knight, W. (2023, April 17). OpenAI’s CEO says the age of giant AI models is already over. WIRED. https://www.wired.com/story/openai-ceo-sam-altman-the-age-of-giant-ai-models-is-already-over/
Lewis, P., Perez, E., Piktus, A., Petroni, F., Karpukhin, V., Goyal, N., Küttler, H., Lewis, M., Yih, W., Rocktäschel, T., Riedel, S., & Kiela, D. (2021, April 12). Retrieval-augmented generation for knowledge-intensive NLP tasks. arXiv.org. https://arxiv.org/abs/2005.11401
Zhang, Y., Sun, S., Gao, X., Fang, Y., Brockett, C., Galley, M., Gao, J., & Dolan, B. (2022). RetGen: A joint framework for retrieval and grounded text generation modeling. Proceedings of the AAAI Conference on Artificial Intelligence, 36(10), 11739–11747. https://doi.org/10.1609/aaai.v36i10.21429
Lazaridou, A., Gribovskaya, E., Stokowiec, W., & Grigorev, N. (2022, May 23). Internet-augmented language models through few-shot prompting for open-domain question answering. arXiv.org. https://arxiv.org/abs/2203.05115
Madaan, A., Tandon, N., Clark, P., & Yang, Y. (2023, February 18). Memory-assisted prompt editing to improve GPT-3 after deployment. arXiv.org. https://arxiv.org/abs/2201.06009
Isbister, T., Carlsson, F., & Sahlgren, M. (2021). Should we Stop Training More Monolingual Models, and Simply Use Machine Translation Instead? ACL Anthology. https://aclanthology.org/2021.nodalida-main.42/
Liu, D., & Hsu, H. (2009). An international comparison of empirical generalized double diamond model approaches to Taiwan and Korea. Competitiveness Review: An International Business Journal, 19(3), 160–174. https://doi.org/10.1108/10595420910962043
Saabith, A. S., Fareez, M. M. M., & Vinothraj, T. (2019). Python current trend applications-an overview. International Journal of Advance Engineering and Research Development, 6(10). https://www.ijaerd.com/index.php/IJAERD/article/view/4419
Topsakal, O., & Akinci, T. C. (2023). Creating large language model applications utilizing LangChain: A primer on developing LLM Apps Fast. International Conference on Applied Engineering and Natural Sciences, 1(1), 1050–1056. https://doi.org/10.59287/icaens.1127
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2023 Articles are open access under the Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Authors who publish a manuscript in this journal agree to the following terms:
- The authors reserve the right to authorship of their work and transfer to the journal the right of first publication under the terms of the Creative Commons Attribution License, which allows others to freely distribute the published work with a mandatory link to the the original work and the first publication of the work in this journal.
- Authors have the right to conclude independent additional agreements that relate to the non-exclusive distribution of the work in the form in which it was published by this journal (for example, to post the work in the electronic repository of the institution or publish as part of a monograph), providing the link to the first publication of the work in this journal.
- Other terms stated in the Copyright Agreement.