全部 标题 作者
关键词 摘要

OALib Journal期刊
ISSN: 2333-9721
费用:99美元

查看量下载量

相关文章

更多...

对比经微调的ERNIE-Lite-8K-0922和GPT-4在使用Prompt策略后在英语对话系统中的表现:以心理咨询师角色为例
Comparison of the Performance of Fine-Tuned ERNIE-Lite-8K-0922 and GPT-4 in English Dialogue Systems after Using the Prompt Strategy: A Case Study of the Role of a Psychological Counselor

DOI: 10.12677/airr.2024.132029, PP. 272-281

Keywords: 模型微调,提示词工程,大语言模型,英语对话系统,人工智能
Fine-Tuning
, Prompt Engineering, Large Language Models, English Dialogue System, Artificial Intelligence

Full-Text   Cite this paper   Add to My Lib

Abstract:

本研究基于大模型在英语对话系统中的实际应用对比了经过微调的ERNIE-Lite-8K-0922和GPT-4模型在采用Prompt策略后在英语对话系统中的能力表现。本研究采用了一系列定量指标,如BLEU、ROUGE分数、训练损失等指标,展示了模型微调的效果,使用自然度、逻辑性、上下文理解、多轮对话处理和情感表达等指标,评估了模型生成回复的质量。本研究在指出了ERNIE-Lite-8K-0922和GPT-4在英语对话系统中的性能差异的同时,还提出了需要进一步完善数据集与微调参数等方法以提高微调后的ERNIE-Lite-8K-0922在英语对话系统及特定领域的表现能力。本研究为探索是否有更加经济高效的方法在实际应用场景中将大语言模型部署为英语对话系统提供了重要参考,也为英语对话系统及相关领域的进一步发展做出了贡献。
Based on the practical application of the Large Language Models in the English dialogue system, this study compares the performance of the fine-tuned ERNIE-Lite-8K-0922 and gpt-4 model in the English dialogue system after using the prompt strategy. This study uses a series of quantitative indicators, such as BLEU, ROUGE, training loss and other indicators, to show the effect of model fine-tuning, and uses indicators such as naturalness, logicality, context understanding, multiple rounds of dialogue processing and emotional expression to evaluate the quality of response generated by the model. While pointing out the performance differences between ERNIE-Lite-8K-0922 and gpt-4 in English dialogue system, this study also proposes the need to further improve the data set and fine tune parameters to improve the performance of fine-tuned ERNIE-Lite-8K-0922 in English dialogue system and specific fields. This study provides an important reference for exploring whether there is a more cost-effective method to deploy the large language model into the English dialogue system in the actual application scenario, and also makes a contribution to the further development of the English dialogue system and related fields.

References

[1]  Lin, T.Y., Wang, Y.X., Liu, X.Y., et al. (2022) A Survey of Transformers. AI Open, 3, 111-132.
https://doi.org/10.1016/j.aiopen.2022.10.001
[2]  Tay, Y., Dehghani, M., Bahri, D., et al. (2022) Efficient Transformers: A Survey. ACM Computing Surveys, 55, 109.
https://doi.org/10.1145/3530811
[3]  Radford, A., Narasimhan, K., Salimans, T., et al. (2023) Improving Language Understanding by Generative Pre-Training. Openai.
https://www.semanticscholar.org/paper/Improving-Language-Understanding-by-Generative-Radford-Narasimhan/cd18800a0fe0b668a1cc19f2ec95b5003d0a5035
[4]  Radford, A., Wu, J., Child, R., et al. (2019) Language Models Are Unsupervised Multitask Learners. Openai.
https://paperswithcode.com/paper/language-models-are-unsupervised-multitask
[5]  Brown, T.B., Mann, B., Ryder, N., et al. (2020) Language Models Are Few-Shot Learners. Proceedings of the 34th International Conference on Neural Information Processing Systems, Vancouver, 1877-1901.
https://arxiv.org/abs/2005.14165
[6]  Long, O.Y., Wu, J., Slama, K., et al. (2022) Training Language Models to Follow Instructions with Human Feedback. Computation and Language.
https://arxiv.org/abs/2203.02155
[7]  Openai. (2023) GPT-4 Technical Report.
https://arxiv.org/abs/2303.08774.2023
[8]  王昀, 胡珉, 塔娜, 等. 大语言模型及其在政务领域的应用[J/OL]. 清华大学学报(自然科学版), 1-10.
https://doi.org/10.16511/j.cnki.qhdxxb.2023.26.042
[9]  贾骥业. 《中国人工智能大模型地图研究报告》发布, 发展态势可圈可点[EB/OL].
http://news.cyol.com/gb/articles/2023-05/28/content_pQ3l4YfYBE.html, 2024-03-15.
[10]  曹静. 《AI大模型技术能力评估报告, 2023》发布 百度AI大模型整体竞争力位于领先水平[EB/OL].
http://tech.chinadaily.com.cn/a/202307/19/WS64b7bddca3109d7585e459ae.html, 2024-03-15.
[11]  姜莎, 赵明峰, 张高毅. 生成式人工智能(AIGC)应用进展浅析[J]. 移动通信, 2023, 47(12): 71-78.
[12]  徐月梅, 胡玲, 赵佳艺, 等. 大语言模型的技术应用前景与风险挑战[J/OL]. 计算机应用, 1-10.
http://kns.cnki.net/kcms/detail/51.1307.TP.20230911.1048.006.html, 2024-03-15.
[13]  赵广立. 文心一言是如何炼成的? [N]. 中国科学报, 2023-03-23(003).
https://doi.org/10.28514/n.cnki.nkxsb.2023.000712
[14]  Qiu, H.C., He, H.L., Zhang, S., Li, A.Q. and Lan, Z.Z. (2023) SMILE: Single-Turn to Multi-turn Inclusive Language Expansion via ChatGPT for Mental Health Support.
https://github.com/qiuhuachuan/smile/tree/main/data

Full-Text

comments powered by Disqus

Contact Us

service@oalib.com

QQ:3279437679

WhatsApp +8615387084133

WeChat 1538708413