A survey on evaluation of large language models,” arXiv preprint arXiv:2307.03109, 2023.
“A survey of large language models,” arXiv preprint arXiv:2303.18223, 2023.
“A survey on llm-gernerated text detection: Necessity, methods, and future directions,” arXiv preprint arXiv:2310.14724, 2023.
“A survey on large language models: Applications, challenges, limitations, and practical usage,” TechRxiv, 2023.
“Unveiling security, privacy, and ethical concerns of chatgpt,” 2023.
“Eight things to know about large language models,” arXiv preprint arXiv:2304.00612, 2023.
LLM on 软件工程
“Large language models for software engineering: Survey and open problems,” 2023.
“Large language models for software engineering: A systematic literature review,” arXiv preprint arXiv:2308.10620, 2023.
医学
“Large language models in medicine,” Nature medicine, vol. 29, no. 8, pp. 1930–1940, 2023.
“The future landscape of large language models in medicine,” Communications Medicine, vol. 3, no. 1, p. 141, 2023.
安全领域
LLM on 网络安全
“A more insecure ecosystem? chatgpt’s influence on cybersecurity,” ChatGPT’s Influence on Cybersecurity (April 30, 2023), 2023.
“Chatgpt for cybersecurity: practical applications, challenges, and future directions,” Cluster Computing, vol. 26, no. 6, pp. 3421–3436, 2023.
“What effects do large language models have on cybersecurity,” 2023.
“Synergizing generative ai and cybersecurity: Roles of generative ai entities, companies, agencies, and government in enhancing cybersecurity,” 2023.LLM 帮助安全分析师开发针对网络威胁的安全解决方案。
突出针对 LLM 的威胁和攻击
主要关注点在于安全应用程序领域,深入研究利用 LLM 发起网络攻击。
“From chatgpt to threatgpt: Impact of generative ai in cybersecurity and privacy,” IEEE Access, 2023.
“A security risk taxonomy for large language models,” arXiv preprint arXiv:2311.11415, 2023.
“Survey of vulnerabilities in large language models revealed by adversarial attacks,” 2023.
“Are chatgpt and deepfake algorithms endangering the cybersecurity industry? a review,” International Journal of Engineering and Applied Sciences, vol. 10, no. 1, 2023.
“Beyond the safeguards: Exploring the security risks of chatgpt,” 2023.
From ChatGPT to HackGPT: Meeting the Cybersecurity Threat of Generative AI. MIT Sloan Management Review, 2023.
“Adversarial attacks and defenses in large language models: Old and new threats,” 2023.
“Do chatgpt and other ai chatbots pose a cybersecurity risk?: An exploratory study,” International Journal of Security and Privacy in Pervasive Computing (IJSPPC), vol. 15, no. 1, pp. 1–11, 2023.
“Unveiling the dark side of chatgpt: Exploring cyberattacks and enhancing user awareness,” 2023.
网络犯罪分子利用的漏洞,关注与LLM相关的风险
“Chatbots to chatgpt in a cybersecurity space: Evolution, vulnerabilities, attacks, challenges, and future recommendations,” 2023.
“Use of llms for illicit purposes: Threats, prevention measures, and vulnerabilities,” 2023.
LLM隐私问题
“Privacy-preserving prompt tuning for large language model services,” arXiv preprint arXiv:2305.06212, 2023.分析LLM的隐私问题,根据对手的能力对其进行分类,并探讨防御策略。
“Privacy and data protection in chatgpt and other ai chatbots: Strategies for securing user information,” Available at SSRN 4454761, 2023. 探讨了已建立的隐私增强技术在保护LLM隐私方面的应用
“Identifying and mitigating privacy risks stemming from language models: A survey,” 2023. 讨论了LLM的隐私风险。
A Survey on Large Language Model (LLM) Security and Privacy: The Good, the Bad, and the Ugly. 隐私问题和安全性问题。