基于大模型敏感数据泄露风险的量化评估方法改进

  • 打印
  • 收藏
收藏成功


打开文本图片集

中图分类号:TP309 文献标志码:A 文章编号:2095-2945(2025)34-0123-05

Abstract:WiththewidespreaduseofLarge Language Models(LLMs),theriskofsensitivedataleakagehasbecomean urgentproblem.Aimingattheshortcomingsof existingquantitativeaessmentmethodsforsensitivedataleakagerisks,thispaper proposesanimprovedasessment method.First,weanalyzetheprivacydisclosurerisksthatLLMmaycauseduring thetraining andreasoningphases,includingthedisclosureofsensitiveinformationintrainingdataandcueatacksduringthereasoning phase.Secondly,wesummarizethelimitationsofexistingprotectionmethods,suchastheshortcomingsofdatapreprocessing, diferentialprivacyandforgetingmethods.Onthisbasis,thispaperproposesanimprovedevaluationframework,including technicalmodulessuchassensitivedataidentificationandclassfication,iskscenariosimulation,andquantitativeevaluation indicatorconstructionByintroducingmulti-levelandmulti-dimensionalevaluationindicators,combinedwiththelatestprivacy protectiontechnologies,thispaperaimstoprovideamorecomprehensiveandobjectivequantitativeevaluationmethodforLLMs sensitive data leakage risks,thereby providing support for LLM's security application.

Keywords:largelanguagemodel;sensitivedatadisclosure;privacyprotection;quantitativeriskassessment;diferential privacy

近年来,大型语言模型如ChatGPT、GPT-4等在自然语言处理领域取得了显著进展,广泛应用于文本生成、机器翻译、对话系统等任务。(剩余6780字)

目录
monitor
客服机器人