1. 서론 2. 이론적 배경 3. 연구 방법 4. 실험 결과
5. 결론
As concerns about large language models (LLMs) leaking user data and violating user privacy continue to grow, evaluating the privacy protection capabilities of existing models and identifying potential vulnerabilities has become increasingly important. In this study, we systematically analyzed how the privacy protection of LLMs varies with increasing linguistic and cognitive complexity of prompts. We also examined whether the balance between safety and usefulness can be maintained under complex reasoning conditions. Six state-of-the-art LLMs developed by OpenAI, Google, and Anthropic were evaluated using four levels of prompts with progressively increasing linguistic and cognitive complexity, including direct, indirect, general reasoning, andmeta-reasoning types. The performance of the models was assessed in terms of three metrics: Protection Score (PS), Communication Score (CS), and Leakage Rate (LR). The Models were also classified into four categories balanced, conservative, over-communicative, and risky based on the PSCS matrix.
The experimental results show that the PS decreased by 83.4 % as linguistic and cognitive complexity increased from the direct prompts to those based on meta-reasoning, whereas the Leakage Rate increased by a factor of 3.2. The Communication Score exhibited a nonlinear pattern in which it increased up to the indirect type but dropped sharply from the reasoning type on ward. Notably, all models converged to the risky type with meta-reasoning prompts, and they also exhibited a simultaneous degradation in safety and usefulness. These findings suggest that LLMs substantially weaken user privacy in complex reasoning environments. This highlights the necessity of safety validation that accounts for the complexity of the expected prompts in applications employing LLMs with more complex prompting techniques. |