From Prompt Injections to SQL Injection Attacks: How Protected is Your LLM-Integrated Web Application?

Abstract

Large Language Models (LLMs) have found widespread applications in various domains, including web applications, where they facilitate human interaction via chatbots with natural language interfaces. Internally, aided by an LLM-integration middleware such as Langchain, user prompts are translated into SQL queries used by the LLM to provide meaningful responses to users. However, unsanitized user prompts can lead to SQL injection attacks, potentially compromising the security of the database. Despite the growing interest in prompt injection vulnerabilities targeting LLMs, the specific risks of generating SQL injection attacks through prompt injections have not been extensively studied. In this paper, we present a comprehensive examination of prompt-to-SQL (P2_2SQL) injections targeting web applications based on the Langchain framework. Using Langchain as our case study, we characterize P2_2SQL injections, exploring their variants and impact on application security through multiple concrete examples. Furthermore, we evaluate 7 state-of-the-art LLMs, demonstrating the pervasiveness of P2_2SQL attacks across language models. Our findings indicate that LLM-integrated applications based on Langchain are highly susceptible to P2_2SQL injection attacks, warranting the adoption of robust defenses. To counter these attacks, we propose four effective defense techniques that can be integrated as extensions to the Langchain framework. We validate the defenses through an experimental evaluation with a real-world use case application.Comment: 12 pages, 3 figures, 3 tables, 5 listing

    Similar works

    Full text

    thumbnail-image

    Available Versions