Egor Zverev

PhD
Institute of Science and Technology Austria (ISTA)
Instruction-Data Separation in Language Models

Instruction-tuned Large Language Models (LLMs) show impressive results in numerous practical applications, but they lack essential safety features that are common in other areas of computer science, particularly an explicit separation of instructions and data. This makes them vulnerable to manipulations such as indirect prompt injections and generally unsuitable for safety-critical tasks. There is currently no established definition or benchmark to quantify this phenomenon, and neither are there solutions to improve instruction-data separation. In this project, we will develop such techniques.

Track:
Academic Track
PhD Duration:
September 1st, 2022 - September 30th, 2027
ELLIS Edge Newsletter
Join the 6,000+ people who get the monthly newsletter filled with the latest news, jobs, events and insights from the ELLIS Network.