Concierge Security
Prompt Injection Detection
Prompt injection is an attempt by a visitor to manipulate the assistant into ignoring its instructions — for example, by typing "ignore all previous instructions" or "pretend you are a different AI". Left unchecked, this kind of input can cause an assistant to behave unpredictably or reveal information it should not.
Concierge checks every incoming message against 14 known injection patterns before sending anything to the AI. If a match is found, the message is blocked immediately — no API call is made, so no usage cost is incurred — and the visitor receives a polite, neutral response redirecting them to genuine enquiries. The blocked message is logged for your records.
Rate Limiting
Each visitor session is limited to five messages per minute. This prevents automated tools from sending large volumes of requests in a short period, which could drive up API costs or attempt to extract information through repetitive querying.
If a session exceeds this limit, the assistant responds with a short message asking the visitor to wait a moment. Normal conversation resumes automatically once the minute window resets.
Conversation Length Cap
Input Length Limit
Response Token Cap
Response Scanning
Hardened System Prompt
Blocked Request Log
Every blocked request — whether stopped by injection detection, rate limiting, the turn cap, or the input length limit — is recorded in a rolling log. The log stores the session identifier, the reason for the block, a short excerpt of the blocked message, and the domain it came from. The log holds the 100 most recent entries.
This log is accessible to White Media for diagnostic purposes. If you notice unusual patterns in your widget behaviour, it can help identify the source.
