Commit Graph

3 Commits

Author SHA1 Message Date
85dec4908f security: hide defense mechanism from user-facing prompt display
Split system prompt and user message into public/private versions:
- Private versions (sent to LLM): include delimiter tags, anti-injection
  instructions, and 'never reveal' directives
- Public versions (shown to user via 'Show prompt'): clean prompt
  without any defense details, raw user text without tag wrappers

The user never sees:
- The ###### delimiter tags wrapping their input
- The instruction to ignore embedded instructions
- The instruction to never reveal the system prompt
- The instruction not to acknowledge delimiter tags

This prevents an attacker from learning the defense mechanism
and crafting injections that work around it.
2026-04-12 23:42:31 -04:00
90bb701068 fix: eliminate redundancy in system prompt
The old prompt had two problems:
1. {style} placeholder was filled with the full promptModifier sentence,
   producing gibberish like "rewrite strongly in a Rewrite in a
   sarcastic... style"
2. The promptModifier was then repeated as its own line

New design separates concerns cleanly:
- intensityMap no longer uses {style} placeholder — instructions are
  pure intensity adverbs ("strongly", "subtly, with a light touch", etc.)
- buildSystemPrompt strips the leading "Rewrite" verb from the style
  modifier and combines both into one non-redundant instruction:
  "Rewrite the text strongly: in a sarcastic, snarky tone with biting wit"

Example outputs by intensity:
  1: Rewrite the text subtly, with a light touch: in a sarcastic...
  3: Rewrite the text strongly: in a sarcastic...
  5: Rewrite the text with absolute maximum intensity, no restraint: ...
2026-04-12 23:23:58 -04:00
a12afb792e feat: implement English Style Converter
- SvelteKit project scaffolded with TypeScript
- Type definitions for Style, StyleCategory, ConversionRequest, ConversionResponse, LLMConfig
- Style definitions with 6 categories and 25 sub-styles
- Intensity mapping (1-5) with prompt modifier placeholders
- LLM client using OpenAI-compatible API (Ollama default)
- POST /api/convert endpoint with input validation
- Animated loading modal with per-letter animations
- Main page UI with category/style selectors, intensity slider
- Copy to clipboard, collapsible prompt display
- Vitest tests for styles, LLM prompt building, and API validation
- Environment configuration for LLM settings
2026-04-12 21:53:27 -04:00