Comment by bambax
1 day ago
It is indeed a working demo, hitting
https://llm.koomen.dev/v1/chat/completions
in the OpenAI API format, and it responds to any prompt without filtering. Free tokens, anyone?
More seriously, I think the reason companies don't want to expose the system prompt is because they want to keep some of the magic alive. Once most people understand that the universal interface to AI is text prompts, then all that will remain is the models themselves.
That's right. llm.koomen.dev is a cloudflare worker that forwards requests to openai. I was a little worried about getting DDOSed but so far that hasn't been an issue, and the tokens are ridiculously cheap.
Blog author seems smart (despite questionable ideas about how much real world users would want to interact with any of his elaborate feature concepts), you hope he's actually just got a bunch of responses cached and you're getting a random one each time from that endpoint... and that freely sent content doesn't actually hit OpenAI's APIs.
I tested it with some prompts, it does answer properly. My guess is it just forwards the queries with a key with a cap, and when the cap is reached it will stop responding...