Ask HN: What Online LLM / Chat do you use?

2 days ago

I have been wanting to try more LLMs than the standard Anthropic/Grok/ChatGPT/Qwen

Are there other LLM chat sites you use or recommend?

Deepseek all the way, for both chat and coding. I can't complain as their price is the less expensive of all, started with just $2, worked all February and still $1.95 left in balance

Working on my first AI app for medical advice and I've been using it myself extensively to the point of addiction, it's magical

https://mediconsulta.net

  • I have heard mixed reviews about deepseek regarding coding how's your experience has been using deepseek when coding

    • I love it for the things that I do. It's not comparable to Claude which is the top coding experience right now but for typing chunks of code, refactoring, etc is more than enough.

      I don't have a full CLI agent, I just installed deepseek as a plugin in my Sublime Editor, some python code that takes the selected text or file and can do things to it on command, sometimes I just enter a prompt without selection so it can create code for me.

      All in all, enough for me as a lone coder, it doesn't have continuous integration or team edition which I don't care for, and for the price (cents per month) I couldn't ask for more.

      Let me know if I can be of any help

Through my Kagi subscription I get access to quite a few models [1] but I tend to rely on Qwen3 (fast) for quick questions and Qwen3 (reasoning) when I want a more structured approach, for example, when I am researching a topic.

I have tried the same approach with Kimi K2.5 and GLM 5 but I keep going back fo Qwen3.

I also have access to Perplexity which is quite decent to be honest, but I prefer to keep everything in Kagi.

1: https://help.kagi.com/kagi/ai/assistant.html#available-llms

DeepSeek (reasoning) and Gemini (multimodal) have been useful for me — especially when I want stronger pushback or a different angle. What are you hoping to get that you’re not getting from the usual set?

  • I recently deleted my ChatGPT and was just looking around for what the other options are. I kinda like fast, but Grok / Perplexity are behind a perpetual CloudFlare check for me. I'm looking for something that doesn't spend forever reasoning out the answer to some basic quick question.

    It's interesting, as I type that out it makes me wonder why not just go back to the search engine since it has the AI summaries that have been getting better.

    Finally, I do also like the longer reasoning when I have a tough question and usually like to copy paste it around to various models and compare their responses.

    • I’ve had similar friction experiences — especially when reasoning-heavy modes take longer or get retried. That repels me too.

      On the search engine comparison: do you feel LLMs reduce cognitive load because they maintain context, whereas search requires more manual synthesis?

      Also curious — do you think the frustration is mostly with the model itself, or with the serving/infrastructure layer (Cloudflare, routing, batching, etc.) around it? Both comments seem to point at that layer in different ways.

  • Do you customize or personalize these at all? Or use it just out of the box? I use gemini for image generation and as an, assistant through my android although not frequently. As a chat assistant the online Google UI is a turnoff personally.

    • For tone, I prefer efficient — higher signal-to-noise.

      In custom instructions I’ll sometimes write something like “tell it like it is.” Interestingly, some models then start with “okay, I’ll tell you like it is…” which makes me wonder whether that’s psychological framing, operational behavior, or both.

      Curious why you asked — have you found customization meaningfully changes performance?

      1 reply →

I’ve been using the Kagi Assistant with various models. It’s more of a glorified search summarizer but it’s essentially free with my existing subscription.

ChatGPT just because, then Gemini because it loads much faster. I don't have to wait a few seconds for the web UI to be ready. The frequent ChatGPT downtimes are what got me to look for alternatives.

I tend to rotate between a few depending on the task — some are better at summaries, others handle logic better. One thing that really helps is trying smaller niche models for specific use-cases instead of always defaulting to the big ones.

Mistral is still a lot of fun, especially with the ChatGPT/Claude voice becoming too common. Openrouter gives you access to it for free, or you can download it to LM Studio.

Lumo by Proton is pretty performant. Mostly, it has an epic free tier and I dig their stance on privacy-first AI.