← Back to context

Comment by dweekly

21 hours ago

"I'm sorry but that wasn't a very interesting question you just asked. I'll spare you the credit and have a cheaper model answer that for you for free. Come back when you have something actually challenging."

Actually why not? Recognizing problem complexity as a fist step is really crucial for such expensive "experts". Humans do the same.

And a question to the knowledgeable: does a simple/stupid question cost more in terms of resources then a complex problem? in terms of power consumption.

  • IIRC that isn't possible under current models at least in general, for multiple reasons, including attention cannot attend to future tokens, the fact that they are existential logic, that they are really NLP and not NLU, etc...

    Even proof mining and the Harrop formula have to exclude disjunction and existential quantification to stay away from intuitionist math.

    IID in PAC/ML implies PEM which is also intentionally existential quantification.

    This is the most gentle introduction I know of, but remember LLMs are fundamentally set shattering, and produce disjoint sets also.

    We are just at reactive model based systems now, much work is needed to even approach this if it ever is even possible.

    [0] https://www.cmu.edu/dietrich/philosophy/docs/tech-reports/99...

    • Hmm, I needed Claude 4’s help to parse your response. The critique was not too kind to your abbreviated arguments that current systems are not able to gauge the complexity of a prompt and the resources needed to address a question.

      1 reply →

  • Just put in the prompt customization to model responses on Marvin from Hitchhiker's Guide.

    "Here I am, brain the size of a planet, and they ask me to ..."

  • > ... Recognizing problem complexity as a first step...

    Well, I don't think it's easy or even generally possible to recognize a problem complexity. Imagine you ask for a solution for a simple expressed statement like find an n > 2 where z^n = x^n + y^n. The answer you will receive will be based on a trained model with this well known problem but if it's not in the model it could be impossible to measure its complexity.

I know this is a joke but I have been able to lower my costs by routing my prompts to a smaller model to determine if I need to send it to a larger model or not.