← Back to context

Comment by tptacek

2 months ago

I'm a software security person! This is not irrelevant to me.

In summary: the existing program analysis tooling in this space has been ineffective for decades, despite hundreds of millions of dollars invested in the tooling. If it is effective now, that strongly indicates that the LLM component of it isn't irrelevant; nothing else in the field has changed.

Note that everybody in this story concedes the LLM involvement. The only person who isn't is you, and you're not actually involved. (I'm not either, but I'm agreeing with --- checks again --- everybody involved in the story).

I concede the LLM involvement. But I want to be more specific in the description of the role it plays in the solution.

If it is a central role, then there is nothing to loose from describing it better. That's why this feels so strange. You disagree with me, but you don't present an arrangement in which the LLM plays a role different to what I described. In fact, no one here did. It's like you're not disagreeing with me, but trying to make me stop describing how to achieve a similar quality system out of free pieces.

  • Motte/bailey[1].

    Also, somehow, you keep coming back to this uninteresting conversation where no one offers you anything new.

    I recommending being kinder to people who offer their time. Even when we disagree or are having a rollicking discussion, there's a fundamental respect we should have for each other, if begrudging.

    [1] Where you are: "[it seems you are] trying to make me stop describing how to achieve a similar quality system out of free pieces."

    Where you started: "Do you believe AI is at the core of these security analyzers? If so, why the personal story blogpost? You can just explain me in technical terms why is that so.

    Claiming to work for Google does not work as an authority card for me, you still have to deliver a solid argument.

    Look, AI is great for many things, but to me these products sounds like chocolate that is actually just 1% real chocolate. Delicious, but 99% not chocolate."

    • Irrelevant. I actually started by describing the system, which was my first comment on the post.

      I am not responsible for anyone that gets offended if I say that something is not as AI as it seems to be. It's obviously not a personal offense. If you took it as such, it's really not my problem.

      Rejecting an argument by authority like "I'm an ex Googler!" or "I'm a security engineer" is also common sense. Maybe it works a lot and you folks are upset it's not working here, but that's just the way it is.

      6 replies →