← Back to context

Comment by peteforde

5 hours ago

Reasonable question and hopefully an interesting answer...

The simple lack of reasons to use TTL logic in 2026 was exactly why I didn't know what the deal was. It'd never come up, but I'd see it referenced.

I'm self-taught and in defiance of the people who insist that LLMs turn our brains to passive mush, the more things I learn the more things I have to be curious about.

LLMs remove the gatekeeping around asking "simple" questions that tend to make EEs roll their eyes. I didn't know, so I asked and now I know!

What was the answer?

I’m just curious at this point about what the quality of the answer is, just because you made a point about LLM use not turning your brain into mush.

I’ve not really used LLMs to answer questions, since it hasn’t gotten me the answers I wanted, but maybe I’m just set in my ways.

  • I'm actually pretty thrilled that you asked, because I think that this chat is an extremely solid example of LLM usage in the EE domain, and I'm happy to share.

    https://chatgpt.com/share/69a184b0-7c38-8012-b36d-c3f2cefc13...

    I definitely led some questions to try and squeeze new-to-me perspectives out of it; for example, there could be tricks that make the active high variant more useful in some scenarios.

    I think it does a good job of surfacing adjacent questions you might not realize you were eager to ask, as well as showing how it's able to critically evaluate real-world part suitability. I do find that ChatGPT in particular does better with a screengrab of the most likely parts vs a URL to the search engine.