Comment by auxiliarymoose
3 days ago
> Today we have AI models that can already do as good, almost as good, or even better than the average human in many many tasks, including the ones you mentioned.
We really don't. There are demos that look cool onstage, but there is a big difference between "in store good" and "at home good" in the sense that products aren't living up to their marketing during actual use.
IMO there is a lot of room to grow within the traditional approaches of "yesterday" - The problem is that large orgs get bogged down in legacy + bureaucracy, and most startups don't understand the business problems well enough to make a better solution. And I don't think that there is any technical silver bullet that can solve either of these problems (AI or otherwise)
I am wondering how often do you use AI models? Because I do it on a daily basis, and as much as they have limitations, I find them to be performing incredibly well. It's far very far from being a demo - last time it was a demo that looked "cool" was around 2020/21 when they were cool for spitting out the haiku poetry, and perhaps 2022 when capabilities were not as good. But today? Completely mind-blowing.
If you're not convinced, I suggest you to search for the law firms, hospitals, and laboratories ... all of which are using AI models as of today to do both the research and boiler-plate work. Creative industries are being literally erased by the generative AI as we are speaking. What will happen with the Photoshop and other similar tools when I can create whatever I want using the free AI model in literally 2 seconds without prior knowledge? What will happen with majority of movie effect makers when single guy will be able to do the work of 5 people at the same time? Or interior designers? The heck, what will happen with the Google search - I anticipate nobody will be using it in a year or two. I already don't because it's a massive sink of time compared to what I can do with perplexity for example.
There's many many examples. You just need to have your mind open to see it.
You're making a ridiculously overconfident statement.
* Show me a discrete manufacturing company using AI models for statistical process control or quality reporting
* Show me a pharmaceutical company using AI models for safety data analysis
* Show me an engineering company using AI models for structural design
The list goes on and on. There are precious few industries or companies that have replaced traditional analysis & prediction with AI. Why? Because one of two things are true: 1) their data is already in highly structured relational stores that have long legacies of SQL-based extraction and analysis, 2) they're in regulated industries and have to have audit-proof, explainable reporting, or 3) they need evidence-based design and analysis that has a key component coming from real people observing real processes in action.
For all the hyped "AI Automation" you read about, there are 100 other things that aren't, or where firms don't believe they can be, or where they'll struggle to for [reasons].
Right right, I get it. Pharma, structural engineering, discrete manufacturing, ..., all of the industries which are "too hard" to be conquered by some stupid statistical parrot. You're being delusional my friend but I am not going to be the one trying to persuade you to believe otherwise. I am here for sharing experiences and interesting discussions from which I can learn and I am not here for combating triggered and defensive strangers on the internet. And FWIW both of your conclusion and premise, and interpretation of my comment is wrong.
I try them from time to time, but I have yet to see AI models produce a useful output for my work. The problems I work on are not well-represented in training data and internet-based resources, and correctness matters far more than speed.
For my work, it's important to form strong and correct mental models of complex systems so I can reason about them well. It's more about thinking and writing clearly than anything else. LLMs tend to include subtle mistakes or even completely incorrect information (and reasoning!) which disrupts this process.
On the creative industry side...well, you can produce some results that look fine by themselves, but producing large-scale cohesive artwork (games, movies, etc.)? It's mostly human elbow-grease for the foreseeable future.
> LLMs tend to include subtle mistakes or even completely incorrect information (and reasoning!) which disrupts this process.
You see, so many humans do that as well but yet we make it believe that LLMs are somehow special here. Yes, they make mistakes, we make mistakes, I make mistakes, your colleague makes mistakes, but that's not the point of this discussion at all and it's a form of confirmation bias.
Why this reflex gag happens, I think, is because people are biased to catch somebody in making the mistake so that they can feel superior and irreplaceable. We do that to keep our position strong (in society, work), and this is completely natural - it's called the survival instinct, and it is present in our species regardless of LLMs. LLMs are just one of the ways we can obviously trigger this condition.
So, your response is no more special than other people combating the AI but take into account that "The problems I work on are not well-represented in training data" or "correctness matters far more than speed" or "it's important to form strong and correct mental models of complex systems so I can reason about them well" makes a great deal of strong assumptions. Almost any domain can literally copy-paste this into their defense, and this is also something interesting I observed through the course of years - every domain I worked in, and there were plenty, each domain thought that it is their domain which is the "hardest". Vanity.
4 replies →