It’s possible translate has gotten better with AI. The old versions, however, were not necessarily using AI principles.
I remember learning about image recognition tools that were simply based around randomized goal-based heuristics. It’s tricky programming, but I certainly wouldn’t call it AI. Now, it’s a challenge to define what is and isn’t; and likely a lot of labeling is just used to gather VC funding. Much like porn, it becomes a “know it when I see it” moment.
Image recognition depends on the amount of resources you can offer for your system. There are traditional methods of feature extractions like edge detection, histogram of oriented gradients and viola-jones, but the best performers are all convolutional neural networks.
While the term can be up for debate, you cannot separate these cases and things like LLMs and image generators, they are the same field. Generative models try to capture the distribution of the data, whereas discriminitive models try to capture the distribution of labels given the data. Unlike traditional programming, you do not directly encode a sequence of steps that manipulate data into what you want as a result, but instead you try to recover the distributions based on the data you have, and then you use the model you have made in new situations.
And generative and discriminative/diagnostic paradigms are not mutually exclusive either, one is often used to improve the other.
I understand that people are angry with the aggressive marketing and find that LLMs and image generators do not remotely live up to the hype (I myself don’t use them), but extending that feeling to the entire field to the point where people say that they “loathe machine learning” (which as a sentence makes as much sense as saying that you loathe the euclidean algorithm) is unjustified, just like limiting the term AI to a single digit use cases of an entire family of solutions.
It’s possible translate has gotten better with AI. The old versions, however, were not necessarily using AI principles.
I remember learning about image recognition tools that were simply based around randomized goal-based heuristics. It’s tricky programming, but I certainly wouldn’t call it AI. Now, it’s a challenge to define what is and isn’t; and likely a lot of labeling is just used to gather VC funding. Much like porn, it becomes a “know it when I see it” moment.
Image recognition depends on the amount of resources you can offer for your system. There are traditional methods of feature extractions like edge detection, histogram of oriented gradients and viola-jones, but the best performers are all convolutional neural networks.
While the term can be up for debate, you cannot separate these cases and things like LLMs and image generators, they are the same field. Generative models try to capture the distribution of the data, whereas discriminitive models try to capture the distribution of labels given the data. Unlike traditional programming, you do not directly encode a sequence of steps that manipulate data into what you want as a result, but instead you try to recover the distributions based on the data you have, and then you use the model you have made in new situations.
And generative and discriminative/diagnostic paradigms are not mutually exclusive either, one is often used to improve the other.
I understand that people are angry with the aggressive marketing and find that LLMs and image generators do not remotely live up to the hype (I myself don’t use them), but extending that feeling to the entire field to the point where people say that they “loathe machine learning” (which as a sentence makes as much sense as saying that you loathe the euclidean algorithm) is unjustified, just like limiting the term AI to a single digit use cases of an entire family of solutions.