They said AI would take you places. They never said they were places you wanted to go.
Is this loss?
“Cohesive”
When we say LLMs don’t know or understand anything, this is what we mean. This is a perfect example of an “AI” just not having any idea what it’s doing.
- I’ll start with a bit of praise: It does do a fairly good job of decomposing the elements of Python and the actuary profession into bits that would be representative of those realms.
But:
- In the text version of the response, there are already far too many elements for a good tattoo, demonstrating it doesn’t understand tattoo design or even just design
- In the drawn version, the design uses big blocks of color with no detail, which (even if they looked good on a white background; and they don’t;) would look like shit inked on someone’s skin. So again, no understand of tattoo art.
- It produces a “simplified version” of the python logo. I assume those elements are the blue and yellow hexagons, which are at least the correct colors. But it doesn’t understand that, for this to be PART OF THE SAME DESIGN, they must be visually connected, not just near each other. It also doesn’t understand that the design is more like a plus; nor that the design is composed of two snakes; nor that the Python logo is ALREADY VERY SIMPLE, nor that the logo, lacking snakes, loses any meaning in its role of representing Python.
- It says there’s a briefcase and glasses in there. Maybe the brown rectangle? Or is the gray rectangle meant to be a briefcase lying on its side so the handle is visible? No understanding here of how humans process visual information, or what makes a visual representation recognizable to a human brain.
- Math stuff can be very visually interesting. Lots of mathematical constructs have compelling visuals that go with them. A competent designer could even tie them into the Python stuff in a unified way; like, imagine a bar graph where the bars were snakes, twining around each other in a double helix. You got math, you got Python, you got data analysis. None of this ties together, or is even made to look good on its own. No understanding of what makes something interesting.
- Everything is just randomly scattered. Once again, no understanding of what design is.
AIs do not understand anything. They just regurgitate in ways that the algorithm chooses. There’s no attempt to make the algorithm right, or smart, or relevant, or anything except an algorithm that’s just mashing up strings and vectors.
My main concern with people making fun of such cases is about deficiencies of “AI” being harder to find/detect but obviously present.
Whenever someone publishes a proof of a system’s limitations, the company behind it gets a test case to use to improve it. The next time we - the reasonable people arguing that cybernetic hallucinations aren’t AI yet and are dangerous - try using such point, we would only get a reply of “oh yeah, but they’ve fixed it”. Even people in IT often don’t understand what they’re dealing with, so the non-IT people may have even more difficulties…
Myself - I just boycott this rubbish. I’ve never tried any LLM and don’t plan to, unless it’s used to work with language, not knowledge.
Here’s gemini’s attempt:
The last one is a cool concept, but pie charts are pretty useless lmao
I’m going to get that last one. “What’s on your arm?“ ‘oh it’s just a third of it’ put my three arms together
I’m pretty sure it’s from different perspectives since it’s wrapping around the arm
Duhhhhh
I’m not much of a tattoo expert. Designs on my body include:
Yup normal picture for tats that you can’t get in one pic. I mean that’s why it generated like that, a ton of those pics exist and got shoved into the magic picture box.
If you want a snake and a pie chart, at least have the snake do something with it like carrying the chart in its mouth.
Perhaps you can do the biblical scene of the snake tempting Adam and Eve but this time it’s the snake tempting managers with a useless pie chart.