Why are hands difficult for ai images?

536 views

Why are hands difficult for ai images?

In: 250

63 Answers

Anonymous 0 Comments

Specifically, it should be asked why *the current* AI models have trouble with hands.

Give it a year, won’t be an issue. Give it five years and people won’t even remember it bring a problem. You’ll only remember because you posted it here.

I’m sure as they worked on the model, it would make mistakes on other things like faces or body proportions and the developers noticed quickly and applied modifications. Hands just slipped through the cracks. Probably because it is less noticeable than an eye out of place.

It’s a common misconception that all AI models are slowly learning from us using them. That’s not true for most AIs. There are three major categories for Machine learning:

**Supervised** – You have a dataset where you know the outcome. Like a list of dog and cat pictures. This is good for identifying fake news titles or the reddit bots that tell you how sarcastic a reddit post is.

**Unsupervised** – You have a dataset but you don’t have outcomes. Just a large set of unspecified data and the AI itself has to detect patterns and reproduce. **This is what image AIs use.** Also good for natural language stuff like chatGPT.

**Reinforcement** – This is the kind that learns as it goes. You set up parameters, when it succeeds, it is rewarded. Fails are punished. They tend to be inaccurate but improve over time. This is what people think of when they hear AI. It’s good for solving video games or finding solutions to things where the answer isn’t obvious. Example: youtube algorithm.

The only way to improve an unsupervised model is to change how it functions by adding layers to the way AI handles the data. Or get a stronger dataset to train it on. The latter is already optimal these days, so it’s almost exclusively improving from developers. They’ll add a hand layer of some kind and it’ll fix the issue.

Anonymous 0 Comments

Can someone comment if this is why hands are difficult for dreams as well?

Anonymous 0 Comments

Can someone comment if this is why hands are difficult for dreams as well?

Anonymous 0 Comments

Can someone comment if this is why hands are difficult for dreams as well?

Anonymous 0 Comments

Hands and feet are a struggle for human artists as well, so let’s not be too hard on AI. The main difference is we realize there are 5 fingers meant to be there 🙂

Anonymous 0 Comments

Hands and feet are a struggle for human artists as well, so let’s not be too hard on AI. The main difference is we realize there are 5 fingers meant to be there 🙂

Anonymous 0 Comments

Hands and feet are a struggle for human artists as well, so let’s not be too hard on AI. The main difference is we realize there are 5 fingers meant to be there 🙂

Anonymous 0 Comments

There was a youtube video that explained it so well… can’t remember if it was steve Mold or Sabine or someone else. Anyways, AIs that have difficulty with hands are those that are only trained on 2D pictures and lack a model of how things are in the real world in 3D.

Those AIs recognize a human face whether it is a front picture (2 eyes, 2 ears) or profile (1 eye, 1 ear), BUT they don’t know that the difference between the two pictures is a rotation on the 3D space.

If you follow subs like r/confusingperspective you’ll soon realize that there is plenty of pictures where people seem to have 3 legs or 3 hands. But there is less confusing perspective pictures about heads, maybe because when we take picture we make sure that head and body are prominent in the shots we take.

So, a lot of confusing perspective around hands and feet… confusing to us, humans who have knowledge of the 3D world. Now imagine you are a software that only analyzes pixels in 2D.

Anonymous 0 Comments

There was a youtube video that explained it so well… can’t remember if it was steve Mold or Sabine or someone else. Anyways, AIs that have difficulty with hands are those that are only trained on 2D pictures and lack a model of how things are in the real world in 3D.

Those AIs recognize a human face whether it is a front picture (2 eyes, 2 ears) or profile (1 eye, 1 ear), BUT they don’t know that the difference between the two pictures is a rotation on the 3D space.

If you follow subs like r/confusingperspective you’ll soon realize that there is plenty of pictures where people seem to have 3 legs or 3 hands. But there is less confusing perspective pictures about heads, maybe because when we take picture we make sure that head and body are prominent in the shots we take.

So, a lot of confusing perspective around hands and feet… confusing to us, humans who have knowledge of the 3D world. Now imagine you are a software that only analyzes pixels in 2D.

Anonymous 0 Comments

There was a youtube video that explained it so well… can’t remember if it was steve Mold or Sabine or someone else. Anyways, AIs that have difficulty with hands are those that are only trained on 2D pictures and lack a model of how things are in the real world in 3D.

Those AIs recognize a human face whether it is a front picture (2 eyes, 2 ears) or profile (1 eye, 1 ear), BUT they don’t know that the difference between the two pictures is a rotation on the 3D space.

If you follow subs like r/confusingperspective you’ll soon realize that there is plenty of pictures where people seem to have 3 legs or 3 hands. But there is less confusing perspective pictures about heads, maybe because when we take picture we make sure that head and body are prominent in the shots we take.

So, a lot of confusing perspective around hands and feet… confusing to us, humans who have knowledge of the 3D world. Now imagine you are a software that only analyzes pixels in 2D.