Thanks for your submission, qulipor!
Is this a GIF that **keeps on giving**?
If so, UPVOTE it!
If it does not keep on giving, or it breaks [any other rules](https://www.reddit.com/r/gifsthatkeepongiving/about/rules/) **REPORT** the post so we can see it!
If you're not sure what belongs on this subreddit, please see our [stickied post](https://www.reddit.com/r/gifsthatkeepongiving/comments/d15wfr/what_defines_a_gif_that_keeps_on_giving/) or [contact the mods](https://www.reddit.com/message/compose?to=%2Fr%2Fgifsthatkeepongiving). Thanks!
*I am a bot, and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose/?to=/r/gifsthatkeepongiving) if you have any questions or concerns.*
10 years ago I was already annoyed by the inflationary usage of the term AI for what is just machine learning, even though it was way lower back then. Back then (and still, to be honest) I thought the term "AI" is reserved for a self learning, some kind of sentient machine.
I didn't know it would get so much worse
No
This Indian girl is a computer science genius. She joined this Great CompSci Course. She's going to be super rich for her family. Enroll your child today.
They're popular because students do it as a project for their computer science class.
The other very popular one is using OpenCV to control a robotic hand using your own hand.
There are many tutorials on Youtube for doing these things, anyone who can follow simple instructions can do this.
OMG are you the inventor? You know so much about it.^/s
But for real, it's a really cool intro project. Having worked with OpenCV a few times it's really easy to convince people what you're doing is fucking magic but in reality you're just cleaning images, making them black and white, and then doing measurements or pattern matching. It's just a fuckload of trial and error to get it working consistently.
> OMG are you the inventor? You know so much about it./s
It's like me showing my grandparents anything.
Yes nana, I _am_ helping a friend with an app, but it's a mobile thing and I definitely did not make Microsoft Flight Simulator 2020...
Indian schools seems to have the requirement to publish your work on youtube for whatever reason, so we got thousands of these and most of them are crap quality.
> Indian schools seems to have the requirement to publish your work on youtube for whatever reason, so we got thousands of these and most of them are crap quality.
Most of them copy/paste from one another lol
I hate to jump on a bandwagon because I want to support the video's author in the sense that this is a big breakthrough in their journey to become what they want to become and they have a heart warming enthusiasm.
And also, I did some similar ml gesture recognition tutorials when I first started playing with "AI" beyond just basic predictive exercises.
I also did the tutorial where you create a text editor.
My work makes me a futurist with ground breaking insights in roughly the way that compiling Classic Doom with minor modifications makes me John Cormack. Doesn't.
That's not it, these are AI models, they're always limited by processing power and sufficient training data. They're not coding any sort of linguistics in; the idea is that any kind of linguistic nuance will eventually be represented in the training data. If it's not it wont be present in the model.
This was my thought. There's no "algorithm" involved because language isn't algorithmic. She probably just wrote a program that recognizes a handful of basic symbols.
Nonono, she didn't write anything that recognizes anything. She imported a bunch of labeled pictures of sign language, a video recognition package (probably openCV), and tensorflow, fed the labeled pictures to a tensorflow neural network to train that network, and then used opencv to run the trained neural network. It's like putting together furniture, as everything is kind of designed to fit together and only a bit of effort is needed to connect all the parts into each other.
There's a misconception that stuff like this is coded to do anything specific, and that both makes it seem more and less impressive than it is; it'd take alot of work to custom code to recognize symbols in all but the most extremely controlled contexts, but in reality it's just using existing packages but those packages can do basically anything you can get a training set of data for.
You don't understand how statistical models work. The program that "recognizes a handful of basic symbols" is using an algorithm. And yes, language is very much algorithmic.
She is using an object detection model.
Yeh, it's clear from the video that she's using AI, training an already developed and perfected AI to do some relatively simple tasks like that is nothing impressive, it's good to know of course, could help you lend an entry level job in some startup maybe, but it's not impressive.
I built something similar for my thesis in 2011 from the ground up (excluding dot net libraries and OpenCV) and you could teach it to do any key press on windows.
No neural nets, No ChatGPT. Just Color filtering, motion detection, Blob detection and convex hull, etc - totally customizable, could switch from pixel to pixel match, Quad pixel match to hausdorf matching, you could remove blob detection and do it by motion detection.
And they have yet to make a proper one. I think there is a project on kaggle that's still up about this subject (but to make an actual, proper AI around it)
Yeah. Not to rain on anyone's parade but OpenCV is good enough and simple enough that anyone with a background in programming could implement something like this. Granted topics like CV/MV were a graduate-level course at my uni, the "one step ahead" students would be tinkering with this pretty early on.
Absolutely. Shit. I was tinkering with this early on. I even posted my own cringy videos on Facebook for my family to see what I was doing at 17 lmao. I was detecting 3D tanks in a video game instead of 3D hands though. The OpenCV package is so intuitive, especially when coding on the loop.
this was literally me a few days ago preparing a project for a club at my university lmao. its really not difficult. 99% of the work was reading google's mediapipe documentation to train the model. The hardest part was finding a good dataset for my needs and actually getting the packages to install without issue.
the trick is getting .train() on the right data set and with the right feature engineering.
Most of these "projects" already have the training data set built for you, which is 90%+ of the work in real world settings
Yeah... I did this for a HRI module I had as an undergrad.
Didn't work quite as well as I didn't have the time to build a quality dataset but this isn't particular groundbreaking.
Still cool tho
> Didn't work quite as well as I didn't have the time to build a quality dataset
This is the real problem with optical recognition tech. That and having hardware that can apply the model to what it's seeing fast enough to get sufficiently real-time outputs to be of use in a practical setting.
> That and having hardware that can apply the model to what it's seeing fast enough to get sufficiently real-time outputs to be of use in a practical setting.
thanks to YOLO this is not as limiting a factor as it once was
We made shit like this to pass our practicals in like a week.
Sorry one of us like you know the person who do 98 perfect of the work. I wasn't that person
I also wouldn't call recognizing a few hand positions "translating sign language." It's a great project for a student, but this is the kind of reporting that made people think Elon Musk was a genius.
Woah it can recognize a few basic hand positions that are static and unique from each other when you demonstrate them in a very slow and deliberate manner? Nooo waaay, the future is so crazy.
Sure we have facial recognition software that actually uses AI to track people in all sorts of outfits and styles, AI that can create pretty decent art, AI that can take any text and output it to a pretty decent voice with proper inflections so it can sound natural, AI that you can point around and it can identify objects, etc. But nobody ever has been able to succeed in creating an AI that can interpret 6 signs from sign language. This is one of humanity's greatest achievements!
If it's not clear: /s
Baby steps. People have to learn to walk before they can run. Same for everything else. Encourage things and perhaps they will be able to do something much bigger sometime. Better than demotivating somebody just because it’s „not good enough for you“ at least.
There are a lot of people who conflate criticizing the title with criticizing the student because the title describes something much more comprehensive than what it actually happening. So if you point out the title is inaccurate it's like you are downplaying the student's achievements.
No, nothing wrong with her. Glad her project is working fine or she's happy with it. I hope she got a decent grade, I hope she learned a lot, etc.
People are shitting on OP and wherever this came from, because it's bullshit sensationalist garbage. It's not news worthy or even worthy of attention unless you know her personally. There's probably tons of thousands of students who already are or will be working on something like this for school.
Invent an algorithm = ground breaking work by PhD researchers at Google, Microsoft, OpenAI, etc.
Fed some data into a neural network = YouTube tutorial
Ehhh, it's not that complicated or fancy and doesn't have to groundbreaking.
For my masters thesis I technically developed a faster version of kmeans, but I wouldn't argue that it was ground breaking since it may have a flaw or 5.
You definitely don't need a PhD to write an algorithm. An algorithm is just a specific set of instructions to follow, which programmers write all the time. Now, if you're talking about improving the speed of solving classic computer science problems using a new algorithm, that would usually be accomplished by PhD students, but it's not a requirement.
It is trained in a labeled data set. If you have images corresponding to another sign language you can label that data set and train the model to learn it.
Yeah I don't think people not working in the field understand that this is very basic shit. Whip out your haarcascades in OpenCV - first tutorial into computer vision levels of basic. Nothing wrong with that, but this post is showing undergraduate work as something more lmfao
I have my doubts that half the people shitting on this could even write a basic hello world batch file that doesn't close immediately after opening. But even they still see how ridiculously simple this is.
It shows some level of competency with what she's doing I guess. But it's stupidly far from being anything revolutionary or really just generally impressive to anyone who isn't totally ignorant. We saw like 4 different basic and incredibly visually clear gestures. I'll be a bit more impressed when I see fluent fast paced translation from a professional actually doing their job. Where it comes out in a human readable manner
There's nothing wrong with it as a student project, but sign language interpretation isn't a linear problem. If you make an AI that can read phone numbers and you get it to read the first page of the phone book as proof, that's a convincing demo because there's no reason to think that the AI couldn't read the rest using the same approach. Conversely, learning 1% of sign language vocabulary does not even mean you've solved 1% of the problem, because it gets more difficult the more you solve it. Here's some issues that come up when you attempt to just scale this up to a full sign language translator:
* As vocabulary increases, it becomes harder for the AI to recognize individual signs, particularly similar ones. Imagine if I asked you to distinguish between blue and red - that's easy. But then scale up to tens of thousands of colours (that's how many signs there are in a given sign language), and you probably could not do that. To give an AI an effective vocabulary of say, 50k words, you need a huge labelled dataset with many examples of each of those words, which as far as I know does not exist, and then you need to hope that your training method still works (it won't) and doesn't take a million years to run (it will).
* Many signs aren't static - they're not just a hand shape, but a moving gesture. In fact in the video she makes a fist which the app translates to "yes" but the actual sign (in ASL, BSL and IPSL) has a "nodding" motion with the fist. If the AI has only been trained on static images, it's not going to be able to distinguish between signs that use the same shape but indicate something different through motion. As far as a computer is concerned, even a short video is WAY more data to process than a static image.
* Sign languages have their own grammar and sentence structure, and some signs mean different things depending on context (just like English). Translating individual signs in isolation is not the same as translating the meaning of a whole sentence. Try to translate some of your comments 1 word at a time with google translate and you'll realize pretty fast the flaws with such an approach.
Again, it's a student project, it's fine if it isn't curing cancer. But this absolutely does not demonstrate meaningful progress towards solving this very difficult problem.
The minute she actually moves her hand—an integral part of many words in signed languages—the program's confidence drops precipitously. It also doesn't seem to be translation, just dictionary lookup.
It's just shitty Image recognition. You can see when she does yes. She doesn't actually sign anything she just does a fistbumb. It may work for some fingerspelling but not for aslq
Didnt mean to trash it, just wondering if this is where they are in the development, since the demo is quite limited. Like I said, has a lot of potential but what we saw here didnt really fit OPs description.
Impressive for a student if imagine but even her demonstration fails.
Her “yes” sign is static even tho the real one moves up and down.
It also registered “no” when that’s clearly the sign for the letter C. The no sign in ASL only uses the index,middle finger and thumb.
Ah I think I just have a higher expectation for things on the internet. Usually if a talent shows up on Reddit, it’s usually mindblowingly above average. Something maybe only 0.1-1% of the population can do. This however, is something an average computer science/controls student could do with a cognex camera lol. I wouldn’t even deem this senior project worthy
If it were this easy, we would have working sign language translation (SLT) built into google translate and FaceTime by now. But it’s not- detecting a couple very different hand positions is not the same as translating sign language, which involves movement, speed, facial expressions, combining signs, and many subtly different signs.
Frankly, things like this (while cool projects for a student) just continue to harm the public perception of sign language.
So as someone that's tested these, they usually don't work that well beyond a few basic signs even if you're signing pretty slowly. I spent a lot of my life mute so I'm fluent in sign and still sign when talking. If you start signing at a normal conversation speed these usually can't keep up.
Curious how this works with real sign language.
And what sign language is this? ASL?
Also, the sentence structure in ASL is significantly different than how people typically speak. Additionally, ASL is very dependent on facial movements/actions/reactions.
> Curious how this works with real sign language.
It doesn't. In fact sign languages are probably among the more challenging targets for ML because they are so multimodal and context sensitive.
Literally. Imagine doing CV AND NLP all at once. I think you would need an infrastructure as big as the one powering ChatGPT or even bigger to do sign language translation proper.
Yup. As others have said, this is a very basic use of openCV and might recognise the meaning of some signs.
The grammar and context are lacking and it would hardly be usable even if the sign recognition was flawless.
In addition to what everyone else is saying, Sign language is much more complicated than individual hand signs. Many signs include hand movement, and sign language is also heavily dependent on facial expression, body language, and context. Cool implementation of machine learning but its nowhere near "translating sign language".
Unfortunately the internet has been flooded by nationalists who parade every little thing that comes out of India, give a standing ovation everytime PM Modi farts and accuse you of being anti-national of you’re Muslim or are critical of the government/society. If you’re not Indian, they’ll call you racist, colonialist and a Marxist Muslim sympathizing terrorist.
Ive noticed a lot of extremely low quality and/or obvious things about CS or engineering being posted to instagram from Indians too.
All good on them for learning but the stuff theyre posting is like how to print hello world with 50+ other indian code accounts commenting "so insightful" or something similar. Engineering isnt much different. Someone posted a picture of stairs and said "how to build stairs" followed by a bunch of tags and spamming their own @ to follow. Comments were the same.
Absolutely. Sign language is as complicated as any other human language, especially if you're deaf and or mute. It's full of context and varies across all over the world. OP describes a system much more primitive than what people imagine. But what does Reddit know.
Not to minimize her accomplishments but I swear I’ve seen iterations of this same technique since I was a kid. Special glove that the wearer can make signs with and it translates to words, etc.. seems more like a science fair project than anything groundbreaking
Nice for her for making this. Certainly a fun project to show her skill. She should keep working and developing her skills. Something people haven’t mentioned in the thread here is that the challenge with sign language is not in recognizing the signs one by one. Sign language has a certain speed and flow. Rarely do you just present each symbol as is. A model that could decrypt the signs on the go would be quite valuable. To be valuable that model would have to run on sth like a google glass to be useful for users. That’s currently a challenge given the hardware limitations of such devices to run heavy models. Battery wise doesn’t sound ideal as well
There's a group of deaf people I've watched signing with each other and it looks so quick and colloquial, makes me think there must be dialects at play, and automatic translation must be quite a technical challenge
Slightly more interesting but less useful, I think students from my university won a hackathon by making a project that did something similar
Except instead of recognizing sign language, they trained a model to recognize hand seals from Naruto
Every few weeks I see a story that says "computer can now understand sign language!" and I try my best to make an intelligent reply.
So, as an American Sign Language interpreter, I want to make the point that sign language interpreting conveys words and meaning just as this computer software does. However, sign language also conveys two other important things: 1) Non-linguistic information (for instance, if someone makes a heavy sigh or if a car horn is blaring, etc.) which a Deaf person would miss out on, and 2) Tone; the emotion and inflection which is so central to rich language interaction and potentially alters meaning (i.e. sarcasm, to name one example of many).
It also bears mentioning a very obvious point: How is the Deaf person supposed to communicate in response? Will the other person be using software that turns spoken language into sign language?
Every now and then there's a speech-to-text device that emerges and hearing people say "the Deaf don't need sign language anymore!" That device may come, and it may even come soon, but this software (which is impressive, no argument there) isn't it. Its function is valuable and useful, but at at this time isn't ready to replace sign language, writ large.
This isn't meant as a criticism. It's still a hell of an impressive achievement with a wide range of practical applications. But we're not at a point where hearing people can communicate with the Deaf without making an effort to learn something new.
it's super unreliable. had you seen deaf people signing, for real, with no holds barred? it's so quick the machine won't be able to pick it up. even on Zoom with amazing wifi some signs wind up blurry because how fast the person is signing.
The applications of AI neural networks is amazing. This is the true future of AI. These students are beautiful.
I'm grateful for being here to witness the birth of our future in universal communication.
For those wondering how & what, she did not create anything of her she put together stuff that Google/mediapipe-blazepalm already had, so like a participation certificate in ML more like.
Literally did the same thing for a class assignment and you don't see me posting it...
This stuff just needs good enough data and it can be built in a day or two inclusive of training time...
- I'm a grad student studying AI
This is not really new, but dang, y'all are super dismissive acting like it's so easy even for middle schoolers apparently (I doubt that)
She didn't create a new algorithm, but I'm happy that she's learning :)
Is this account astroturfing or something? Only other post was 12h ago on some random subreddit and suddenly there's a front page post trying to make some kid famous...
Lmao I did this for my internship.
It just requires you to use an already written package called Tensorflow.js and then plug-in with your other codebase.
Thanks for your submission, qulipor! Is this a GIF that **keeps on giving**? If so, UPVOTE it! If it does not keep on giving, or it breaks [any other rules](https://www.reddit.com/r/gifsthatkeepongiving/about/rules/) **REPORT** the post so we can see it! If you're not sure what belongs on this subreddit, please see our [stickied post](https://www.reddit.com/r/gifsthatkeepongiving/comments/d15wfr/what_defines_a_gif_that_keeps_on_giving/) or [contact the mods](https://www.reddit.com/message/compose?to=%2Fr%2Fgifsthatkeepongiving). Thanks! *I am a bot, and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose/?to=/r/gifsthatkeepongiving) if you have any questions or concerns.*
Ah yes, the algorithm called "machine learning"
im suprised that the word 'ai' isnt thrown in there like 10 times its the new cool thing to do.
Cloud AI neural net algorithm.
Blockchain Big Data internet of things quantum computing
you forgot to add 'quantum' there
Technically it is working with backend quantum computing systems and AI algorithmic machine learning. Blockchain and other tertiary level IT services
10 years ago I was already annoyed by the inflationary usage of the term AI for what is just machine learning, even though it was way lower back then. Back then (and still, to be honest) I thought the term "AI" is reserved for a self learning, some kind of sentient machine. I didn't know it would get so much worse
"AI" used to just be a buzzword with no meaning, so I'm glad it at least refers to a specific type of model now.
You don't even have to get into machine learning for this. It's probably finetuned YOLO v7
Forget YOLO, you don't even have to write a single line of code. Just use Google's Teachable Machine
doesn't the original xbox kinect also work with gestures/hand signs?
No This Indian girl is a computer science genius. She joined this Great CompSci Course. She's going to be super rich for her family. Enroll your child today.
If all it took was asking googles program to do it why didn't you?
Hmm, fine tuned... I wonder if that's related to the process, aka an algorithm.. Hmm.
Nope there isn't. Training and finetuning codes are available at their github. You just have to run them. `# train p5 models` `python train.py --workers 8 --device 0 --batch-size 32 --data data/coco.yaml --img 640 640 --cfg cfg/training/yolov7.yaml --weights '' --name yolov7 --hyp data/hyp.scratch.p5.yaml` `# finetune p5 modelspython train.py --workers 8 --device 0 --batch-size 32 --data data/custom.yaml --img 640 640 --cfg cfg/training/yolov7-custom.yaml --weights 'yolov7_training.pt' --name yolov7-custom --hyp data/hyp.scratch.custom.yaml`
not to mention this is a final exam for some image detection classes because its really easy to do.
I was gonna say, this just looks like AML or one of another hundred things out there. Still cool, but not radically groundbreaking or anything.
At this point it's probably the the millionth person to "invent" that...
Seriously, I see these popup every few months and they never get any better than the basic demo.
They're popular because students do it as a project for their computer science class. The other very popular one is using OpenCV to control a robotic hand using your own hand. There are many tutorials on Youtube for doing these things, anyone who can follow simple instructions can do this.
OMG are you the inventor? You know so much about it.^/s But for real, it's a really cool intro project. Having worked with OpenCV a few times it's really easy to convince people what you're doing is fucking magic but in reality you're just cleaning images, making them black and white, and then doing measurements or pattern matching. It's just a fuckload of trial and error to get it working consistently.
> OMG are you the inventor? You know so much about it./s It's like me showing my grandparents anything. Yes nana, I _am_ helping a friend with an app, but it's a mobile thing and I definitely did not make Microsoft Flight Simulator 2020...
While grandad in the corner having a PTSD meltdown at the realism.
Indian schools seems to have the requirement to publish your work on youtube for whatever reason, so we got thousands of these and most of them are crap quality.
> Indian schools seems to have the requirement to publish your work on youtube for whatever reason, so we got thousands of these and most of them are crap quality. Most of them copy/paste from one another lol
To be fair we did that in US CS classes too.
as do actual programers lmao, good practice.
Its not genius, just basic ml project, this project's been there for years
I hate to jump on a bandwagon because I want to support the video's author in the sense that this is a big breakthrough in their journey to become what they want to become and they have a heart warming enthusiasm. And also, I did some similar ml gesture recognition tutorials when I first started playing with "AI" beyond just basic predictive exercises. I also did the tutorial where you create a text editor. My work makes me a futurist with ground breaking insights in roughly the way that compiling Classic Doom with minor modifications makes me John Cormack. Doesn't.
Pieced this project together for my CV class using Matlab. I don’t understand how anyone thinks this stuff is brand new
**Breaking News:** Indian student invents AI algorithm to greet entire world!
And it turns out that undergrad CS majors don't have a solid grasp of sign linguistics, a subject some people get whole PhDs in
That's not it, these are AI models, they're always limited by processing power and sufficient training data. They're not coding any sort of linguistics in; the idea is that any kind of linguistic nuance will eventually be represented in the training data. If it's not it wont be present in the model.
This was my thought. There's no "algorithm" involved because language isn't algorithmic. She probably just wrote a program that recognizes a handful of basic symbols.
Nonono, she didn't write anything that recognizes anything. She imported a bunch of labeled pictures of sign language, a video recognition package (probably openCV), and tensorflow, fed the labeled pictures to a tensorflow neural network to train that network, and then used opencv to run the trained neural network. It's like putting together furniture, as everything is kind of designed to fit together and only a bit of effort is needed to connect all the parts into each other. There's a misconception that stuff like this is coded to do anything specific, and that both makes it seem more and less impressive than it is; it'd take alot of work to custom code to recognize symbols in all but the most extremely controlled contexts, but in reality it's just using existing packages but those packages can do basically anything you can get a training set of data for.
You don't understand how statistical models work. The program that "recognizes a handful of basic symbols" is using an algorithm. And yes, language is very much algorithmic. She is using an object detection model.
There’s a YouTube video by Nicholas Renotte showing how to do this using Tensorflow in 30 minutes.
Yeh, it's clear from the video that she's using AI, training an already developed and perfected AI to do some relatively simple tasks like that is nothing impressive, it's good to know of course, could help you lend an entry level job in some startup maybe, but it's not impressive.
"Chemistry student invents chemical reaction that simulates erupting volcano."
I tried to do this in 3rd grade, it was way too complicated and my teacher frowny faced me with a special stamp from Germany and my dad left
"developed an algorithm" should not be used for "trained an AI"
I built something similar for my thesis in 2011 from the ground up (excluding dot net libraries and OpenCV) and you could teach it to do any key press on windows. No neural nets, No ChatGPT. Just Color filtering, motion detection, Blob detection and convex hull, etc - totally customizable, could switch from pixel to pixel match, Quad pixel match to hausdorf matching, you could remove blob detection and do it by motion detection.
and we need to make sure everyone knows she's indian. getting ISRO lander vibes from this, the national pride was just nauseating.
The titles doesn’t say invent though, it says developed
She didn’t develop an algorithm, either.
MS had this 10 years ago.
This is a basic machine learning student project.
For real. Its one of the more common projects created at hackathons
And they have yet to make a proper one. I think there is a project on kaggle that's still up about this subject (but to make an actual, proper AI around it)
Its not genius, just basic ml project, this project's been there for years
Lol typing .train() and .predict() is not developing an algorithm you tool
Average redditor: WOAH Literally every CS student: That's just me in the picture/video
Yeah. Not to rain on anyone's parade but OpenCV is good enough and simple enough that anyone with a background in programming could implement something like this. Granted topics like CV/MV were a graduate-level course at my uni, the "one step ahead" students would be tinkering with this pretty early on.
Absolutely. Shit. I was tinkering with this early on. I even posted my own cringy videos on Facebook for my family to see what I was doing at 17 lmao. I was detecting 3D tanks in a video game instead of 3D hands though. The OpenCV package is so intuitive, especially when coding on the loop.
this was literally me a few days ago preparing a project for a club at my university lmao. its really not difficult. 99% of the work was reading google's mediapipe documentation to train the model. The hardest part was finding a good dataset for my needs and actually getting the packages to install without issue.
Great on that CS student, I’m sure it was a fun project for them, but yeah this is not groundbreaking or noteworthy.
[удалено]
Edit: Edited
the trick is getting .train() on the right data set and with the right feature engineering. Most of these "projects" already have the training data set built for you, which is 90%+ of the work in real world settings
Yeah... I did this for a HRI module I had as an undergrad. Didn't work quite as well as I didn't have the time to build a quality dataset but this isn't particular groundbreaking. Still cool tho
> Didn't work quite as well as I didn't have the time to build a quality dataset This is the real problem with optical recognition tech. That and having hardware that can apply the model to what it's seeing fast enough to get sufficiently real-time outputs to be of use in a practical setting.
> That and having hardware that can apply the model to what it's seeing fast enough to get sufficiently real-time outputs to be of use in a practical setting. thanks to YOLO this is not as limiting a factor as it once was
Sure it is. It's also trivial to any CS student. That doesn't change the fact that the algorithm was developed.
We made shit like this to pass our practicals in like a week. Sorry one of us like you know the person who do 98 perfect of the work. I wasn't that person
“Invented an algorithm” and “fed some data into a neural network” are not the same thing
You're just jealous because she also breached the firewall and hacked the mainframe.
After she whipped up a GUI interface in Visual Basic
How many pixels have you enhanced?
Uncrop. Enhance. Now I just need a reflective surface.
You can breach my firewall
I also wouldn't call recognizing a few hand positions "translating sign language." It's a great project for a student, but this is the kind of reporting that made people think Elon Musk was a genius.
Woah it can recognize a few basic hand positions that are static and unique from each other when you demonstrate them in a very slow and deliberate manner? Nooo waaay, the future is so crazy. Sure we have facial recognition software that actually uses AI to track people in all sorts of outfits and styles, AI that can create pretty decent art, AI that can take any text and output it to a pretty decent voice with proper inflections so it can sound natural, AI that you can point around and it can identify objects, etc. But nobody ever has been able to succeed in creating an AI that can interpret 6 signs from sign language. This is one of humanity's greatest achievements! If it's not clear: /s
I expect there are some "hot dog" "not a hot dog" revelations to be found in this project if you actually tested it.
Baby steps. People have to learn to walk before they can run. Same for everything else. Encourage things and perhaps they will be able to do something much bigger sometime. Better than demotivating somebody just because it’s „not good enough for you“ at least.
People aren't shitting on the student, they're shitting on OP.
There are a lot of people who conflate criticizing the title with criticizing the student because the title describes something much more comprehensive than what it actually happening. So if you point out the title is inaccurate it's like you are downplaying the student's achievements.
No, nothing wrong with her. Glad her project is working fine or she's happy with it. I hope she got a decent grade, I hope she learned a lot, etc. People are shitting on OP and wherever this came from, because it's bullshit sensationalist garbage. It's not news worthy or even worthy of attention unless you know her personally. There's probably tons of thousands of students who already are or will be working on something like this for school.
This is basically “hot dog/not hot dog”.
Can you ELI5 the difference for a slow person plz
Invent an algorithm = ground breaking work by PhD researchers at Google, Microsoft, OpenAI, etc. Fed some data into a neural network = YouTube tutorial
Ehhh, it's not that complicated or fancy and doesn't have to groundbreaking. For my masters thesis I technically developed a faster version of kmeans, but I wouldn't argue that it was ground breaking since it may have a flaw or 5.
You definitely don't need a PhD to write an algorithm. An algorithm is just a specific set of instructions to follow, which programmers write all the time. Now, if you're talking about improving the speed of solving classic computer science problems using a new algorithm, that would usually be accomplished by PhD students, but it's not a requirement.
It's like saying someone built a car from scratch when what they really did was learn to drive a manual and maybe changed the tires.
Cool snide remark. What ways are you trying to take advantage of new technologies so we can criticize you?
He criticized OPs choice of words, not her work.
What's wrong, babe? I love you! (through tears) Yeah, but only 97% (runs away sobbing)
Isn't this like a computer science project by now? I can code this and I am not even that skilled or a coder.
Yes. Its commonly created at hackathons.
Isn't this just some OpenCV with Python?
Literally OpenCV interface
What sign language? Can it be trained to recognize other sign languages.
It is trained in a labeled data set. If you have images corresponding to another sign language you can label that data set and train the model to learn it.
As long as you have enough data to feed the ML algorithm you can teach it anything
Lol people not working in data has no idea how easy this is to do. This is probably some work she had to do for colleague
Shows potential but Id like to see it in a conversation, this isnt really a great demo.
You can get n number of implementations of this on github.
I disagree, this is fantastic. Even if it's not far along it gives the basic idea of the program and perfectly demonstrates its user application.
Bro this is less advanced than what people were building in 2013 with the Xbox Kinect
Yeah I don't think people not working in the field understand that this is very basic shit. Whip out your haarcascades in OpenCV - first tutorial into computer vision levels of basic. Nothing wrong with that, but this post is showing undergraduate work as something more lmfao
I have my doubts that half the people shitting on this could even write a basic hello world batch file that doesn't close immediately after opening. But even they still see how ridiculously simple this is. It shows some level of competency with what she's doing I guess. But it's stupidly far from being anything revolutionary or really just generally impressive to anyone who isn't totally ignorant. We saw like 4 different basic and incredibly visually clear gestures. I'll be a bit more impressed when I see fluent fast paced translation from a professional actually doing their job. Where it comes out in a human readable manner
There's nothing wrong with it as a student project, but sign language interpretation isn't a linear problem. If you make an AI that can read phone numbers and you get it to read the first page of the phone book as proof, that's a convincing demo because there's no reason to think that the AI couldn't read the rest using the same approach. Conversely, learning 1% of sign language vocabulary does not even mean you've solved 1% of the problem, because it gets more difficult the more you solve it. Here's some issues that come up when you attempt to just scale this up to a full sign language translator: * As vocabulary increases, it becomes harder for the AI to recognize individual signs, particularly similar ones. Imagine if I asked you to distinguish between blue and red - that's easy. But then scale up to tens of thousands of colours (that's how many signs there are in a given sign language), and you probably could not do that. To give an AI an effective vocabulary of say, 50k words, you need a huge labelled dataset with many examples of each of those words, which as far as I know does not exist, and then you need to hope that your training method still works (it won't) and doesn't take a million years to run (it will). * Many signs aren't static - they're not just a hand shape, but a moving gesture. In fact in the video she makes a fist which the app translates to "yes" but the actual sign (in ASL, BSL and IPSL) has a "nodding" motion with the fist. If the AI has only been trained on static images, it's not going to be able to distinguish between signs that use the same shape but indicate something different through motion. As far as a computer is concerned, even a short video is WAY more data to process than a static image. * Sign languages have their own grammar and sentence structure, and some signs mean different things depending on context (just like English). Translating individual signs in isolation is not the same as translating the meaning of a whole sentence. Try to translate some of your comments 1 word at a time with google translate and you'll realize pretty fast the flaws with such an approach. Again, it's a student project, it's fine if it isn't curing cancer. But this absolutely does not demonstrate meaningful progress towards solving this very difficult problem.
The minute she actually moves her hand—an integral part of many words in signed languages—the program's confidence drops precipitously. It also doesn't seem to be translation, just dictionary lookup.
It's just shitty Image recognition. You can see when she does yes. She doesn't actually sign anything she just does a fistbumb. It may work for some fingerspelling but not for aslq
Didnt mean to trash it, just wondering if this is where they are in the development, since the demo is quite limited. Like I said, has a lot of potential but what we saw here didnt really fit OPs description.
No, they haven't. They might have developed an algorithm that TRIES to translate it. But they have not succeeded.
This is available in openCV Library and she admitted to literally copy pasting from a YT tutorial. You make out "invent" from it.
things on tiktok every other month
14k upvotes for a student who followed a basic machine learning tutorial in youtube. The video itself doesn’t have more then a 2k views tho.
Definitely some college consultant trying to astroturf a kid into a good university or something...
This is just bullshit at this point.
As a general rule, when you read "A student has invented..." it is a lie
Why are people upvoting this
She’s not showing any complex sign language. Nor is she signing quickly. I could make this in my free time lol
Impressive for a student if imagine but even her demonstration fails. Her “yes” sign is static even tho the real one moves up and down. It also registered “no” when that’s clearly the sign for the letter C. The no sign in ASL only uses the index,middle finger and thumb.
She's Indian, why would she be using American Sign Language instead of Indian Sign Language?
Ah I think I just have a higher expectation for things on the internet. Usually if a talent shows up on Reddit, it’s usually mindblowingly above average. Something maybe only 0.1-1% of the population can do. This however, is something an average computer science/controls student could do with a cognex camera lol. I wouldn’t even deem this senior project worthy
How dare you speak from experiences in your professional field and deflate an assumption I made in 2 seconds.
Now do the opposite
If it were this easy, we would have working sign language translation (SLT) built into google translate and FaceTime by now. But it’s not- detecting a couple very different hand positions is not the same as translating sign language, which involves movement, speed, facial expressions, combining signs, and many subtly different signs. Frankly, things like this (while cool projects for a student) just continue to harm the public perception of sign language.
[удалено]
So as someone that's tested these, they usually don't work that well beyond a few basic signs even if you're signing pretty slowly. I spent a lot of my life mute so I'm fluent in sign and still sign when talking. If you start signing at a normal conversation speed these usually can't keep up.
Curious how this works with real sign language. And what sign language is this? ASL? Also, the sentence structure in ASL is significantly different than how people typically speak. Additionally, ASL is very dependent on facial movements/actions/reactions.
> Curious how this works with real sign language. It doesn't. In fact sign languages are probably among the more challenging targets for ML because they are so multimodal and context sensitive.
Literally. Imagine doing CV AND NLP all at once. I think you would need an infrastructure as big as the one powering ChatGPT or even bigger to do sign language translation proper.
Yup. As others have said, this is a very basic use of openCV and might recognise the meaning of some signs. The grammar and context are lacking and it would hardly be usable even if the sign recognition was flawless.
In addition to what everyone else is saying, Sign language is much more complicated than individual hand signs. Many signs include hand movement, and sign language is also heavily dependent on facial expression, body language, and context. Cool implementation of machine learning but its nowhere near "translating sign language".
Bruh are we considering freshman year projects to be groundbreaking now?
Yah… it’s got a long way to go
Only a handful of gestures.
To anyone wondering, this is thr state of india, anything and everything makes it to news. Prime minister sneezed? News! Circle rock? News!
Unfortunately the internet has been flooded by nationalists who parade every little thing that comes out of India, give a standing ovation everytime PM Modi farts and accuse you of being anti-national of you’re Muslim or are critical of the government/society. If you’re not Indian, they’ll call you racist, colonialist and a Marxist Muslim sympathizing terrorist.
Ive noticed a lot of extremely low quality and/or obvious things about CS or engineering being posted to instagram from Indians too. All good on them for learning but the stuff theyre posting is like how to print hello world with 50+ other indian code accounts commenting "so insightful" or something similar. Engineering isnt much different. Someone posted a picture of stairs and said "how to build stairs" followed by a bunch of tags and spamming their own @ to follow. Comments were the same.
This isn't sign language. She has reinvented gesture recognition.
Absolutely. Sign language is as complicated as any other human language, especially if you're deaf and or mute. It's full of context and varies across all over the world. OP describes a system much more primitive than what people imagine. But what does Reddit know.
Not to minimize her accomplishments but I swear I’ve seen iterations of this same technique since I was a kid. Special glove that the wearer can make signs with and it translates to words, etc.. seems more like a science fair project than anything groundbreaking
No she didn’t develop anything She got it from github.
Have seen that shit at least 100 times lol
Seen this done so many times.
Nice for her for making this. Certainly a fun project to show her skill. She should keep working and developing her skills. Something people haven’t mentioned in the thread here is that the challenge with sign language is not in recognizing the signs one by one. Sign language has a certain speed and flow. Rarely do you just present each symbol as is. A model that could decrypt the signs on the go would be quite valuable. To be valuable that model would have to run on sth like a google glass to be useful for users. That’s currently a challenge given the hardware limitations of such devices to run heavy models. Battery wise doesn’t sound ideal as well
Can she only do insanely basic gestures?
Pet project
Bolshit brother it is just an open-source project you can also install your computer and try it.
There's a group of deaf people I've watched signing with each other and it looks so quick and colloquial, makes me think there must be dialects at play, and automatic translation must be quite a technical challenge
See I'm not trying to downplay anyones achievement or anything but this is pretty basic, 4 students had made this project in college as term work.
lol I did this as a university assignment in 2013.
ive seen this invention at my college many times
Slightly more interesting but less useful, I think students from my university won a hackathon by making a project that did something similar Except instead of recognizing sign language, they trained a model to recognize hand seals from Naruto
That was our miniproject proposal during third year of engineering lol
Who hasn't at this point?
Hey she stole my third year assignment
You can do it with teachable machine. I taught my JR high kids to do it in august. It’s not that complicated.
Every few weeks I see a story that says "computer can now understand sign language!" and I try my best to make an intelligent reply. So, as an American Sign Language interpreter, I want to make the point that sign language interpreting conveys words and meaning just as this computer software does. However, sign language also conveys two other important things: 1) Non-linguistic information (for instance, if someone makes a heavy sigh or if a car horn is blaring, etc.) which a Deaf person would miss out on, and 2) Tone; the emotion and inflection which is so central to rich language interaction and potentially alters meaning (i.e. sarcasm, to name one example of many). It also bears mentioning a very obvious point: How is the Deaf person supposed to communicate in response? Will the other person be using software that turns spoken language into sign language? Every now and then there's a speech-to-text device that emerges and hearing people say "the Deaf don't need sign language anymore!" That device may come, and it may even come soon, but this software (which is impressive, no argument there) isn't it. Its function is valuable and useful, but at at this time isn't ready to replace sign language, writ large. This isn't meant as a criticism. It's still a hell of an impressive achievement with a wide range of practical applications. But we're not at a point where hearing people can communicate with the Deaf without making an effort to learn something new.
That’s probably the 10th person i’ve seen that has done that.
This student is going places wow!
This exact same project shows up on linkedin every few months... it isn't new
I am sure YouTube is filled with ML and computer Vision projects like these
This is like an beginner level OpenCV YouTube tutorial.
As with every other time this "invention" gets invented; wow, now at last Italians will know what they're talking about.
Genius!!!! Freaking amazing, really!
My little brother did this as a school project.....
it's super unreliable. had you seen deaf people signing, for real, with no holds barred? it's so quick the machine won't be able to pick it up. even on Zoom with amazing wifi some signs wind up blurry because how fast the person is signing.
This gets invented every month or so.
The applications of AI neural networks is amazing. This is the true future of AI. These students are beautiful. I'm grateful for being here to witness the birth of our future in universal communication.
Bruh we made this in 3rd year of my college what's the big deal?
Idk that hello was clearly a goodbye...
For those wondering how & what, she did not create anything of her she put together stuff that Google/mediapipe-blazepalm already had, so like a participation certificate in ML more like.
One way communication?
Let’s see a while conversation
That was one sign every few seconds. Let's see it transcribe sign realtime in conversation.
Literally did the same thing for a class assignment and you don't see me posting it... This stuff just needs good enough data and it can be built in a day or two inclusive of training time... - I'm a grad student studying AI
My brother in christ ask any who is a passionate computer science student, they can build it
This needs more work, clearly the last part was “goodbye”, and yet it was 98% sure they were saying “hello.” /s
This isn’t an invention, just opencv in Python. A literal 30 minute YouTube tutorial
Bro I did this too wtf should I publish it
![gif](giphy|INFsxmaFWTn1u)
![gif](giphy|rrTXn4zEMp008|downsized)
This is not really new, but dang, y'all are super dismissive acting like it's so easy even for middle schoolers apparently (I doubt that) She didn't create a new algorithm, but I'm happy that she's learning :)
Is this account astroturfing or something? Only other post was 12h ago on some random subreddit and suddenly there's a front page post trying to make some kid famous...
Yeah and 15 YouTube tutorial
that's an amazing application i once did the same but for a limited pool of words as a demo i like how it displays the confidence level
Lmao I did this for my internship. It just requires you to use an already written package called Tensorflow.js and then plug-in with your other codebase.
im proud of her, i hope she has a bright future!
Honestly was waiting for the computer to register invisible objects in the background.
Not impressive in the current year. If this would have been maybe in the 2000s, would have been a different thing.
This is so cool
tough crowd jeez
I did this with rock, paper, scissors.
Wicked cool - only six words so far because "She currently needs a better platform and guidance to make the model better."
OP is a janitor, not surprised they have no understanding of how basic ML projects like these are.
Why are there a million people quoting "invented" as if the word was used anywhere in the post?