Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Perhaps you could help by inventing a written version of ASL, since one doesn't currently exist. [EDIT: I was wrong, sorry, see my response below.] Seems like that would be a prerequisite for a model based on written language.

And of course if you could also create an entire corpus of training material in your written ASL?



It is called SignWriting for all you multiple naysayers.

https://en.m.wikipedia.org/wiki/SignWriting


Very fair point. I've never seen it used, but SignWriting is already in the Unicode standard[0] (at U+1D800 - U+1DAAF).

I suspect Meta/Facebook doesn't have a lot of content to work off of. I've only been able to find community-generated examples of SignWriting on the official website[1], and none of those seem to be using Unicode characters. MMS is an audio-to-text tool, so it seems unlikely that it can be trivially expanded to take in visual data (pictures of text or video of ASL being performed).

I suspect the process of turning viewed ASL into SignWriting text will be very difficult to automate. I would not be surprised if such a project would either use a different textual encoding or directly translate out to English (which also sounds terribly hard, but these LLM advances recently have surprised me).

[0] https://www.unicode.org/charts/PDF/U1D800.pdf

[1] https://signwriting.org/#RecentPostingsUS


I stand corrected, thanks. There's a lot of info on the internet that says a written form of ASL doesn't exist, which is what I found when I Googled it.

Looking into it, it seems very much at the experimental stage in terms of digital representation -- while Unicode symbols exist, they require being placed in 2D boxes (using a drawing tool like SVG). It seems like it's only in the past few years that there have been proposals for how to turn it into a linear canonical text encoding?

Is anyone actually using those linear encodings -- SignPuddle or Formal SignWriting, they seem to be called -- in the wild, outside of demonstration texts or academia? Especially since they only date to 2016 and 2019.

Is there anywhere close at all to a corpus that Meta could train on? Because it still seems like the answer is no, but I also got my research wrong when Google gave no indication that SignWriting existed in the first place.


SignWriting has been documented at National Deafness section in California State University of Northridge South Library since 1968 and at Gallaudet University, Washington, D.C. since 1950s.


That's great, but it doesn't have anything to do with the points I raised.

Which is that I can't find any indication of a linear digital encoding that has been used to any appreciable extent that Meta could train on a corpus of it.

Which is why I'm struggling to understand why you're criticizing Meta? How could they realistically train a linear text model on ASL when the necessary content doesn't appear to exist?


Most hearing impaired people have never heard or don't care to use SignWriting. You are right about it's existence, for what it's worth.


right, a minor point, hearing-impaired is an ablist slang.

Instead, go by hard-of-hearing, people with hearing loss or simply, Deaf.

Also "deaf" is barely ok when use alone for the above generalized replacement however "deaf" often poorly is refers to most severe of hearing-loss as determined by differing standards of hearing loss, but capitalized "D"eaf is a direct reference to those of actively using sign language and engage within deaf culture, whether it would be American Sign Language, British Sign Language, or some 40-odd variants and different nationalities.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: