mstdn.social is one of the many independent Mastodon servers you can use to participate in the fediverse.
A general-purpose Mastodon server with a 500 character limit. All languages are welcome.

Administered by:

Server stats:

12K
active users

#imagedescription

2 posts2 participants0 posts today
Replied in thread
@Karen E. Lund 💙💛 Would you lose points for very long alt-texts/image descriptions?

I tend to describe my original images in extremely high details. Recently (for any definition of "recently" because I haven't posted a single image in over a year due to the huge effort of describing them), my alt-texts tend to reach 1,500 characters or one or a few below that. At least ca. 900 characters are actual image descriptions, sometimes up to ca. 1,400.

And that's what I consider a "short" description. Because the rest of the alt-text explains where to find the "long" description. It's in the post itself. It includes verbatim transcripts of every last bit of text anywhere within the borders of the image, readable or not. And it includes all explanations which I deem necessary for everyone to understand my images.

This long description exceeds any known arbitrarily defined character limit anywhere in the Fediverse by magnitudes. I can post such long image descriptions because the only character limit I have here on Hubzilla is the maximum size of the database field for the post text.

Yes, you've read that right. I describe each one of my original images twice.

And I must write my image descriptions that long. I don't post real-life photos, nor do I post social media screen shots. I post renderings from extremely obscure 3-D virtual worlds. Maybe one in 200,000 Fediverse users has even only heard of the technology that drives them.

Thus, I cannot assume anything in my images to be familiar to anyone out there. I can't assume that anyone out there knows what anything in my images looks like, also because my images tend to contain things which simply do not exist in real life in any shape or form.

At the same time, my impression is that especially Mastodon users expect all information which they don't have to be served on a silver platter immediately with the image description. If you mention something in your image, and somebody doesn't know what it looks like, you're obliged to describe it right away. Expecting anyone to ask you anything about your image after the fact feels like being considered ableist. I mean, you could just as well expect people to ask you to describe the whole image in the first place, right?

Same goes for explanations. Given the choice between looking stuff up themselves, being given links where they can look stuff up and being explained everything right there, right then, Mastodon users appear to greatly prefer the latter and only consider the latter really accessible.

And I have to explain a lot. When I tell you where I've taken an image, this alone takes me more characters than some of you use for a whole day's worth of alt-texts. The whole topic is so obscure that I have to explain explanations of explanations.

My personal record (warning: technically outdated image descriptions): 1,500 characters of alt-text, 1,400+ of which are image description; 60,000+ characters of long description for one image. That's about 10,000 words. It took me two full days, morning to evening, to research for and write the image descriptions. It takes a screen reader about three hours to read the long description out loud. But someone somewhere out there might be interested in all this information and displeased if they had to ask me about it to get it.

As for bilinguality, I should add to my WIP wiki on image descriptions and alt-texts that an alt-text must never include more than one image because screen readers cannot switch between languages mid-alt-text.

What I do, and I'm not even sure if that's such a good idea, is transcribe text in images that is not in English verbatim, literally letter by letter, and then translate it into English as closely as possible. I'm torn between a verbatim transcript which a screen reader cannot read out correctly and only giving a translation which would not be a verbatim transcript.

In fact, I've once had a situation in which I had to transcribe a sign in English, (broken) German and French. So I gave
  • a 100% verbatim transcript of the English text
  • a 100% verbatim transcript of the German text, all mistakes included
  • an English translation of the German text that's as close to the original as possible
  • a 100% verbatim transcript of the French text
  • an English translation of the German text that's as close to the original as possible

Nowadays, I'd simply avoid posting images with non-English text anywhere in it like the plague.

CC: @Kim Possible :kimoji_fire: @Jayne

#Long #LongPost #CWLong #CWLongPost #AltText #AltTextMeta #CWAltTextMeta #ImageDescription #ImageDescriptions #ImageDescriptionMeta #CWImageDescriptionMeta
MastodonKaren E. Lund 💙💛 (@Karen5Lund@mastodon.social)10.2K Posts, 1.75K Following, 688 Followers · My posts may be shared to other Mastodon and PixelFed instances. I DO NOT CONSENT TO SHARING OR COPYING TO NOSTR, MOSTR, THREADS OR ANY OTHER PLATFORM. Pedestrian advocate. I enjoy walking, especially in parks. Newsletter author on environment, history, infrastructure. I have an IQ in the triple digits & am not afraid to use it. If you are a writer of any kind you might prefer to follow me at https://writing.exchange/web/@Karen5Lund. #TwitterQuitter. Pronouns: She/her, Ꝥey/Ꝥem, Yes, ma'am.
Replied in thread

@djamilaknopf
alternative image description suggestion: a painting of a cat shopkeeper. Surrounded by their wares, sits an orange and white cat behind a low open cabinet similar to a TV bench. Their feet and bum have wiggled itself in the opening —because of course a cat has to occupy an opening if it can— while their paws are leaning on the top. They look quite pleased and at peace.
Their wares are a variety of jars, tins, boxes and other containers, made from a variety of materials. Some more notable objects include a big wireframe basket, perhaps containing yarn?, a drawing of a cat, a maneki-neko statue (one of those Japanese 'beckoning cats' with their paw raised; moving back and forth beckoning in any visitors), paintings of a house and of a lidded pan, something that looks like a bigger than usual brass service bell?, a framed fish, and a large ball.
Above their shop is a wooden sign with Japanese glyphs.

Replied in thread
@Christian Kent   𝘊𝘒 :\﹥ I can't really go with that proposal.

First of all, if you turn the automatic generation of alt-text on by default, then fully-abled people with tons of spoons who could write their own alt-texts with no problems will use that automatic alt-text-generating AI, too. Especially newbies who'll keep everything at default settings for quite a while. This would make alt-text quality and especially accuracy plummet.

Trust me. I've pitted LLaVA against my own hand-written image descriptions twice. In both cases, the images were renderings from very obscure 3-D virtual worlds with a resolution of only 800x533. The results were utterly devastating by my standards. Granted, the AI only had the images whereas I described the images while looking around in-world, being able to move the camera and zoom in on things and such.

So instead of turning AI-generated alt-text on by default, people should be educated about alt-text and told to write them themselves. Only if they can't, they should be told about the alt-text generator switch.

Besides, having this established on Mastodon is one thing. Trying to force the whole Fediverse to implement an image-describing AI is something else. And: It will never come to pass.

The Fediverse is far from being only Mastodon. The Fediverse is not even a) Mastodon, b) Mastodon forks and c) stuff (actually or seemingly) glued onto Mastodon as add-ons after the fact. There's stuff in the Fediverse that's vastly different from Mastodon while at the same time being in direct competition against Mastodon.

For example, you have rather minimalist stuff like snac2 or GoToSocial. I'm not sure if their devs are too keen on including an AI-based automatic alt-text generator even if the AI is external.

On the other end of the spectrum, you have Friendica, Hubzilla, (streams) and Forte. These four are vastly more powerful than Hubzilla by various degrees. And they're so modular that, in theory, it'd be possible to build an add-on that uses an AI to automatically describe images.

Now, the first issue is that neither of these four strives to be more like Mastodon. In fact, they've got their own cultures that date back to times when Mastodon wasn't even an idea. After all, Friendica is five and a half years older than Mastodon, and the others are part of the same big family tree of forks and forks of forks. And, unlike Mastodon, they're Facebook alternatives with a side of fully-featured long-form blogging and not minimalist Twitter alternatives (and Hubzilla is much much more than that even).

Their cultures are vastly different from Mastodon's and will always be. They don't aim for brevity, what with "character limits" with eight digits. They don't ask for CWs in a dedicated CW field. What's the CW field on Mastodon has been the abstract field on Friendica for almost seven years longer and on StatusNet for some nine years longer, and like Friendica, the other three use it for summaries. All four have always been able to automatically generate CWs individually for each reader, depending on a) the presence of keywords in a message and b) the keywords in that reader's "NSFW" keyword list. Something that Mastodon didn't have until October, 2022 when Mastodon 4 came out, and that is not part of Mastodon's culture because it came that late.

As for alt-text, users on all four are completely unaware of its existence at best. That's because they're completely out of touch with Mastodon and its culture. In fact, some Hubzilla users chose to leave ActivityPub switched off, and some (streams) users decided to turn it off themselves, in both cases with the explicit intent of locking Mastodon out. Of those few who have heard about alt-text in the Fediverse, most see it as just another stupid Mastodon fad that Mastodon is trying hard to force upon the rest of the Fediverse. Like that 500-character limit.

Also, especially the Hubzilla devs and the (streams) and Forte dev are not too keen on implementing Mastodon features to say the least. For now, they only do what's absolutely necessary to federate with Mastodon, but not much beyond that. These three will not make an image-describing AI add-on a priority task. Or a task at all. And don't ask for building that image-describing feature into the core. This will happen even less. Also, don't expect either of them to hard-code a connection to a power-hungry corporation like Alphabet (Google) or OpenAI into an add-on, much less into the core. Despite the fact that Hubzilla still has its old (optional, off-by-default for both hubs and channels) Twitter connector.

By the way: (streams) and Forte are the only server applications in the Fediverse with a server-wide user agent filter that's not only capable of blocking Mastodon in its entirety, but that was actually announced and advertised as having that capability while having actually been designed to keep Threads out. This should give you to think.

There's another obstacle: Where would the alt-text be stored? After all, the handling of images is, again, vastly different on these four than on Mastodon. On Mastodon, images are files attached to posts with their own alt-text data fields. And upon uploading the images, they end up in some data nirvana.

In stark contrast, Friendica, Hubzilla, (streams) and Forte all have a cloud file storage built into each account (Friendica)/channel (Hubzilla, (streams), Forte), complete with its own simple file manager and, except for Friendica, even WebDAV connectivity. It's there where image files are uploaded. And then they aren't attached to posts as files. Rather, they're embedded into posts just like into webpages or blog posts. You can have text, then an image, then more text, than another image, then more text. These four can generate something that Mastodon still staunchly refuses to even display.

Now, I know that (streams) and Forte do have a field for an image description. It can be filled out during upload or in the image editor after the upload. I'm not sure if Friendica has at least got a data field for alt-text. But I know for a fact from daily-driving it that Hubzilla does not have any data field for alt-text, not in the UI and most likely not in the database either.

In the posts, the alt-text is not added to the images by referencing that data field. The posts don't work on some murky hocus-pocus in the background. By default, they're composed in a markdown language in plain sight. Friendica uses expanded BBcode or optionally Markdown. Hubzilla uses even more expanded BBcode. (streams) and Forte can use the same expanded BBcode as Hubzilla and Markdown and HTML within the same post. And the alt-text is always part of the image-embedding code.

And this is also the only place on Hubzilla where alt-text can come into play. This means that if Hubzilla wanted to include automatic alt-text generation, the alt-text would have to a) be generated whenever an image is embedded into a post while composing the post (= you embed the same image file a hundred times, and a hundred times will the AI generate a new alt-text for it from scratch) and then b) be woven into the image-embedding BBcode.

Speaking of image storage, WriteFreely, a specialised long-form blogging server application and basically a Medium alternative, doesn't even have that. You can embed images into posts on WriteFreely, but only as long as you can store them someplace that allows for hotlinking because you will have to hotlink your images from there. But if WriteFreely doesn't handle images itself at all, if images in WriteFreely posts don't actually go through WriteFreely until after the post is sent (namely when someone views the post for the first time), it can't use an AI to automatically describe them either.

Unlike WriteFreely, Plume does have its own dedicated image storage. However, while Plume is not dead, its development has come to a complete halt for the foreseeable future because its devs don't have any time to work on it. And even if they got back to it, they'd probably have more important things to work on after several years of neglect.

And then there is software that is actually dead, that was officially pronounced dead by its own last maintainers, but that still has running servers. Calckey. Firefish. /kbin. But if it's dead, it won't even get security patches anymore, much less new features such as automatic AI-driven alt-text generation.

CC: @James Edwards @Danny Boling ☮️

#Long #LongPost #CWLong #CWLongPost #FediMeta #FediverseMeta #CWFediMeta #CWFediverseMeta #AltText #AltTextMeta #CWAltTextMeta #ImageDescription #ImageDescriptions #ImageDescriptionMeta #CWImageDescriptionMeta #AIVsHuman #HumanVsAI #CW #CWs #CWMeta #ContentWarning #ContentWarnings #ContentWarningMeta #Fediverse #MastodonCulture #FediverseCulture #snac #snac2 #GoToSocial #Friendica #Hubzilla #Streams #(streams) #Forte #Plume #WriteFreely #Calckey #Firefish #/kbin
hub.netzgemeinde.euNetzgemeinde/Hubzilla
Replied in thread
@James Edwards This is highly ableist.

This would mean that people who are constantly low on spoons due to Long Covid, Post-Covid, ME/CFS or the like would not be allowed to post images. Ever.

This would mean that autistic people who simply cannot turn images into words (no matter how often your likes tell them to "just try harder", they simply lack the ability) would not be allowed to post images. Ever.

This would mean that blind people would not be allowed to post images. Ever. And yes, they do post images.

CC: @Danny Boling ☮️ @Christian Kent   𝘊𝘒 :\﹥

#Long #LongPost #CWLong #CWLongPost #ImageDescription #ImageDescriptions #ImageDescriptionMeta #CWImageDescriptionMeta #LongCovid #PostCovid #ME/CFS #Autism #Autistic #Blind #VisuallyImpaired #Disabled #Disability #Ableist #Ableism
hub.netzgemeinde.euNetzgemeinde/Hubzilla
Replied in thread
@Danny Boling ☮️ @Christian Kent   𝘊𝘒 :\﹥ AI will never be a 100% drop-in replacement for human image descriptions in absolutely all cases. AI will never be able to describe absolutely all images 100% accurately.

That would require each image-describing AI to be absolutely omniscient, all the way to the tiniest and most obscure niche topics. It would literally require each image-describing AI to have knowledge which otherwise only exists in the head of whoever posts the image.

#Long #LongPost #CWLong #CWLongPost #ImageDescription #ImageDescriptions #ImageDescriptionMeta #CWImageDescriptionMeta #AIVsHuman #HumanVsAI
hub.netzgemeinde.euNetzgemeinde/Hubzilla
Replied in thread
@Gemma ⭐🔰🇺🇸 🇵🇭 🎐 I do 1 and 3.

1 to such extents that the actual alt-text only contains a short description where "short" means anything between ca. 900 and ca. 1,400 characters. The long description goes into the post, and it regularly measures several tens of thousands of characters. Also, I don't describe what's in the image as I can see it in the image, I describe what's in the image as I can see it at the place where the image was made, i.e. at an almost infinitely higher resolution and, if need be, with the ability of looking around obstacles.

Someone somewhere out there might be interested in these details and at the same time consider having to ask for further descriptions lazy or maybe even ableist.

What I no longer do, however, is describe images within my image at more details than visible in the place where I've taken the image. In one of my last image descriptions, I would otherwise have had to describe not only multiple images in my image, but dozens of images in one image in my image and probably even more images in these images.

3 to such extents that I even transcribe text that's unreadable in the image, but that I can read at the place where the image was made. Also, I've once had a sign (unreadable of course) in English, French and rather broken German. I transcribed all three languages character by character, and I translated the French and the German text into English right after transcribing each one of it. Another reason why my long image descriptions are so long. This irritates screen readers because they can't switch languages mid-text, but if 100% verbatim transcripts are the rule, then so be it.

The only thing I no longer do regarding this is transcribe all-caps as all-caps because screen readers may or may not misinterpret them. Also, I don't transcribe Roman numbers as such.

#Long #LongPost #CWLong #CWLongPost #AltText #AltTextMeta #CWAltTextMeta #ImageDescription #ImageDescriptions #ImageDescriptionMeta #CWImageDescriptionMeta #Transcripts
hub.netzgemeinde.euNetzgemeinde/Hubzilla
Replied in thread
@Baranduin
Oh does this feel like my inner monologue when I post a photo. It is a bummer that it, at times, prevents me from posting more photos, but I hope this make me a little more quality over quantity.

I actually keep entire categories of things out of my images because I can't describe them up to my own standards. This includes realistic buildings. I would first have to look up loads of architectural terms to describe all details of a building, and then I would have to explain each and every one of these architectural buildings in a way that absolute laypeople understand my image description without ever having to ask me anything or look anything up themselves.

The last time I posted an image with a building was this post. I actually went around and looked for a nice motive for a new image post for quite a while. There was one harbour scene which I thought looked spectacular enough to show, but which was impossible to describe. So I fell back to this motive. I thought it's not too bland, not too simple and at the same time not too complex. Besides, the one building in the image is totally unrealistic and without all the tiny details that would make up a realistic building.

And then I ended up taking some 30 hours over two days to describe the image in over 60,000 characters. The building alone took up some 40,000 or so. This is still the longest image description in the whole Fediverse, I think. Here is the image describing log thread.

My last image post before that was this one with still over 25,000 characters of description for one image, and I consider it outdated slop.

It was the last time that I described an image in my image with more details than visible in the original of that image itself. And that's where I got sloppy. I completely forgot to transcribe what's written on the license plate above the office door of the motel in that image in my image. And I couldn't be bothered to give detailed descriptions of the two 1957 Chevy Bel Airs parked in front of the motel because I really wanted to get that description done. In the actual image, all of this is sub-pixel-sized. You wouldn't know it's even there if I didn't mention it. I did describe the motel, but it's a fairly simple building, and I decided against describing what's visible through the windows with open blinds from the camera angle in the image in my image.

In the next image, the one with 60,000+ characters of description, I stopped describing images in the image beyond what I can see in the place where the image itself was taken. That was because one image is a destination preview image on a teleporter. The destination is a kind of teleport hub. The preview actually (if only barely so) shows over 300 single-destination teleporters, a few dozen of them with their own preview images.

So I teleported to that hub to describe it in detail. And I looked at the teleporters and their preview images. Turned out, not only do these preview images pretty much all have text in them and not necessarily few bits of text, but some of them actually have images within themselves again.

I would have had to describe that image in my image, dozens of images in that image in my image and a number of images in these images in that image in my image. For each of the latter, I would have had to teleport three times from the place that I originally wanted to describe. I would also have had a whole lot more text to transcribe. All on a sub-pixel scale several times over.

Not only would that have been a humongous task, but more importantly, it would have inflated my image description and my whole post to more than 100,000 characters. Mastodon would probably have rejected my post for being too long. And this would have rendered the whole effort futile. In the few places in the Fediverse that would still have accepted my post, nobody cares for image descriptions.

AI certainly can't get inside my brain well enough to write accurate descriptions. Even if it could would I? hmmm.

I've only used AI to describe images twice. And in both cases, that was to show just how bad AI is at describing images about an extremely obscure and quickly changing niche topic at the level of accuracy and detail which I deem necessary for that topic.

I guess one problem that you're facing is that next to nobody in the Fediverse can even grasp what you're thinking about, what you're taking into consideration for your image descriptions. That's why you got next to no feedback upon your first comment in this thread.

I have one advantage here: What you're pondering, I have actually done. If I feel like people won't understand what I'm thinking about, I point them at one or several of my actual image posts, and/or I post a quote from one of my actual image descriptions. Still, almost nobody actually goes and reads through any of my image descriptions, but I guess they get the gist, especially when I post snippets from my actual descriptions.

CC: @Icarosity

#Long #LongPost #CWLong #CWLongPost #AltText #AltTextMeta #CWAltTextMeta #ImageDescription #ImageDescriptions #ImageDescriptionMeta #CWImageDescriptionMeta
hub.netzgemeinde.euUniversal Campus: The mother of all mega-regionsOpenSim's famous Universal Campus and a picture of its main building; CW: long (62,514 characters, including 1,747 characters of actual post text and 60,553 characters of image description)
Replied in thread
@ScotsBear 🏴󠁧󠁢󠁳󠁣󠁴󠁿 Just for me to be on the safe side: What are your minimum requirements for alt-texts and image descriptions so you refrain from sanctioning a user?

Full, to-the-point adherence to the Accessible Social guidelines, the Cooper Hewitt guidelines, Veronica With Four Eyes' various guidelines etc., even though they contradict each other?

Do you demand image descriptions be detailed and informative enough so that nobody will ever have to ask the poster about explanations and/or details because they're all already in the descriptions, no matter how niche and obscure the content of the image is?

If there is already a lengthy image description in the post itself (imagine all character limits you know in the Fediverse; it's longer than all of them by magnitudes), do you still demand there be another description in the alt-text, even though the alt-text actually points the user to the description in the post, because there absolutely must be a sufficiently detailed and accurate image description in the alt-text, full stop?

In fact, do you sanction image descriptions in general or alt-texts in particular if you think they are too long? For example, if you stumble upon an image post from me that has a "short" image description of 1,400 characters in the alt-text and a "long" image description of over 60,000 characters in the post itself (and I've actually posted such a thing into the Fediverse; here's the link to the source), will you demand I discard two days and some 30 hours of work, delete the long description and cut the short description down to no more than 200 characters? Maybe even while still retaining the same amount of information? Lest you have me dogpiled and mass-blocked or worse?

By the way, I think I've gathered a whole lot of experience and knowledge about describing images generally and specifically for the Fediverse, and I also see the high level of detail in my image descriptions as fully justified.

#Long #LongPost #CWLong #CWLongPost #FediMeta #FediverseMeta #CWFediMeta #CWFediverseMeta #AltText #AltTextMeta #CWAltTextMeta #ImageDescription #ImageDescriptions #ImageDescriptionMeta #CWImageDescriptionMeta
www.accessible-social.comWriting Image DescriptionsTips on how to write effective image descriptions to make visuals accessible.
Replied in thread
@Icarosity It's similar for me, only that I always put a gigantic effort into describing my own images twice, once not exactly briefly in the alt-text and once with even more details in the post itself. Sometimes I find an interesting motive, but when I start thinking about how to describe it, I don't even render an image because it isn't worth doing so if I can't post it.

I haven't posted a new image in almost a year. In fact, I've got a series of fairly simple images for which I've started writing the descriptions late last year, and I'm still not done. So much about "it only takes a few seconds".

Before someone suggests I could use Altbot: I'm not even sure if it'll work with Hubzilla posts. And besides, no AI on this planet is fit for the task of properly, appropriately and accurately describing the kind of images that I post.

@Baranduin And then there's me who has managed to describe one image in a bit over ten thousand words last year. Good thing I have a post character limit of over 16.7 million. And I actually limited myself this time: I did not describe images within my image in detail, in stark contrast to about two years ago when I described a barely visible image in an image in well over 4,000 characters of its own, and that wasn't the only image within that image that I described.

CC: @Logan 5 and 999 others

#Long #LongPost #CWLong #CWLongPost #AltText #AltTextMeta #CWAltTextMeta #ImageDescription #ImageDescriptions #ImageDescriptionMeta #CWImageDescriptionMeta
MastodonIcarosity (@nancywisser@mastodon.social)5.71K Posts, 72 Following, 463 Followers · mostly harmless
Observer: always looking and curious about overlooked things, especially plants, especially native plants. I take a lot of pictures. I have a cat and I grow slipper orchids. Oh yeah also—I’m an old
Just a visitor here—Tumblr is my home and there I am geopsych 
death trap clad happily
Replied in thread
@Logan 5 and 999 others First of all: You must never put line breaks into alt-text. Ever. (https://www.tpgi.com/short-note-on-coding-alt-text/, https://joinfediverse.wiki/Draft:Captions#Line_breaks)

Besides, that will certainly not be the day that I'll post my first image after more than a year.

It's tedious enough to properly describe my original images at the necessary level of detail, and one image takes me many hours to describe, sometimes up to two full days, morning to evening. Not joking here. I certainly won't put extra effort into turning at least the 900 characters of "short" description that go into the alt-text into a poem. And I definitely will not also turn the additional 20,000, 40,000, 60,000 characters of long description that go into the post into a poem as well. (And yes, I can post 60,000+ characters in one go, and I have done so in the past. My character limit is 16,777,215.)

#Long #LongPost #CWLong #CWLongPost #AltText #AltTextMeta #CWAltTextMeta #ImageDescription #ImageDescriptions #ImageDescriptionMeta #CWImageDescriptionMeta
TPGi · Short note on coding alt text - TPGiThe other day, in relation to a github comment, I was asked by my friend Mike[tm]Smith “Can alt have line breaks in it or does that do weird things to...
Replied in thread
@Freeeeeeeeeeeeeeeeeeeee :ablobfoxbongohyper:‮‮‮‮‮‮

Here are some more:

  • Always keep in mind that the primary target audience of your alt-text are blind or visually-impaired people. Do not write your alt-text for sighted people first and foremost.
  • Do not start with "picture of", "image of" or "photo of". "Picture of" and "image of" are redundant because screen readers announce images anyway, and "photo of" is redundant because digital photographs are the default for online images. Do mention all other media, though.
  • Do not add line breaks. Line breaks are not a standard part of alt-text, and they are useless for blind or visually-impaired people because screen readers won't read them out aloud. Not all frontends support them, and they may cause nasty side-effects.
  • Do not add quotation marks from your keyboard. Quotation marks are not a standard part of alt-text, and they are useless for blind or visually-impaired people because screen readers won't read them out aloud. Not all frontends support them, and they may cause nasty side-effects or even severe breakage.
  • Do not add emoji or fancy Unicode characters. They will disturb the reading flow for screen reader users because screen readers will call all of them by name.
  • Do not add hashtags. They won't be parsed as hashtags, and links don't work in alt-text anyway.
  • Describe dimensions and distances relatively to what other dimensions and/or distances are in the image.
  • Do describe colours. Not everyone who needs a screen reader was born completely blind. Some people used to be sighted, and they have lost their eyesight later in their lives, but they still remember colours. But always describe colours using brightness, saturation and one or two familiar basic colours: red, blue, yellow, green, orange, purple, violet, pink, brown, gold, silver, black, grey, white. Do not simply use the name of a specific hue or shade. Not everyone who isn't sighted knows what it looks like. If you absolutely have to mention the name, do describe the colour afterwards.
  • Think about what someone who comes across your post knows and what they don't know. Don't assume that everyone is familiar with the same things as you.
  • Transcribe all text within the borders of the image 100% verbatim. (Sorry, I don't have any steadfast rules for text in foreign languages, for misspellings or other mistakes or for text that's unreadable in the image, but that you can read elsewhere and therefore transcribe.)
  • Do not use all-caps for words or entire sentences. Some screen readers will spell them out, letter by letter, even if that isn't your intention.
  • Do not use technical terms and/or jargon, no matter what you think people are or should be familiar with.
  • Do not mention a person's race. If at all, mention their skin tone (light-skinned, medium-light-skinned, medium-skinned, medium-dark-skinned, dark-skinned).
  • Do not mention a person's gender if you have to assume it. Only do it when a person actually performs their gender in an image, or that person has personally declared their gender to you, or you have a source of that person publicly declaring their gender, or you have a similarly rock-solid, undeniable source for that person's gender.
  • Do not add any information that is not available in the post text and/or the image itself. This includes explanations, image credits and license information. Not everyone can access alt-text. Some people have physical disabilities that prevent them from accessing alt-text. Any information that's only available in alt-text is inaccessible and lost to them.

#Long #LongPost #CWLong #CWLongPost #AltText #AltTextMeta #CWAltTextMeta #ImageDescription #ImageDescriptions #ImageDescriptionMeta #CWImageDescriptionMeta
hub.netzgemeinde.euNetzgemeinde/Hubzilla
Replied in thread
@Georgiana Brummell Isn't the A2I post quite a bit out-dated?

Some two years ago, I've read about screen readers not supporting more than 200 characters in alt-text. But people who actually use screen readers told me that all screen reader software available has long been upgraded to support an infinite number of characters. And next to nobody uses old versions with a 200-character limit anymore.

And now I often see posts and articles, even recent ones, mention a hard limit of 125 characters for alt-text in screen readers. This must actually be leftover information from the mid-2010s at best.

Case in point: I've never seen anyone in the Fediverse being criticised for what would be absolutely excessively long alt-text by Web design standards. Proof enough that screen readers can easily handle 800 or 1,000 or more characters of alt-text.

As far as I'm informed, the only issue is that screen readers cannot navigate alt-texts, i.e. you cannot rewind to a certain point within an alt-text and have it re-read from there. You can only jump back to the beginning of the alt-text and have the whole alt-text re-read. The longer an alt-text is, the less convenient this is.

By the way: I've started working on an entire wiki on how to describe images properly and write image descriptions in general and alt-texts specifically for the Fediverse. It will take quite a number of existing guides and how-tos and the like into consideration and link to them. It will also take both Mastodon's culture and the special perks of the various places in the Fediverse outside of Mastodon into consideration. When guides contradict each other, I'll mention that as well.

It has to be a wiki because it will contain so much information that it simply wouldn't fit onto one page anymore. Also, I want to be able to point people at certain aspects of describing images or writing alt-texts such as how colours should be described, why people's races should never be mentioned and why explanations do not belong into alt-texts. I don't want to tell them to scroll down to a certain paragraph. I want to show them one page that specialises in that particular topic.

I'm not sure if that's utter overkill, if that'll stand in the way of just "doing it" and actually drive people away from describing images. But in my opinion, someone has to tell people how to do it properly.

#Long #LongPost #CWLong #CWLongPost #FediMeta #FediverseMeta #CWFediMeta #CWFediverseMeta #AltText #AltTextMeta #CWAltTextMeta #ImageDescription #ImageDescriptions #ImageDescriptionMeta #CWImageDescriptionMeta #A11y #Accessibility
hub.netzgemeinde.euNetzgemeinde/Hubzilla

@yuriposting so pretty,,

Image description: It’s a digital illustration of a cute noontime scene featuring a lady wearing a white gown and a maid in a long black dress dozing off in a lavish victorian style room. The picture is watermarked “Illust by Rrrrrrice.”

A draft blows from the window sending the curtains billowing toward the lady and her maid with sheer fabric fingers that wrap around them, carrying with it the scent of hickory and currants. Warm rays of sunlight filter in. Having finished reading a book for now, the lady rests in an olive colored high-backed chair with a brown tasseled blanket and cozy white pillow. Her maid, brunette with her hair kept in a bun, is doubled over on a red carpet and pressed against the lady’s legs. She lies her head in her lady’s lap and allows their fingers to intertwine. There is no other option in this situation—this is what must occur.

Precious metal and ceramic furnishings fill the room: china, a clock, and vases resting on the mantel. To the side of the chair is a table holding a porcelain lamp with a tan tasseled shade. On the wall are some candlesticks in gold holders and, of course, there are several portraits on the wall, presumably other members of the house, dressed in fancy clothes, which are held in gold frames.

Overall it has a serene but very wealthy aesthetic to it. Stylistically it has a lot of big blocks of color with most of the detailed rendering concentrated on the focal points like the lady’s face and cascading blonde hair.

#ImageDescription #AltText #Alt4You

RE: https://sakurajima.moe/@yuriposting/114700242549863247

Danbooru tags: 2girls apron artist_name black_dress blanket blonde_hair book brown_hair candle candlestand chair closed_eyes closed_mouth curtains dress fireplace hair_bun highres indoors kneeling long_hair maid maid_apron master_and_servant multiple_girls open_book original painting_(object) picture_frame puffy_sleeves reaching rrr_(reason) scrunchie sitting sleeping table vase victorian victorian_maid waist_apron white_apron white_dress white_scrunchie yuri
Sakurajima (桜島)Yuri Posting (@yuriposting@sakurajima.moe)Attached: 1 image Artist: rrr (reason) Media: original Source: https://www.pixiv.net/en/artworks/121425842
Replied in thread
@nihilistic_capybara LLMs aren't omniscient, and they will never be.

If I make a picture on a sim in an OpenSim-based grid (that's a 3-D virtual world) which has only been started up for the first time 10 minutes ago, and which the WWW knows exactly zilch about, and I feed that picture to an LLM, I do not think the LLM will correctly pinpoint the place where the image was taken. It will not be able to correctly say that the picture was taken at <Place> on <Sim> in <Grid>, and then explain that <Grid> is a 3-D virtual world, a so-called grid, based on the virtual world server software OpenSimulator, and carry on explaining what OpenSim is, why a grid is called a grid, what a region is and what a sim is. But I can do that.

If there's a sign with three lines of text on it somewhere within the borders of the image, but it's so tiny at the resolution of the image that it's only a few dozen pixels altogether, then no LLM will be able to correctly transcribe the three lines of text verbatim. It probably won't even be able to identify the sign as a sign. But I can do that by reading the sign not in the image, but directly in-world.

By the way: All my original images are from within OpenSim grids. I've probably put more thought into describing images from virtual worlds than anyone. And I've pitted my own hand-written image description against an AI-generated image description of the self-same image twice. So I guess I know what I'm writing about.

CC: @🅰🅻🅸🅲🅴  (🌈🦄) @nihilistic_capybara

#Long #LongPost #CWLong #OpenSim #OpenSimulator #Metaverse #VirtualWorlds #CWLongPost #ImageDescription #ImageDescriptions #ImageDescriptionMeta #CWImageDescriptionMeta #AI #LLM #AIVsHuman #HumanVsAI
hub.netzgemeinde.euNetzgemeinde/Hubzilla