An idea,
Scott Berry
scott at citlink.net
Wed Jul 27 10:25:06 EDT 2005
This really doesn't pertain to how Gnopernicus works. But one
question I really would like to through around out here is why are
the keystrokes so difficult. I know you can change key bindings in
Gnopernicus but it looked like a guy with 20 years of experience
would have to do it. Maybe this has changed in recent versions but I
remember when I started using Gnopernicus wow! Difficult to learn.
At 08:05 AM 7/27/2005, you wrote:
>Hi, Lorenzo:
>
>Others have responded with reference to what an X server does, and
>doesn't do. I want to respond to two other particular points from your
>message.
>
>Lorenzo Taylor writes:
> > -----BEGIN PGP SIGNED MESSAGE-----
> > Hash: SHA1
> >
> > ... Gnopernicus, for example, is using libraries that
> > rely on certain information ent by the underlying application libraries.
> > Unfortunately, this implementation causes only some apps to speak
> while others
> > which use the same widgets but whose libraries don't send messages to the
> > accessibility system will not speak.
>
>This is only partially correct. Any applications using those "same
>widgets," as you put it, will speak. There are no exceptions.
>
>What causes them to not speak is that the properties required to make
>them speak have not been supplied. So, Gnopernicus is getting an empty
>string to renderd, which I suppose it dutifully renders as silence.
>
>Fortunately, these are open source applications and we don't need an
>advocacy campaign to resolve these kinds of problems. A solid example of
>this at work is the Gnome Volume Control. It was written with gtk2, but
>the developers did not supply all the relevant property data. So, a
>blind programmer came along one weekend, fixed it, and submitted the
>patch which has shipped with the rest of Gnome Volume Control ever
>since.
>
>Now the next point ...
>
> > But it occurs to me that X is simply a
> > protocol by which client applications send messages to a server
> which renders
> > the proper text, windows, buttons and other widgets on the
> screen. I believe
> > that a screen reader that is an extension to the X server itself,
> (like Speakup
> > is a set of patches to the kernel) would be a far better
> solution, as it could
> > capture everything sent to the server and correctly translate it
> into humanly
> > understandable speech output without relying on "accessibility
> messages" being
> > sent from the client apps.
>
>
>As other have pointed out, there's nothing to be gained by speaking RGB
>values at some particular X-Y mouse coordinate location. But, I'm sure
>that's not what you really intend. If I interpret you correctly you're
>suggesting some kind of mechanism whereby a widget of some kind can be
>reliably identified and assigned values that the screen reader can
>henceforth utter. This is the approach with Windows OSM that has been
>used over the past decade, and it's what allows screen readers, like
>JFW, to develop interfaces based on scripts. For instance, Take widget
>number 38,492 and call it "volume slider," and speak it before anything
>else on screen when it shows up on screen, and facilitate the method
>that will allow user to use up and down arrow to change it's value,
>etc., etc.
>
>It is arguable, and has been cogently argued over the past 18 months,
>that the failure of the original Desktop Accessibility Architecture
>promoted by Sun and Gnome was to not provide such mechanisms. A great
>part of the intent of the Orca screen reader proof of concept was to
>provide exactly this kind of functionality. I believe this is now being
>addressed, though I'm not aware any code for newer Gnopernicus (or post
>Gnopernicus) readers is yet released. However, I do fully expect that
>Gnopernicus is not the last word in desktop screen readers.
>
> Janina
>
>_______________________________________________
>Speakup mailing list
>Speakup at braille.uwo.ca
>http://speech.braille.uwo.ca/mailman/listinfo/speakup
More information about the Speakup
mailing list