Jump to content

Edit filter log

Details for log entry 31205906

09:25, 1 November 2021: 104.218.65.8 (talk) triggered filter 636, performing the action "edit" on Screen reader. Actions taken: none; Filter description: Unexplained removal of sourced content (examine)

Changes made in edit

=== Command-line (text) ===
=== Command-line (text) ===


In early [[operating system]]s, such as [[MS-DOS]], which employed [[command-line interface]]s (<abbr>CLI</abbr>s), the screen display consisted of [[Character (computing)|characters]] [[Memory-mapped I/O|mapping]] directly to a [[screen buffer]] in [[Computer memory|memory]] and a [[Cursor (computers)|cursor]] position. Input was by keyboard. All this information could therefore be obtained from the system either by [[hooking]] the flow of information around the system and reading the screen buffer or by using a standard hardware output socket<ref>{{cite web
In early operating systems, such as MS-DOS, which employed command-line interfaces (<abbr>CLI</abbr>s), the screen display consisted of characters mapping directly to a screen buffer in memory and a cursor position. Input was by keyboard. All this information could therefore be obtained from the system either by hooking the flow of information around the system and reading the screen buffer or by using a standard hardware output socket and communicating the results to the user.
| title=Talking Terminals. BYTE, September 1982
| url=http://www.edstoffel.com/david/talkingterminals.html
| access-date=September 7, 2006 |archive-url = https://web.archive.org/web/20060625225004/http://www.edstoffel.com/david/talkingterminals.html <!-- Bot retrieved archive --> |archive-date = June 25, 2006}}</ref> and communicating the results to the user.


In the 1980s, the Research Centre for the Education of the Visually Handicapped (<abbr>RCEVH</abbr>) at the [[University of Birmingham]] developed Screen Reader for the [[BBC Micro]] and <!--abbr?-->NEC Portable.<ref>Paul Blenkhorn, "The <abbr>RCEVH</abbr> project on micro-computer systems and computer assisted learning", British Journal of Visual Impairment, 4/3, 101-103 (1986). [http://www.visugate.biz/bjvi/1986/autumn1986.html#RCEVH Free HTML version at Visugate].</ref><ref>{{cite web | title=Access to personal computers using speech synthesis. RNIB New Beacon No.76, May 1992 | date=March 3, 2014| url=http://www.rnib.org.uk/information-everyday-living-using-technology-beginners-guides/beginners-guide-assistive-technology}}</ref>
In the 1980s, the Research Centre for the Education of the Visually Handicapped (<abbr>RCEVH</abbr>) at the University of Birmingham developed Screen Reader for the BBC Micro and <!--abbr?-->NEC Portable.


=== Graphical ===
=== Graphical ===
==== Off-screen models ====
==== Off-screen models ====


With the arrival of [[graphical user interface]]s (<abbr>GUI</abbr>s), the situation became more complicated. A <abbr>GUI</abbr> has characters and graphics drawn on the screen at particular positions, and therefore there is no purely textual representation of the graphical contents of the display. Screen readers were therefore forced to employ new low-level techniques, gathering messages from the [[operating system]] and using these to build up an "off-screen model", a representation of the display in which the required text content is stored.<ref>According to "[ftp://service.boulder.ibm.com/sns/sr-os2/sr2doc/guitalk.txt Making the <abbr>GUI</abbr> Talk]" (by Richard Schwerdtfeger, ''BYTE'' December 1991, p. 118-128), the first screen reader to build an off-screen model was outSPOKEN.</ref>
With the arrival of graphical user interfaces (<abbr>GUI</abbr>s), the situation became more complicated. A <abbr>GUI</abbr> has characters and graphics drawn on the screen at particular positions, and therefore there is no purely textual representation of the graphical contents of the display. Screen readers were therefore forced to employ new low-level techniques, gathering messages from the operating system and using these to build up an "off-screen model", a representation of the display in which the required text content is stored.


For example, the operating system might send messages to draw a command button and its caption. These messages are intercepted and used to construct the off-screen model. The user can switch between controls (such as buttons) available on the screen and the captions and control contents will be read aloud and/or shown on refreshable [[Braille]] display.
For example, the operating system might send messages to draw a command button and its caption. These messages are intercepted and used to construct the off-screen model. The user can switch between controls (such as buttons) available on the screen and the captions and control contents will be read aloud and/or shown on refreshable Braille display.


Screen readers can also communicate information on menus, controls, and other visual constructs to permit blind users to interact with these constructs. However, maintaining an off-screen model is a significant technical challenge; hooking the low-level messages and maintaining an accurate model are both difficult tasks.{{citation needed|date=January 2015}}
Screen readers can also communicate information on menus, controls, and other visual constructs to permit blind users to interact with these constructs. However, maintaining an off-screen model is a significant technical challenge; hooking the low-level messages and maintaining an accurate model are both difficult tasks.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}}


==== Accessibility APIs ====
==== Accessibility APIs ====


Operating system and application designers have attempted to address these problems by providing ways for screen readers to access the display contents without having to maintain an off-screen model. These involve the provision of alternative and accessible representations of what is being displayed on the screen accessed through an <abbr>[[Application programming interface|API]]</abbr>. Existing <abbr>API</abbr>s include:
Operating system and application designers have attempted to address these problems by providing ways for screen readers to access the display contents without having to maintain an off-screen model. These involve the provision of alternative and accessible representations of what is being displayed on the screen accessed through an <abbr>API</abbr>. Existing <abbr>API</abbr>s include:
* Android Accessibility Framework<ref>[https://developer.android.com/training/accessibility/index.html Implementing Accessibility on Android].</ref>
* Android Accessibility Framework
* Apple Accessibility <abbr>API</abbr><ref>[https://developer.apple.com/documentation/Accessibility/Reference/AccessibilityLowlevel/index.html Apple Accessibility <abbr>API</abbr>].</ref>
* Apple Accessibility <abbr>API</abbr>
* [[AT-SPI]]
* AT-SPI
* I Accessible
* [[IAccessible2]]
* [[Microsoft Active Accessibility]] (<abbr>MSAA</abbr>)
* Microsoft Active Accessibility (<abbr>MSAA</abbr>)
* [[Microsoft UI Automation]]
* Microsoft UI Automation
* Java Access Bridge
* [[Java Access Bridge]]<ref>{{cite web|url=http://java.sun.com/products/accessbridge/|title=Oracle Technology Network for Java Developers – Oracle Technology Network – Oracle}}</ref>


Screen readers can query the operating system or application for what is currently being displayed and receive updates when the display changes. For example, a screen reader can be told that the current focus is on a button and the button caption to be communicated to the user. This approach is considerably easier for the developers of screen readers, but fails when applications do not comply with the accessibility <abbr>API</abbr>: for example, [[Microsoft Word]] does not comply with the <abbr>MSAA</abbr> <abbr>API</abbr>, so screen readers must still maintain an off-screen model for Word or find another way to access its contents.{{citation needed|date=January 2015}} One approach is to use available operating system messages and application object models to supplement accessibility <abbr>API</abbr>s.
Screen readers can query the operating system or application for what is currently being displayed and receive updates when the display changes. For example, a screen reader can be told that the current focus is on a button and the button caption to be communicated to the user. This approach is considerably easier for the developers of screen readers, but fails when applications do not comply with the accessibility <abbr>API</abbr>: for example, Microsoft Word does not comply with the <abbr>MSAA</abbr> <abbr>API</abbr>, so screen readers must still maintain an off-screen model for Word or find another way to access its contents.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} One approach is to use available operating system messages and application object models to supplement accessibility <abbr>API</abbr>s.


Screen readers can be assumed to be able to access all display content that is not intrinsically inaccessible. Web browsers, word processors, icons and windows and email programs are just some of the applications used successfully by screen reader users. However, according to some users,{{Who|date=January 2015}} using a screen reader is considerably more difficult than using a GUI, and many applications have specific problems resulting from the nature of the application (e.g. animations) or failure to comply with accessibility standards for the platform (e.g. Microsoft Word and Active Accessibility).{{citation needed|date=January 2015}}
Screen readers can be assumed to be able to access all display content that is not intrinsically inaccessible. Web browsers, word processors, icons and windows and email programs are just some of the applications used successfully by screen reader users. However, according to some users,{{Who|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} using a screen reader is considerably more difficult than using a GUI, and many applications have specific problems resulting from the nature of the application (e.g. animations) or failure to comply with accessibility standards for the platform (e.g. Microsoft Word and Active Accessibility).{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}}


=== Self-voicing programs and applications ===
=== Self-voicing programs and applications ===


Some programs and applications have voicing technology built in alongside their primary functionality. These programs are termed [[self-voicing]] and can be a form of [[assistive technology]] if they are designed to remove the need to use a screen reader.{{citation needed|date=January 2015}}
Some programs and applications have voicing technology built in alongside their primary functionality. These programs are termed self-voicing and can be a form of assistive technology if they are designed to remove the need to use a screen reader.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}}


=== Cloud-based ===
=== Cloud-based ===


Some telephone services allow users to interact with the internet remotely. For example, TeleTender can read web pages over the phone and does not require special programs or devices on the user side.{{citation needed|date=January 2015}}
Some telephone services allow users to interact with the internet remotely. For example, TeleTender can read web pages over the phone and does not require special programs or devices on the user side.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}}


=== Web-based ===
=== Web-based ===


A relatively new development in the field is web-based applications like Spoken-Web that act as web portals, managing content like news updates, weather, science and business articles for visually-impaired or blind computer users.{{citation needed|date=January 2015}} Other examples are ReadSpeaker or [[BrowseAloud]] that add [[text-to-speech]] functionality to web content.{{citation needed|date=January 2015}} The primary audience for such applications is those who have difficulty reading because of learning disabilities or language barriers.{{citation needed|date=January 2015}} Although functionality remains limited compared to equivalent desktop applications, the major benefit is to increase the accessibility of said websites when viewed on public machines where users do not have permission to install custom software, giving people greater "freedom to roam".{{citation needed|date=January 2015}}
A relatively new development in the field is web-based applications like Spoken-Web that act as web portals, managing content like news updates, weather, science and business articles for visually-impaired or blind computer users.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} Other examples are ReadSpeaker or BrowseAloud that add text-to-speech functionality to web content.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} The primary audience for such applications is those who have difficulty reading because of learning disabilities or language barriers.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} Although functionality remains limited compared to equivalent desktop applications, the major benefit is to increase the accessibility of said websites when viewed on public machines where users do not have permission to install custom software, giving people greater "freedom to roam".{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}}


With the development of [[smartphone]]s, the ability to listen to written documents (textual web content, <abbr>[[Portable Document Format|PDF]]</abbr> documents, e-mails etc.) while driving or during a similar activity in the same way that listening to music, will benefit a much broader audience than visually-impaired people.{{citation needed|date=January 2015}} The best-known examples are [[Siri (software)|Siri]] for [[iOS]], and [[Google Now]] and [[Iris (software)|Iris]] for [[Android (operating system)|Android]]. With the release of the [[Samsung Galaxy S III|Galaxy S III]], [[Samsung]] also introduced a similar [[intelligent personal assistant]] called [[S Voice]]. On the [[BlackBerry 10 operating system voice control|BlackBerry 10 operating system]], their [[BlackBerry Z30|Z30 smartphone]] also features spoken interaction features, which are similar to the other [[mobile operating system]]s.{{citation needed| date= October 2013| reason= Similarity unverified.}}
With the development of smartphones, the ability to listen to written documents (textual web content, <abbr>PDF</abbr> documents, e-mails etc.) while driving or during a similar activity in the same way that listening to music, will benefit a much broader audience than visually-impaired people.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} The best-known examples are Siri for iOS, and Google Now and Iris for Android. With the release of the Galaxy S III, Samsung also introduced a similar intelligent personal assistant called S Voice. On the BlackBerry 10 operating system, their Z30 smartphone also features spoken interaction features, which are similar to the other mobile operating systems.{{citation needed| date= {{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}| reason= Similarity unverified.}}


This functionality depends on the quality of the software but also on a logical structure of the text. Use of headings, punctuation, presence of alternate attributes for images, etc. is crucial for a good vocalization. Also a web site may have a nice look because of the use of appropriate two dimensional positioning with CSS but its standard linearization, for example, by suppressing any CSS and Javascript in the browser may not be comprehensible.{{citation needed|date=January 2015}}
This functionality depends on the quality of the software but also on a logical structure of the text. Use of headings, punctuation, presence of alternate attributes for images, etc. is crucial for a good vocalization. Also a web site may have a nice look because of the use of appropriate two dimensional positioning with CSS but its standard linearization, for example, by suppressing any CSS and Javascript in the browser may not be comprehensible.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}}


== Customization ==
== Customization ==

Action parameters

VariableValue
Edit count of the user (user_editcount)
null
Name of the user account (user_name)
'104.218.65.8'
Age of the user account (user_age)
0
Groups (including implicit) the user is in (user_groups)
[ 0 => '*' ]
Rights that the user has (user_rights)
[ 0 => 'createaccount', 1 => 'read', 2 => 'edit', 3 => 'createtalk', 4 => 'writeapi', 5 => 'viewmywatchlist', 6 => 'editmywatchlist', 7 => 'viewmyprivateinfo', 8 => 'editmyprivateinfo', 9 => 'editmyoptions', 10 => 'abusefilter-log-detail', 11 => 'urlshortener-create-url', 12 => 'centralauth-merge', 13 => 'abusefilter-view', 14 => 'abusefilter-log', 15 => 'vipsscaler-test' ]
Whether the user is editing from mobile app (user_app)
false
Whether or not a user is editing through the mobile interface (user_mobile)
true
Page ID (page_id)
418544
Page namespace (page_namespace)
0
Page title without namespace (page_title)
'Screen reader'
Full page title (page_prefixedtitle)
'Screen reader'
Edit protection level of the page (page_restrictions_edit)
[]
Last ten users to contribute to the page (page_recent_contributors)
[ 0 => 'Willondon', 1 => '104.218.65.8', 2 => 'PerpetuityGrat', 3 => '149.62.204.75', 4 => 'Citation bot', 5 => 'Materialscientist', 6 => '2401:4900:5D1B:4F0E:A91C:9372:EE3E:EE16', 7 => 'Graham87', 8 => 'AnomieBOT', 9 => '2001:44C8:4721:78B8:815D:BAB2:A23:2E3D' ]
Page age in seconds (page_age)
562692861
Action (action)
'edit'
Edit summary/reason (summary)
''
Old content model (old_content_model)
'wikitext'
New content model (new_content_model)
'wikitext'
Old page wikitext, before the edit (old_wikitext)
'{{short description|Assistive technology that converts text or images to speech or Braille}} {{More citations needed|date=July 2017}} [[File:Accessible Books Consortium explains - a digital file is not necessarily accessible.webm|thumb|An example of someone using a screen reader showing documents that are inaccessible, readable and accessible.]] {{Use mdy dates|date=July 2017}} A '''screen reader''' is a form of [[assistive technology]] (<abbr>AT</abbr>)<ref>{{cite web|url=https://www.microsoft.com/enable/at/types.aspx|title=Types of Assistive Technology Products|publisher=Microsoft Accessibility|access-date=13 June 2016}}</ref> that renders text and image content as speech or braille output. Screen readers are essential to people who are [[blindness|blind]],<ref name="Braille display"/> and are useful to people who are [[visual impairment|visually impaired]],<ref name="Braille display"/> [[Illiteracy|illiterate]], or have a [[learning disability]].<ref name="Screen1">{{cite web|url=http://www.vadsa.org/ace/reader.htm|title=Screen Readers and how they work with E-Learning|publisher=Virginia.gov|access-date=31 March 2019|archive-url=https://web.archive.org/web/20181113075826/https://www.vadsa.org/ace/reader.htm|archive-date=13 November 2018}}</ref> Screen readers are [[Application software|software applications]] that attempt to convey what people with normal eyesight see on a [[Display device|display]] to their users via non-visual means, like [[text-to-speech]],<ref>{{cite web|url=http://windows.microsoft.com/en-us/windows/hear-text-read-aloud-narrator#1TC=windows-8|title=Hear text read aloud with Narrator|publisher=[[Microsoft Office|Microsoft]]|access-date=13 June 2016}}</ref> sound icons,<ref>{{cite web|url=https://css-tricks.com/accessiblility-basics-turn-your-css-off/|title=Accessibility Basics: How Does Your Page Look To A Screen Reader?|last=Coyier|first=Chris|date=29 October 2007|publisher=CSS-Tricks|access-date=13 June 2016}}</ref> or a [[Refreshable Braille display|Braille device]].<ref name="Braille display">{{cite web|url=http://www.afb.org/info/living-with-vision-loss/using-technology/assistive-technology-videos/screen-reading-technology/1235|title=Screen Readers Technology and Refreshable Braille Displays|publisher=[[American Foundation for the Blind|AFB]]|access-date=13 June 2016}}</ref> They do this by applying a wide variety of techniques that include, for example, interacting with dedicated [[#Accessibility APIs|accessibility APIs]], using various [[operating system]] features (like [[inter-process communication]] and querying [[user interface]] properties), and employing [[hooking]] techniques.<ref name="SR Overview">{{cite web|url=https://www.nomensa.com/blog/2005/what-screen-reader|title=What is a Screen Reader|publisher=[[Nomensa]]|access-date=9 July 2017}}</ref> [[Microsoft Windows]] [[operating systems]] have included the [[Microsoft Narrator]] screen reader since [[Windows 2000]], though separate products such as [[Freedom Scientific]]'s commercially available [[JAWS (screen reader)|JAWS]] screen reader and [[ZoomText]] screen magnifier and the [[free and open source]] screen reader [[NonVisual Desktop Access|NVDA]] by NV Access are more popular for that operating system.<ref>{{cite web|url=https://webaim.org/projects/screenreadersurvey9/|title=Screen Reader User Survey #9|publisher=[[WebAIM]]|access-date=July 1, 2021}}</ref> [[Apple Inc.]]'s [[macOS]], [[iOS]], and [[tvOS]] include [[VoiceOver]] as a built-in screen reader, while [[Google]]'s [[Android (operating system)|Android]] provides the [[Google TalkBack|Talkback screen reader]] and its [[Chrome OS]] can use ChromeVox.<ref>{{cite web|url=http://www.chromevox.com/|title=ChromeVox|publisher=Google|access-date=March 9, 2020}}</ref> Similarly, Android-based devices from Amazon provide the VoiceView screen reader. There are also free and open source screen readers for [[Linux]] and [[Unix-like]] systems, such as Speakup and [[Orca (assistive technology)|Orca]]. == Types == === Command-line (text) === In early [[operating system]]s, such as [[MS-DOS]], which employed [[command-line interface]]s (<abbr>CLI</abbr>s), the screen display consisted of [[Character (computing)|characters]] [[Memory-mapped I/O|mapping]] directly to a [[screen buffer]] in [[Computer memory|memory]] and a [[Cursor (computers)|cursor]] position. Input was by keyboard. All this information could therefore be obtained from the system either by [[hooking]] the flow of information around the system and reading the screen buffer or by using a standard hardware output socket<ref>{{cite web | title=Talking Terminals. BYTE, September 1982 | url=http://www.edstoffel.com/david/talkingterminals.html | access-date=September 7, 2006 |archive-url = https://web.archive.org/web/20060625225004/http://www.edstoffel.com/david/talkingterminals.html <!-- Bot retrieved archive --> |archive-date = June 25, 2006}}</ref> and communicating the results to the user. In the 1980s, the Research Centre for the Education of the Visually Handicapped (<abbr>RCEVH</abbr>) at the [[University of Birmingham]] developed Screen Reader for the [[BBC Micro]] and <!--abbr?-->NEC Portable.<ref>Paul Blenkhorn, "The <abbr>RCEVH</abbr> project on micro-computer systems and computer assisted learning", British Journal of Visual Impairment, 4/3, 101-103 (1986). [http://www.visugate.biz/bjvi/1986/autumn1986.html#RCEVH Free HTML version at Visugate].</ref><ref>{{cite web | title=Access to personal computers using speech synthesis. RNIB New Beacon No.76, May 1992 | date=March 3, 2014| url=http://www.rnib.org.uk/information-everyday-living-using-technology-beginners-guides/beginners-guide-assistive-technology}}</ref> === Graphical === ==== Off-screen models ==== With the arrival of [[graphical user interface]]s (<abbr>GUI</abbr>s), the situation became more complicated. A <abbr>GUI</abbr> has characters and graphics drawn on the screen at particular positions, and therefore there is no purely textual representation of the graphical contents of the display. Screen readers were therefore forced to employ new low-level techniques, gathering messages from the [[operating system]] and using these to build up an "off-screen model", a representation of the display in which the required text content is stored.<ref>According to "[ftp://service.boulder.ibm.com/sns/sr-os2/sr2doc/guitalk.txt Making the <abbr>GUI</abbr> Talk]" (by Richard Schwerdtfeger, ''BYTE'' December 1991, p. 118-128), the first screen reader to build an off-screen model was outSPOKEN.</ref> For example, the operating system might send messages to draw a command button and its caption. These messages are intercepted and used to construct the off-screen model. The user can switch between controls (such as buttons) available on the screen and the captions and control contents will be read aloud and/or shown on refreshable [[Braille]] display. Screen readers can also communicate information on menus, controls, and other visual constructs to permit blind users to interact with these constructs. However, maintaining an off-screen model is a significant technical challenge; hooking the low-level messages and maintaining an accurate model are both difficult tasks.{{citation needed|date=January 2015}} ==== Accessibility APIs ==== Operating system and application designers have attempted to address these problems by providing ways for screen readers to access the display contents without having to maintain an off-screen model. These involve the provision of alternative and accessible representations of what is being displayed on the screen accessed through an <abbr>[[Application programming interface|API]]</abbr>. Existing <abbr>API</abbr>s include: * Android Accessibility Framework<ref>[https://developer.android.com/training/accessibility/index.html Implementing Accessibility on Android].</ref> * Apple Accessibility <abbr>API</abbr><ref>[https://developer.apple.com/documentation/Accessibility/Reference/AccessibilityLowlevel/index.html Apple Accessibility <abbr>API</abbr>].</ref> * [[AT-SPI]] * [[IAccessible2]] * [[Microsoft Active Accessibility]] (<abbr>MSAA</abbr>) * [[Microsoft UI Automation]] * [[Java Access Bridge]]<ref>{{cite web|url=http://java.sun.com/products/accessbridge/|title=Oracle Technology Network for Java Developers – Oracle Technology Network – Oracle}}</ref> Screen readers can query the operating system or application for what is currently being displayed and receive updates when the display changes. For example, a screen reader can be told that the current focus is on a button and the button caption to be communicated to the user. This approach is considerably easier for the developers of screen readers, but fails when applications do not comply with the accessibility <abbr>API</abbr>: for example, [[Microsoft Word]] does not comply with the <abbr>MSAA</abbr> <abbr>API</abbr>, so screen readers must still maintain an off-screen model for Word or find another way to access its contents.{{citation needed|date=January 2015}} One approach is to use available operating system messages and application object models to supplement accessibility <abbr>API</abbr>s. Screen readers can be assumed to be able to access all display content that is not intrinsically inaccessible. Web browsers, word processors, icons and windows and email programs are just some of the applications used successfully by screen reader users. However, according to some users,{{Who|date=January 2015}} using a screen reader is considerably more difficult than using a GUI, and many applications have specific problems resulting from the nature of the application (e.g. animations) or failure to comply with accessibility standards for the platform (e.g. Microsoft Word and Active Accessibility).{{citation needed|date=January 2015}} === Self-voicing programs and applications === Some programs and applications have voicing technology built in alongside their primary functionality. These programs are termed [[self-voicing]] and can be a form of [[assistive technology]] if they are designed to remove the need to use a screen reader.{{citation needed|date=January 2015}} === Cloud-based === Some telephone services allow users to interact with the internet remotely. For example, TeleTender can read web pages over the phone and does not require special programs or devices on the user side.{{citation needed|date=January 2015}} === Web-based === A relatively new development in the field is web-based applications like Spoken-Web that act as web portals, managing content like news updates, weather, science and business articles for visually-impaired or blind computer users.{{citation needed|date=January 2015}} Other examples are ReadSpeaker or [[BrowseAloud]] that add [[text-to-speech]] functionality to web content.{{citation needed|date=January 2015}} The primary audience for such applications is those who have difficulty reading because of learning disabilities or language barriers.{{citation needed|date=January 2015}} Although functionality remains limited compared to equivalent desktop applications, the major benefit is to increase the accessibility of said websites when viewed on public machines where users do not have permission to install custom software, giving people greater "freedom to roam".{{citation needed|date=January 2015}} With the development of [[smartphone]]s, the ability to listen to written documents (textual web content, <abbr>[[Portable Document Format|PDF]]</abbr> documents, e-mails etc.) while driving or during a similar activity in the same way that listening to music, will benefit a much broader audience than visually-impaired people.{{citation needed|date=January 2015}} The best-known examples are [[Siri (software)|Siri]] for [[iOS]], and [[Google Now]] and [[Iris (software)|Iris]] for [[Android (operating system)|Android]]. With the release of the [[Samsung Galaxy S III|Galaxy S III]], [[Samsung]] also introduced a similar [[intelligent personal assistant]] called [[S Voice]]. On the [[BlackBerry 10 operating system voice control|BlackBerry 10 operating system]], their [[BlackBerry Z30|Z30 smartphone]] also features spoken interaction features, which are similar to the other [[mobile operating system]]s.{{citation needed| date= October 2013| reason= Similarity unverified.}} This functionality depends on the quality of the software but also on a logical structure of the text. Use of headings, punctuation, presence of alternate attributes for images, etc. is crucial for a good vocalization. Also a web site may have a nice look because of the use of appropriate two dimensional positioning with CSS but its standard linearization, for example, by suppressing any CSS and Javascript in the browser may not be comprehensible.{{citation needed|date=January 2015}} == Customization == Most screen readers allow the user to select whether most [[punctuation]] is announced or silently ignored. Some screen readers can be tailored to a particular application through [[script (computing)|'''scripting''']]. One advantage of scripting is that it allows customizations to be shared among users, increasing accessibility for all. <abbr>[[Job Access With Speech|JAWS]]</abbr> enjoys an active script-sharing community, for example.{{citation needed|date=January 2015}} === Verbosity === Verbosity is a feature of screen reading software that supports vision-impaired computer users. Speech verbosity controls enable users to choose how much speech feedback they wish to hear. Specifically, verbosity settings allow users to construct a mental model of web pages displayed on their computer screen. Based on verbosity settings, a screen-reading program informs users of certain formatting changes, such as when a frame or table begins and ends, where graphics have been inserted into the text, or when a list appears in the document.{{citation needed|date=January 2015}} === Language === Some screen readers can read text in more than one [[language]], provided that the language of the material is encoded in its [[Metadata Encoding and Transmission Standard|metadata]].<ref>{{cite web|url=https://developer.yahoo.com/blogs/ydn/yahoo-search-results-now-natural-language-support-7318.html|title=Yahoo! search results now with natural language support|date=March 13, 2008|author=Chris Heilmann|work=[[Yahoo! Developer Network]] Blog|access-date=February 28, 2015|archive-url=https://web.archive.org/web/20090125024422/http://developer.yahoo.net/blog/archives/2008/03/yahoo_search_re.html|archive-date=January 25, 2009|url-status=live}}</ref> Some screen reading programs{{Which|date=January 2015}} also include language verbosity, which automatically detects verbosity settings related to speech output language. For example, if a user navigated to a website based in the United Kingdom, the text would be read with [[British English|an English accent]].{{citation needed|date=January 2015}} == See also == * [[List of screen readers]] * [[Screen magnifier]] * [[Self-voicing]] * [[Speech processing]] * [[Speech recognition]] * [[Speech synthesis]] * [[Vinux]] * [[VoiceOver]] == References == {{Reflist}} == External links == {{Wiktionary}} * [http://www.standards-schmandards.com/projects/fangs Fangs screen reader emulator]{{snds}}an open-source Mozilla [[List of Firefox extensions|Firefox extension]] that simulates how a web page would look in [[JAWS (screen reader)|JAWS]] {{authority control}} [[Category:Assistive technology]] [[Category:Screen readers| ]]'
New page wikitext, after the edit (new_wikitext)
'{{short description|Assistive technology that converts text or images to speech or Braille}} {{More citations needed|date=July 2017}} [[File:Accessible Books Consortium explains - a digital file is not necessarily accessible.webm|thumb|An example of someone using a screen reader showing documents that are inaccessible, readable and accessible.]] {{Use mdy dates|date=July 2017}} A '''screen reader''' is a form of [[assistive technology]] (<abbr>AT</abbr>)<ref>{{cite web|url=https://www.microsoft.com/enable/at/types.aspx|title=Types of Assistive Technology Products|publisher=Microsoft Accessibility|access-date=13 June 2016}}</ref> that renders text and image content as speech or braille output. Screen readers are essential to people who are [[blindness|blind]],<ref name="Braille display"/> and are useful to people who are [[visual impairment|visually impaired]],<ref name="Braille display"/> [[Illiteracy|illiterate]], or have a [[learning disability]].<ref name="Screen1">{{cite web|url=http://www.vadsa.org/ace/reader.htm|title=Screen Readers and how they work with E-Learning|publisher=Virginia.gov|access-date=31 March 2019|archive-url=https://web.archive.org/web/20181113075826/https://www.vadsa.org/ace/reader.htm|archive-date=13 November 2018}}</ref> Screen readers are [[Application software|software applications]] that attempt to convey what people with normal eyesight see on a [[Display device|display]] to their users via non-visual means, like [[text-to-speech]],<ref>{{cite web|url=http://windows.microsoft.com/en-us/windows/hear-text-read-aloud-narrator#1TC=windows-8|title=Hear text read aloud with Narrator|publisher=[[Microsoft Office|Microsoft]]|access-date=13 June 2016}}</ref> sound icons,<ref>{{cite web|url=https://css-tricks.com/accessiblility-basics-turn-your-css-off/|title=Accessibility Basics: How Does Your Page Look To A Screen Reader?|last=Coyier|first=Chris|date=29 October 2007|publisher=CSS-Tricks|access-date=13 June 2016}}</ref> or a [[Refreshable Braille display|Braille device]].<ref name="Braille display">{{cite web|url=http://www.afb.org/info/living-with-vision-loss/using-technology/assistive-technology-videos/screen-reading-technology/1235|title=Screen Readers Technology and Refreshable Braille Displays|publisher=[[American Foundation for the Blind|AFB]]|access-date=13 June 2016}}</ref> They do this by applying a wide variety of techniques that include, for example, interacting with dedicated [[#Accessibility APIs|accessibility APIs]], using various [[operating system]] features (like [[inter-process communication]] and querying [[user interface]] properties), and employing [[hooking]] techniques.<ref name="SR Overview">{{cite web|url=https://www.nomensa.com/blog/2005/what-screen-reader|title=What is a Screen Reader|publisher=[[Nomensa]]|access-date=9 July 2017}}</ref> [[Microsoft Windows]] [[operating systems]] have included the [[Microsoft Narrator]] screen reader since [[Windows 2000]], though separate products such as [[Freedom Scientific]]'s commercially available [[JAWS (screen reader)|JAWS]] screen reader and [[ZoomText]] screen magnifier and the [[free and open source]] screen reader [[NonVisual Desktop Access|NVDA]] by NV Access are more popular for that operating system.<ref>{{cite web|url=https://webaim.org/projects/screenreadersurvey9/|title=Screen Reader User Survey #9|publisher=[[WebAIM]]|access-date=July 1, 2021}}</ref> [[Apple Inc.]]'s [[macOS]], [[iOS]], and [[tvOS]] include [[VoiceOver]] as a built-in screen reader, while [[Google]]'s [[Android (operating system)|Android]] provides the [[Google TalkBack|Talkback screen reader]] and its [[Chrome OS]] can use ChromeVox.<ref>{{cite web|url=http://www.chromevox.com/|title=ChromeVox|publisher=Google|access-date=March 9, 2020}}</ref> Similarly, Android-based devices from Amazon provide the VoiceView screen reader. There are also free and open source screen readers for [[Linux]] and [[Unix-like]] systems, such as Speakup and [[Orca (assistive technology)|Orca]]. == Types == === Command-line (text) === In early operating systems, such as MS-DOS, which employed command-line interfaces (<abbr>CLI</abbr>s), the screen display consisted of characters mapping directly to a screen buffer in memory and a cursor position. Input was by keyboard. All this information could therefore be obtained from the system either by hooking the flow of information around the system and reading the screen buffer or by using a standard hardware output socket and communicating the results to the user. In the 1980s, the Research Centre for the Education of the Visually Handicapped (<abbr>RCEVH</abbr>) at the University of Birmingham developed Screen Reader for the BBC Micro and <!--abbr?-->NEC Portable. === Graphical === ==== Off-screen models ==== With the arrival of graphical user interfaces (<abbr>GUI</abbr>s), the situation became more complicated. A <abbr>GUI</abbr> has characters and graphics drawn on the screen at particular positions, and therefore there is no purely textual representation of the graphical contents of the display. Screen readers were therefore forced to employ new low-level techniques, gathering messages from the operating system and using these to build up an "off-screen model", a representation of the display in which the required text content is stored. For example, the operating system might send messages to draw a command button and its caption. These messages are intercepted and used to construct the off-screen model. The user can switch between controls (such as buttons) available on the screen and the captions and control contents will be read aloud and/or shown on refreshable Braille display. Screen readers can also communicate information on menus, controls, and other visual constructs to permit blind users to interact with these constructs. However, maintaining an off-screen model is a significant technical challenge; hooking the low-level messages and maintaining an accurate model are both difficult tasks.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} ==== Accessibility APIs ==== Operating system and application designers have attempted to address these problems by providing ways for screen readers to access the display contents without having to maintain an off-screen model. These involve the provision of alternative and accessible representations of what is being displayed on the screen accessed through an <abbr>API</abbr>. Existing <abbr>API</abbr>s include: * Android Accessibility Framework * Apple Accessibility <abbr>API</abbr> * AT-SPI * I Accessible * Microsoft Active Accessibility (<abbr>MSAA</abbr>) * Microsoft UI Automation * Java Access Bridge Screen readers can query the operating system or application for what is currently being displayed and receive updates when the display changes. For example, a screen reader can be told that the current focus is on a button and the button caption to be communicated to the user. This approach is considerably easier for the developers of screen readers, but fails when applications do not comply with the accessibility <abbr>API</abbr>: for example, Microsoft Word does not comply with the <abbr>MSAA</abbr> <abbr>API</abbr>, so screen readers must still maintain an off-screen model for Word or find another way to access its contents.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} One approach is to use available operating system messages and application object models to supplement accessibility <abbr>API</abbr>s. Screen readers can be assumed to be able to access all display content that is not intrinsically inaccessible. Web browsers, word processors, icons and windows and email programs are just some of the applications used successfully by screen reader users. However, according to some users,{{Who|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} using a screen reader is considerably more difficult than using a GUI, and many applications have specific problems resulting from the nature of the application (e.g. animations) or failure to comply with accessibility standards for the platform (e.g. Microsoft Word and Active Accessibility).{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} === Self-voicing programs and applications === Some programs and applications have voicing technology built in alongside their primary functionality. These programs are termed self-voicing and can be a form of assistive technology if they are designed to remove the need to use a screen reader.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} === Cloud-based === Some telephone services allow users to interact with the internet remotely. For example, TeleTender can read web pages over the phone and does not require special programs or devices on the user side.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} === Web-based === A relatively new development in the field is web-based applications like Spoken-Web that act as web portals, managing content like news updates, weather, science and business articles for visually-impaired or blind computer users.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} Other examples are ReadSpeaker or BrowseAloud that add text-to-speech functionality to web content.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} The primary audience for such applications is those who have difficulty reading because of learning disabilities or language barriers.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} Although functionality remains limited compared to equivalent desktop applications, the major benefit is to increase the accessibility of said websites when viewed on public machines where users do not have permission to install custom software, giving people greater "freedom to roam".{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} With the development of smartphones, the ability to listen to written documents (textual web content, <abbr>PDF</abbr> documents, e-mails etc.) while driving or during a similar activity in the same way that listening to music, will benefit a much broader audience than visually-impaired people.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} The best-known examples are Siri for iOS, and Google Now and Iris for Android. With the release of the Galaxy S III, Samsung also introduced a similar intelligent personal assistant called S Voice. On the BlackBerry 10 operating system, their Z30 smartphone also features spoken interaction features, which are similar to the other mobile operating systems.{{citation needed| date= {{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}| reason= Similarity unverified.}} This functionality depends on the quality of the software but also on a logical structure of the text. Use of headings, punctuation, presence of alternate attributes for images, etc. is crucial for a good vocalization. Also a web site may have a nice look because of the use of appropriate two dimensional positioning with CSS but its standard linearization, for example, by suppressing any CSS and Javascript in the browser may not be comprehensible.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} == Customization == Most screen readers allow the user to select whether most [[punctuation]] is announced or silently ignored. Some screen readers can be tailored to a particular application through [[script (computing)|'''scripting''']]. One advantage of scripting is that it allows customizations to be shared among users, increasing accessibility for all. <abbr>[[Job Access With Speech|JAWS]]</abbr> enjoys an active script-sharing community, for example.{{citation needed|date=January 2015}} === Verbosity === Verbosity is a feature of screen reading software that supports vision-impaired computer users. Speech verbosity controls enable users to choose how much speech feedback they wish to hear. Specifically, verbosity settings allow users to construct a mental model of web pages displayed on their computer screen. Based on verbosity settings, a screen-reading program informs users of certain formatting changes, such as when a frame or table begins and ends, where graphics have been inserted into the text, or when a list appears in the document.{{citation needed|date=January 2015}} === Language === Some screen readers can read text in more than one [[language]], provided that the language of the material is encoded in its [[Metadata Encoding and Transmission Standard|metadata]].<ref>{{cite web|url=https://developer.yahoo.com/blogs/ydn/yahoo-search-results-now-natural-language-support-7318.html|title=Yahoo! search results now with natural language support|date=March 13, 2008|author=Chris Heilmann|work=[[Yahoo! Developer Network]] Blog|access-date=February 28, 2015|archive-url=https://web.archive.org/web/20090125024422/http://developer.yahoo.net/blog/archives/2008/03/yahoo_search_re.html|archive-date=January 25, 2009|url-status=live}}</ref> Some screen reading programs{{Which|date=January 2015}} also include language verbosity, which automatically detects verbosity settings related to speech output language. For example, if a user navigated to a website based in the United Kingdom, the text would be read with [[British English|an English accent]].{{citation needed|date=January 2015}} == See also == * [[List of screen readers]] * [[Screen magnifier]] * [[Self-voicing]] * [[Speech processing]] * [[Speech recognition]] * [[Speech synthesis]] * [[Vinux]] * [[VoiceOver]] == References == {{Reflist}} == External links == {{Wiktionary}} * [http://www.standards-schmandards.com/projects/fangs Fangs screen reader emulator]{{snds}}an open-source Mozilla [[List of Firefox extensions|Firefox extension]] that simulates how a web page would look in [[JAWS (screen reader)|JAWS]] {{authority control}} [[Category:Assistive technology]] [[Category:Screen readers| ]]'
Unified diff of changes made by edit (edit_diff)
'@@ -11,10 +11,7 @@ === Command-line (text) === -In early [[operating system]]s, such as [[MS-DOS]], which employed [[command-line interface]]s (<abbr>CLI</abbr>s), the screen display consisted of [[Character (computing)|characters]] [[Memory-mapped I/O|mapping]] directly to a [[screen buffer]] in [[Computer memory|memory]] and a [[Cursor (computers)|cursor]] position. Input was by keyboard. All this information could therefore be obtained from the system either by [[hooking]] the flow of information around the system and reading the screen buffer or by using a standard hardware output socket<ref>{{cite web -| title=Talking Terminals. BYTE, September 1982 -| url=http://www.edstoffel.com/david/talkingterminals.html -| access-date=September 7, 2006 |archive-url = https://web.archive.org/web/20060625225004/http://www.edstoffel.com/david/talkingterminals.html <!-- Bot retrieved archive --> |archive-date = June 25, 2006}}</ref> and communicating the results to the user. +In early operating systems, such as MS-DOS, which employed command-line interfaces (<abbr>CLI</abbr>s), the screen display consisted of characters mapping directly to a screen buffer in memory and a cursor position. Input was by keyboard. All this information could therefore be obtained from the system either by hooking the flow of information around the system and reading the screen buffer or by using a standard hardware output socket and communicating the results to the user. -In the 1980s, the Research Centre for the Education of the Visually Handicapped (<abbr>RCEVH</abbr>) at the [[University of Birmingham]] developed Screen Reader for the [[BBC Micro]] and <!--abbr?-->NEC Portable.<ref>Paul Blenkhorn, "The <abbr>RCEVH</abbr> project on micro-computer systems and computer assisted learning", British Journal of Visual Impairment, 4/3, 101-103 (1986). [http://www.visugate.biz/bjvi/1986/autumn1986.html#RCEVH Free HTML version at Visugate].</ref><ref>{{cite web | title=Access to personal computers using speech synthesis. RNIB New Beacon No.76, May 1992 | date=March 3, 2014| url=http://www.rnib.org.uk/information-everyday-living-using-technology-beginners-guides/beginners-guide-assistive-technology}}</ref> +In the 1980s, the Research Centre for the Education of the Visually Handicapped (<abbr>RCEVH</abbr>) at the University of Birmingham developed Screen Reader for the BBC Micro and <!--abbr?-->NEC Portable. === Graphical === @@ -22,40 +19,40 @@ ==== Off-screen models ==== -With the arrival of [[graphical user interface]]s (<abbr>GUI</abbr>s), the situation became more complicated. A <abbr>GUI</abbr> has characters and graphics drawn on the screen at particular positions, and therefore there is no purely textual representation of the graphical contents of the display. Screen readers were therefore forced to employ new low-level techniques, gathering messages from the [[operating system]] and using these to build up an "off-screen model", a representation of the display in which the required text content is stored.<ref>According to "[ftp://service.boulder.ibm.com/sns/sr-os2/sr2doc/guitalk.txt Making the <abbr>GUI</abbr> Talk]" (by Richard Schwerdtfeger, ''BYTE'' December 1991, p. 118-128), the first screen reader to build an off-screen model was outSPOKEN.</ref> +With the arrival of graphical user interfaces (<abbr>GUI</abbr>s), the situation became more complicated. A <abbr>GUI</abbr> has characters and graphics drawn on the screen at particular positions, and therefore there is no purely textual representation of the graphical contents of the display. Screen readers were therefore forced to employ new low-level techniques, gathering messages from the operating system and using these to build up an "off-screen model", a representation of the display in which the required text content is stored. -For example, the operating system might send messages to draw a command button and its caption. These messages are intercepted and used to construct the off-screen model. The user can switch between controls (such as buttons) available on the screen and the captions and control contents will be read aloud and/or shown on refreshable [[Braille]] display. +For example, the operating system might send messages to draw a command button and its caption. These messages are intercepted and used to construct the off-screen model. The user can switch between controls (such as buttons) available on the screen and the captions and control contents will be read aloud and/or shown on refreshable Braille display. -Screen readers can also communicate information on menus, controls, and other visual constructs to permit blind users to interact with these constructs. However, maintaining an off-screen model is a significant technical challenge; hooking the low-level messages and maintaining an accurate model are both difficult tasks.{{citation needed|date=January 2015}} +Screen readers can also communicate information on menus, controls, and other visual constructs to permit blind users to interact with these constructs. However, maintaining an off-screen model is a significant technical challenge; hooking the low-level messages and maintaining an accurate model are both difficult tasks.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} ==== Accessibility APIs ==== -Operating system and application designers have attempted to address these problems by providing ways for screen readers to access the display contents without having to maintain an off-screen model. These involve the provision of alternative and accessible representations of what is being displayed on the screen accessed through an <abbr>[[Application programming interface|API]]</abbr>. Existing <abbr>API</abbr>s include: -* Android Accessibility Framework<ref>[https://developer.android.com/training/accessibility/index.html Implementing Accessibility on Android].</ref> -* Apple Accessibility <abbr>API</abbr><ref>[https://developer.apple.com/documentation/Accessibility/Reference/AccessibilityLowlevel/index.html Apple Accessibility <abbr>API</abbr>].</ref> -* [[AT-SPI]] -* [[IAccessible2]] -* [[Microsoft Active Accessibility]] (<abbr>MSAA</abbr>) -* [[Microsoft UI Automation]] -* [[Java Access Bridge]]<ref>{{cite web|url=http://java.sun.com/products/accessbridge/|title=Oracle Technology Network for Java Developers – Oracle Technology Network – Oracle}}</ref> +Operating system and application designers have attempted to address these problems by providing ways for screen readers to access the display contents without having to maintain an off-screen model. These involve the provision of alternative and accessible representations of what is being displayed on the screen accessed through an <abbr>API</abbr>. Existing <abbr>API</abbr>s include: +* Android Accessibility Framework +* Apple Accessibility <abbr>API</abbr> +* AT-SPI +* I Accessible +* Microsoft Active Accessibility (<abbr>MSAA</abbr>) +* Microsoft UI Automation +* Java Access Bridge -Screen readers can query the operating system or application for what is currently being displayed and receive updates when the display changes. For example, a screen reader can be told that the current focus is on a button and the button caption to be communicated to the user. This approach is considerably easier for the developers of screen readers, but fails when applications do not comply with the accessibility <abbr>API</abbr>: for example, [[Microsoft Word]] does not comply with the <abbr>MSAA</abbr> <abbr>API</abbr>, so screen readers must still maintain an off-screen model for Word or find another way to access its contents.{{citation needed|date=January 2015}} One approach is to use available operating system messages and application object models to supplement accessibility <abbr>API</abbr>s. +Screen readers can query the operating system or application for what is currently being displayed and receive updates when the display changes. For example, a screen reader can be told that the current focus is on a button and the button caption to be communicated to the user. This approach is considerably easier for the developers of screen readers, but fails when applications do not comply with the accessibility <abbr>API</abbr>: for example, Microsoft Word does not comply with the <abbr>MSAA</abbr> <abbr>API</abbr>, so screen readers must still maintain an off-screen model for Word or find another way to access its contents.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} One approach is to use available operating system messages and application object models to supplement accessibility <abbr>API</abbr>s. -Screen readers can be assumed to be able to access all display content that is not intrinsically inaccessible. Web browsers, word processors, icons and windows and email programs are just some of the applications used successfully by screen reader users. However, according to some users,{{Who|date=January 2015}} using a screen reader is considerably more difficult than using a GUI, and many applications have specific problems resulting from the nature of the application (e.g. animations) or failure to comply with accessibility standards for the platform (e.g. Microsoft Word and Active Accessibility).{{citation needed|date=January 2015}} +Screen readers can be assumed to be able to access all display content that is not intrinsically inaccessible. Web browsers, word processors, icons and windows and email programs are just some of the applications used successfully by screen reader users. However, according to some users,{{Who|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} using a screen reader is considerably more difficult than using a GUI, and many applications have specific problems resulting from the nature of the application (e.g. animations) or failure to comply with accessibility standards for the platform (e.g. Microsoft Word and Active Accessibility).{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} === Self-voicing programs and applications === -Some programs and applications have voicing technology built in alongside their primary functionality. These programs are termed [[self-voicing]] and can be a form of [[assistive technology]] if they are designed to remove the need to use a screen reader.{{citation needed|date=January 2015}} +Some programs and applications have voicing technology built in alongside their primary functionality. These programs are termed self-voicing and can be a form of assistive technology if they are designed to remove the need to use a screen reader.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} === Cloud-based === -Some telephone services allow users to interact with the internet remotely. For example, TeleTender can read web pages over the phone and does not require special programs or devices on the user side.{{citation needed|date=January 2015}} +Some telephone services allow users to interact with the internet remotely. For example, TeleTender can read web pages over the phone and does not require special programs or devices on the user side.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} === Web-based === -A relatively new development in the field is web-based applications like Spoken-Web that act as web portals, managing content like news updates, weather, science and business articles for visually-impaired or blind computer users.{{citation needed|date=January 2015}} Other examples are ReadSpeaker or [[BrowseAloud]] that add [[text-to-speech]] functionality to web content.{{citation needed|date=January 2015}} The primary audience for such applications is those who have difficulty reading because of learning disabilities or language barriers.{{citation needed|date=January 2015}} Although functionality remains limited compared to equivalent desktop applications, the major benefit is to increase the accessibility of said websites when viewed on public machines where users do not have permission to install custom software, giving people greater "freedom to roam".{{citation needed|date=January 2015}} +A relatively new development in the field is web-based applications like Spoken-Web that act as web portals, managing content like news updates, weather, science and business articles for visually-impaired or blind computer users.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} Other examples are ReadSpeaker or BrowseAloud that add text-to-speech functionality to web content.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} The primary audience for such applications is those who have difficulty reading because of learning disabilities or language barriers.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} Although functionality remains limited compared to equivalent desktop applications, the major benefit is to increase the accessibility of said websites when viewed on public machines where users do not have permission to install custom software, giving people greater "freedom to roam".{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} -With the development of [[smartphone]]s, the ability to listen to written documents (textual web content, <abbr>[[Portable Document Format|PDF]]</abbr> documents, e-mails etc.) while driving or during a similar activity in the same way that listening to music, will benefit a much broader audience than visually-impaired people.{{citation needed|date=January 2015}} The best-known examples are [[Siri (software)|Siri]] for [[iOS]], and [[Google Now]] and [[Iris (software)|Iris]] for [[Android (operating system)|Android]]. With the release of the [[Samsung Galaxy S III|Galaxy S III]], [[Samsung]] also introduced a similar [[intelligent personal assistant]] called [[S Voice]]. On the [[BlackBerry 10 operating system voice control|BlackBerry 10 operating system]], their [[BlackBerry Z30|Z30 smartphone]] also features spoken interaction features, which are similar to the other [[mobile operating system]]s.{{citation needed| date= October 2013| reason= Similarity unverified.}} +With the development of smartphones, the ability to listen to written documents (textual web content, <abbr>PDF</abbr> documents, e-mails etc.) while driving or during a similar activity in the same way that listening to music, will benefit a much broader audience than visually-impaired people.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} The best-known examples are Siri for iOS, and Google Now and Iris for Android. With the release of the Galaxy S III, Samsung also introduced a similar intelligent personal assistant called S Voice. On the BlackBerry 10 operating system, their Z30 smartphone also features spoken interaction features, which are similar to the other mobile operating systems.{{citation needed| date= {{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}| reason= Similarity unverified.}} -This functionality depends on the quality of the software but also on a logical structure of the text. Use of headings, punctuation, presence of alternate attributes for images, etc. is crucial for a good vocalization. Also a web site may have a nice look because of the use of appropriate two dimensional positioning with CSS but its standard linearization, for example, by suppressing any CSS and Javascript in the browser may not be comprehensible.{{citation needed|date=January 2015}} +This functionality depends on the quality of the software but also on a logical structure of the text. Use of headings, punctuation, presence of alternate attributes for images, etc. is crucial for a good vocalization. Also a web site may have a nice look because of the use of appropriate two dimensional positioning with CSS but its standard linearization, for example, by suppressing any CSS and Javascript in the browser may not be comprehensible.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} == Customization == '
Unified diff of changes made by edit, pre-save transformed (edit_diff_pst)
'@@ -11,10 +11,7 @@ === Command-line (text) === -In early [[operating system]]s, such as [[MS-DOS]], which employed [[command-line interface]]s (<abbr>CLI</abbr>s), the screen display consisted of [[Character (computing)|characters]] [[Memory-mapped I/O|mapping]] directly to a [[screen buffer]] in [[Computer memory|memory]] and a [[Cursor (computers)|cursor]] position. Input was by keyboard. All this information could therefore be obtained from the system either by [[hooking]] the flow of information around the system and reading the screen buffer or by using a standard hardware output socket<ref>{{cite web -| title=Talking Terminals. BYTE, September 1982 -| url=http://www.edstoffel.com/david/talkingterminals.html -| access-date=September 7, 2006 |archive-url = https://web.archive.org/web/20060625225004/http://www.edstoffel.com/david/talkingterminals.html <!-- Bot retrieved archive --> |archive-date = June 25, 2006}}</ref> and communicating the results to the user. +In early operating systems, such as MS-DOS, which employed command-line interfaces (<abbr>CLI</abbr>s), the screen display consisted of characters mapping directly to a screen buffer in memory and a cursor position. Input was by keyboard. All this information could therefore be obtained from the system either by hooking the flow of information around the system and reading the screen buffer or by using a standard hardware output socket and communicating the results to the user. -In the 1980s, the Research Centre for the Education of the Visually Handicapped (<abbr>RCEVH</abbr>) at the [[University of Birmingham]] developed Screen Reader for the [[BBC Micro]] and <!--abbr?-->NEC Portable.<ref>Paul Blenkhorn, "The <abbr>RCEVH</abbr> project on micro-computer systems and computer assisted learning", British Journal of Visual Impairment, 4/3, 101-103 (1986). [http://www.visugate.biz/bjvi/1986/autumn1986.html#RCEVH Free HTML version at Visugate].</ref><ref>{{cite web | title=Access to personal computers using speech synthesis. RNIB New Beacon No.76, May 1992 | date=March 3, 2014| url=http://www.rnib.org.uk/information-everyday-living-using-technology-beginners-guides/beginners-guide-assistive-technology}}</ref> +In the 1980s, the Research Centre for the Education of the Visually Handicapped (<abbr>RCEVH</abbr>) at the University of Birmingham developed Screen Reader for the BBC Micro and <!--abbr?-->NEC Portable. === Graphical === @@ -22,40 +19,40 @@ ==== Off-screen models ==== -With the arrival of [[graphical user interface]]s (<abbr>GUI</abbr>s), the situation became more complicated. A <abbr>GUI</abbr> has characters and graphics drawn on the screen at particular positions, and therefore there is no purely textual representation of the graphical contents of the display. Screen readers were therefore forced to employ new low-level techniques, gathering messages from the [[operating system]] and using these to build up an "off-screen model", a representation of the display in which the required text content is stored.<ref>According to "[ftp://service.boulder.ibm.com/sns/sr-os2/sr2doc/guitalk.txt Making the <abbr>GUI</abbr> Talk]" (by Richard Schwerdtfeger, ''BYTE'' December 1991, p. 118-128), the first screen reader to build an off-screen model was outSPOKEN.</ref> +With the arrival of graphical user interfaces (<abbr>GUI</abbr>s), the situation became more complicated. A <abbr>GUI</abbr> has characters and graphics drawn on the screen at particular positions, and therefore there is no purely textual representation of the graphical contents of the display. Screen readers were therefore forced to employ new low-level techniques, gathering messages from the operating system and using these to build up an "off-screen model", a representation of the display in which the required text content is stored. -For example, the operating system might send messages to draw a command button and its caption. These messages are intercepted and used to construct the off-screen model. The user can switch between controls (such as buttons) available on the screen and the captions and control contents will be read aloud and/or shown on refreshable [[Braille]] display. +For example, the operating system might send messages to draw a command button and its caption. These messages are intercepted and used to construct the off-screen model. The user can switch between controls (such as buttons) available on the screen and the captions and control contents will be read aloud and/or shown on refreshable Braille display. -Screen readers can also communicate information on menus, controls, and other visual constructs to permit blind users to interact with these constructs. However, maintaining an off-screen model is a significant technical challenge; hooking the low-level messages and maintaining an accurate model are both difficult tasks.{{citation needed|date=January 2015}} +Screen readers can also communicate information on menus, controls, and other visual constructs to permit blind users to interact with these constructs. However, maintaining an off-screen model is a significant technical challenge; hooking the low-level messages and maintaining an accurate model are both difficult tasks.{{citation needed|date=November 2021}} ==== Accessibility APIs ==== -Operating system and application designers have attempted to address these problems by providing ways for screen readers to access the display contents without having to maintain an off-screen model. These involve the provision of alternative and accessible representations of what is being displayed on the screen accessed through an <abbr>[[Application programming interface|API]]</abbr>. Existing <abbr>API</abbr>s include: -* Android Accessibility Framework<ref>[https://developer.android.com/training/accessibility/index.html Implementing Accessibility on Android].</ref> -* Apple Accessibility <abbr>API</abbr><ref>[https://developer.apple.com/documentation/Accessibility/Reference/AccessibilityLowlevel/index.html Apple Accessibility <abbr>API</abbr>].</ref> -* [[AT-SPI]] -* [[IAccessible2]] -* [[Microsoft Active Accessibility]] (<abbr>MSAA</abbr>) -* [[Microsoft UI Automation]] -* [[Java Access Bridge]]<ref>{{cite web|url=http://java.sun.com/products/accessbridge/|title=Oracle Technology Network for Java Developers – Oracle Technology Network – Oracle}}</ref> +Operating system and application designers have attempted to address these problems by providing ways for screen readers to access the display contents without having to maintain an off-screen model. These involve the provision of alternative and accessible representations of what is being displayed on the screen accessed through an <abbr>API</abbr>. Existing <abbr>API</abbr>s include: +* Android Accessibility Framework +* Apple Accessibility <abbr>API</abbr> +* AT-SPI +* I Accessible +* Microsoft Active Accessibility (<abbr>MSAA</abbr>) +* Microsoft UI Automation +* Java Access Bridge -Screen readers can query the operating system or application for what is currently being displayed and receive updates when the display changes. For example, a screen reader can be told that the current focus is on a button and the button caption to be communicated to the user. This approach is considerably easier for the developers of screen readers, but fails when applications do not comply with the accessibility <abbr>API</abbr>: for example, [[Microsoft Word]] does not comply with the <abbr>MSAA</abbr> <abbr>API</abbr>, so screen readers must still maintain an off-screen model for Word or find another way to access its contents.{{citation needed|date=January 2015}} One approach is to use available operating system messages and application object models to supplement accessibility <abbr>API</abbr>s. +Screen readers can query the operating system or application for what is currently being displayed and receive updates when the display changes. For example, a screen reader can be told that the current focus is on a button and the button caption to be communicated to the user. This approach is considerably easier for the developers of screen readers, but fails when applications do not comply with the accessibility <abbr>API</abbr>: for example, Microsoft Word does not comply with the <abbr>MSAA</abbr> <abbr>API</abbr>, so screen readers must still maintain an off-screen model for Word or find another way to access its contents.{{citation needed|date=November 2021}} One approach is to use available operating system messages and application object models to supplement accessibility <abbr>API</abbr>s. -Screen readers can be assumed to be able to access all display content that is not intrinsically inaccessible. Web browsers, word processors, icons and windows and email programs are just some of the applications used successfully by screen reader users. However, according to some users,{{Who|date=January 2015}} using a screen reader is considerably more difficult than using a GUI, and many applications have specific problems resulting from the nature of the application (e.g. animations) or failure to comply with accessibility standards for the platform (e.g. Microsoft Word and Active Accessibility).{{citation needed|date=January 2015}} +Screen readers can be assumed to be able to access all display content that is not intrinsically inaccessible. Web browsers, word processors, icons and windows and email programs are just some of the applications used successfully by screen reader users. However, according to some users,{{Who|date=November 2021}} using a screen reader is considerably more difficult than using a GUI, and many applications have specific problems resulting from the nature of the application (e.g. animations) or failure to comply with accessibility standards for the platform (e.g. Microsoft Word and Active Accessibility).{{citation needed|date=November 2021}} === Self-voicing programs and applications === -Some programs and applications have voicing technology built in alongside their primary functionality. These programs are termed [[self-voicing]] and can be a form of [[assistive technology]] if they are designed to remove the need to use a screen reader.{{citation needed|date=January 2015}} +Some programs and applications have voicing technology built in alongside their primary functionality. These programs are termed self-voicing and can be a form of assistive technology if they are designed to remove the need to use a screen reader.{{citation needed|date=November 2021}} === Cloud-based === -Some telephone services allow users to interact with the internet remotely. For example, TeleTender can read web pages over the phone and does not require special programs or devices on the user side.{{citation needed|date=January 2015}} +Some telephone services allow users to interact with the internet remotely. For example, TeleTender can read web pages over the phone and does not require special programs or devices on the user side.{{citation needed|date=November 2021}} === Web-based === -A relatively new development in the field is web-based applications like Spoken-Web that act as web portals, managing content like news updates, weather, science and business articles for visually-impaired or blind computer users.{{citation needed|date=January 2015}} Other examples are ReadSpeaker or [[BrowseAloud]] that add [[text-to-speech]] functionality to web content.{{citation needed|date=January 2015}} The primary audience for such applications is those who have difficulty reading because of learning disabilities or language barriers.{{citation needed|date=January 2015}} Although functionality remains limited compared to equivalent desktop applications, the major benefit is to increase the accessibility of said websites when viewed on public machines where users do not have permission to install custom software, giving people greater "freedom to roam".{{citation needed|date=January 2015}} +A relatively new development in the field is web-based applications like Spoken-Web that act as web portals, managing content like news updates, weather, science and business articles for visually-impaired or blind computer users.{{citation needed|date=November 2021}} Other examples are ReadSpeaker or BrowseAloud that add text-to-speech functionality to web content.{{citation needed|date=November 2021}} The primary audience for such applications is those who have difficulty reading because of learning disabilities or language barriers.{{citation needed|date=November 2021}} Although functionality remains limited compared to equivalent desktop applications, the major benefit is to increase the accessibility of said websites when viewed on public machines where users do not have permission to install custom software, giving people greater "freedom to roam".{{citation needed|date=November 2021}} -With the development of [[smartphone]]s, the ability to listen to written documents (textual web content, <abbr>[[Portable Document Format|PDF]]</abbr> documents, e-mails etc.) while driving or during a similar activity in the same way that listening to music, will benefit a much broader audience than visually-impaired people.{{citation needed|date=January 2015}} The best-known examples are [[Siri (software)|Siri]] for [[iOS]], and [[Google Now]] and [[Iris (software)|Iris]] for [[Android (operating system)|Android]]. With the release of the [[Samsung Galaxy S III|Galaxy S III]], [[Samsung]] also introduced a similar [[intelligent personal assistant]] called [[S Voice]]. On the [[BlackBerry 10 operating system voice control|BlackBerry 10 operating system]], their [[BlackBerry Z30|Z30 smartphone]] also features spoken interaction features, which are similar to the other [[mobile operating system]]s.{{citation needed| date= October 2013| reason= Similarity unverified.}} +With the development of smartphones, the ability to listen to written documents (textual web content, <abbr>PDF</abbr> documents, e-mails etc.) while driving or during a similar activity in the same way that listening to music, will benefit a much broader audience than visually-impaired people.{{citation needed|date=November 2021}} The best-known examples are Siri for iOS, and Google Now and Iris for Android. With the release of the Galaxy S III, Samsung also introduced a similar intelligent personal assistant called S Voice. On the BlackBerry 10 operating system, their Z30 smartphone also features spoken interaction features, which are similar to the other mobile operating systems.{{citation needed| date= November 2021| reason= Similarity unverified.}} -This functionality depends on the quality of the software but also on a logical structure of the text. Use of headings, punctuation, presence of alternate attributes for images, etc. is crucial for a good vocalization. Also a web site may have a nice look because of the use of appropriate two dimensional positioning with CSS but its standard linearization, for example, by suppressing any CSS and Javascript in the browser may not be comprehensible.{{citation needed|date=January 2015}} +This functionality depends on the quality of the software but also on a logical structure of the text. Use of headings, punctuation, presence of alternate attributes for images, etc. is crucial for a good vocalization. Also a web site may have a nice look because of the use of appropriate two dimensional positioning with CSS but its standard linearization, for example, by suppressing any CSS and Javascript in the browser may not be comprehensible.{{citation needed|date=November 2021}} == Customization == '
New page size (new_size)
14070
Old page size (old_size)
15574
Size change in edit (edit_delta)
-1504
Lines added in edit (added_lines)
[ 0 => 'In early operating systems, such as MS-DOS, which employed command-line interfaces (<abbr>CLI</abbr>s), the screen display consisted of characters mapping directly to a screen buffer in memory and a cursor position. Input was by keyboard. All this information could therefore be obtained from the system either by hooking the flow of information around the system and reading the screen buffer or by using a standard hardware output socket and communicating the results to the user.', 1 => 'In the 1980s, the Research Centre for the Education of the Visually Handicapped (<abbr>RCEVH</abbr>) at the University of Birmingham developed Screen Reader for the BBC Micro and <!--abbr?-->NEC Portable.', 2 => 'With the arrival of graphical user interfaces (<abbr>GUI</abbr>s), the situation became more complicated. A <abbr>GUI</abbr> has characters and graphics drawn on the screen at particular positions, and therefore there is no purely textual representation of the graphical contents of the display. Screen readers were therefore forced to employ new low-level techniques, gathering messages from the operating system and using these to build up an "off-screen model", a representation of the display in which the required text content is stored.', 3 => 'For example, the operating system might send messages to draw a command button and its caption. These messages are intercepted and used to construct the off-screen model. The user can switch between controls (such as buttons) available on the screen and the captions and control contents will be read aloud and/or shown on refreshable Braille display.', 4 => 'Screen readers can also communicate information on menus, controls, and other visual constructs to permit blind users to interact with these constructs. However, maintaining an off-screen model is a significant technical challenge; hooking the low-level messages and maintaining an accurate model are both difficult tasks.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}}', 5 => 'Operating system and application designers have attempted to address these problems by providing ways for screen readers to access the display contents without having to maintain an off-screen model. These involve the provision of alternative and accessible representations of what is being displayed on the screen accessed through an <abbr>API</abbr>. Existing <abbr>API</abbr>s include:', 6 => '* Android Accessibility Framework', 7 => '* Apple Accessibility <abbr>API</abbr>', 8 => '* AT-SPI', 9 => '* I Accessible', 10 => '* Microsoft Active Accessibility (<abbr>MSAA</abbr>)', 11 => '* Microsoft UI Automation', 12 => '* Java Access Bridge', 13 => 'Screen readers can query the operating system or application for what is currently being displayed and receive updates when the display changes. For example, a screen reader can be told that the current focus is on a button and the button caption to be communicated to the user. This approach is considerably easier for the developers of screen readers, but fails when applications do not comply with the accessibility <abbr>API</abbr>: for example, Microsoft Word does not comply with the <abbr>MSAA</abbr> <abbr>API</abbr>, so screen readers must still maintain an off-screen model for Word or find another way to access its contents.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} One approach is to use available operating system messages and application object models to supplement accessibility <abbr>API</abbr>s.', 14 => 'Screen readers can be assumed to be able to access all display content that is not intrinsically inaccessible. Web browsers, word processors, icons and windows and email programs are just some of the applications used successfully by screen reader users. However, according to some users,{{Who|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} using a screen reader is considerably more difficult than using a GUI, and many applications have specific problems resulting from the nature of the application (e.g. animations) or failure to comply with accessibility standards for the platform (e.g. Microsoft Word and Active Accessibility).{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}}', 15 => 'Some programs and applications have voicing technology built in alongside their primary functionality. These programs are termed self-voicing and can be a form of assistive technology if they are designed to remove the need to use a screen reader.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}}', 16 => 'Some telephone services allow users to interact with the internet remotely. For example, TeleTender can read web pages over the phone and does not require special programs or devices on the user side.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}}', 17 => 'A relatively new development in the field is web-based applications like Spoken-Web that act as web portals, managing content like news updates, weather, science and business articles for visually-impaired or blind computer users.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} Other examples are ReadSpeaker or BrowseAloud that add text-to-speech functionality to web content.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} The primary audience for such applications is those who have difficulty reading because of learning disabilities or language barriers.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} Although functionality remains limited compared to equivalent desktop applications, the major benefit is to increase the accessibility of said websites when viewed on public machines where users do not have permission to install custom software, giving people greater "freedom to roam".{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}}', 18 => 'With the development of smartphones, the ability to listen to written documents (textual web content, <abbr>PDF</abbr> documents, e-mails etc.) while driving or during a similar activity in the same way that listening to music, will benefit a much broader audience than visually-impaired people.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}} The best-known examples are Siri for iOS, and Google Now and Iris for Android. With the release of the Galaxy S III, Samsung also introduced a similar intelligent personal assistant called S Voice. On the BlackBerry 10 operating system, their Z30 smartphone also features spoken interaction features, which are similar to the other mobile operating systems.{{citation needed| date= {{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}| reason= Similarity unverified.}}', 19 => 'This functionality depends on the quality of the software but also on a logical structure of the text. Use of headings, punctuation, presence of alternate attributes for images, etc. is crucial for a good vocalization. Also a web site may have a nice look because of the use of appropriate two dimensional positioning with CSS but its standard linearization, for example, by suppressing any CSS and Javascript in the browser may not be comprehensible.{{citation needed|date={{subst:CURRENTMONTHNAME}} {{subst:CURRENTYEAR}}}}' ]
Lines removed in edit (removed_lines)
[ 0 => 'In early [[operating system]]s, such as [[MS-DOS]], which employed [[command-line interface]]s (<abbr>CLI</abbr>s), the screen display consisted of [[Character (computing)|characters]] [[Memory-mapped I/O|mapping]] directly to a [[screen buffer]] in [[Computer memory|memory]] and a [[Cursor (computers)|cursor]] position. Input was by keyboard. All this information could therefore be obtained from the system either by [[hooking]] the flow of information around the system and reading the screen buffer or by using a standard hardware output socket<ref>{{cite web', 1 => '| title=Talking Terminals. BYTE, September 1982', 2 => '| url=http://www.edstoffel.com/david/talkingterminals.html', 3 => '| access-date=September 7, 2006 |archive-url = https://web.archive.org/web/20060625225004/http://www.edstoffel.com/david/talkingterminals.html <!-- Bot retrieved archive --> |archive-date = June 25, 2006}}</ref> and communicating the results to the user.', 4 => 'In the 1980s, the Research Centre for the Education of the Visually Handicapped (<abbr>RCEVH</abbr>) at the [[University of Birmingham]] developed Screen Reader for the [[BBC Micro]] and <!--abbr?-->NEC Portable.<ref>Paul Blenkhorn, "The <abbr>RCEVH</abbr> project on micro-computer systems and computer assisted learning", British Journal of Visual Impairment, 4/3, 101-103 (1986). [http://www.visugate.biz/bjvi/1986/autumn1986.html#RCEVH Free HTML version at Visugate].</ref><ref>{{cite web | title=Access to personal computers using speech synthesis. RNIB New Beacon No.76, May 1992 | date=March 3, 2014| url=http://www.rnib.org.uk/information-everyday-living-using-technology-beginners-guides/beginners-guide-assistive-technology}}</ref>', 5 => 'With the arrival of [[graphical user interface]]s (<abbr>GUI</abbr>s), the situation became more complicated. A <abbr>GUI</abbr> has characters and graphics drawn on the screen at particular positions, and therefore there is no purely textual representation of the graphical contents of the display. Screen readers were therefore forced to employ new low-level techniques, gathering messages from the [[operating system]] and using these to build up an "off-screen model", a representation of the display in which the required text content is stored.<ref>According to "[ftp://service.boulder.ibm.com/sns/sr-os2/sr2doc/guitalk.txt Making the <abbr>GUI</abbr> Talk]" (by Richard Schwerdtfeger, ''BYTE'' December 1991, p. 118-128), the first screen reader to build an off-screen model was outSPOKEN.</ref>', 6 => 'For example, the operating system might send messages to draw a command button and its caption. These messages are intercepted and used to construct the off-screen model. The user can switch between controls (such as buttons) available on the screen and the captions and control contents will be read aloud and/or shown on refreshable [[Braille]] display.', 7 => 'Screen readers can also communicate information on menus, controls, and other visual constructs to permit blind users to interact with these constructs. However, maintaining an off-screen model is a significant technical challenge; hooking the low-level messages and maintaining an accurate model are both difficult tasks.{{citation needed|date=January 2015}}', 8 => 'Operating system and application designers have attempted to address these problems by providing ways for screen readers to access the display contents without having to maintain an off-screen model. These involve the provision of alternative and accessible representations of what is being displayed on the screen accessed through an <abbr>[[Application programming interface|API]]</abbr>. Existing <abbr>API</abbr>s include:', 9 => '* Android Accessibility Framework<ref>[https://developer.android.com/training/accessibility/index.html Implementing Accessibility on Android].</ref>', 10 => '* Apple Accessibility <abbr>API</abbr><ref>[https://developer.apple.com/documentation/Accessibility/Reference/AccessibilityLowlevel/index.html Apple Accessibility <abbr>API</abbr>].</ref>', 11 => '* [[AT-SPI]]', 12 => '* [[IAccessible2]]', 13 => '* [[Microsoft Active Accessibility]] (<abbr>MSAA</abbr>)', 14 => '* [[Microsoft UI Automation]]', 15 => '* [[Java Access Bridge]]<ref>{{cite web|url=http://java.sun.com/products/accessbridge/|title=Oracle Technology Network for Java Developers – Oracle Technology Network – Oracle}}</ref>', 16 => 'Screen readers can query the operating system or application for what is currently being displayed and receive updates when the display changes. For example, a screen reader can be told that the current focus is on a button and the button caption to be communicated to the user. This approach is considerably easier for the developers of screen readers, but fails when applications do not comply with the accessibility <abbr>API</abbr>: for example, [[Microsoft Word]] does not comply with the <abbr>MSAA</abbr> <abbr>API</abbr>, so screen readers must still maintain an off-screen model for Word or find another way to access its contents.{{citation needed|date=January 2015}} One approach is to use available operating system messages and application object models to supplement accessibility <abbr>API</abbr>s.', 17 => 'Screen readers can be assumed to be able to access all display content that is not intrinsically inaccessible. Web browsers, word processors, icons and windows and email programs are just some of the applications used successfully by screen reader users. However, according to some users,{{Who|date=January 2015}} using a screen reader is considerably more difficult than using a GUI, and many applications have specific problems resulting from the nature of the application (e.g. animations) or failure to comply with accessibility standards for the platform (e.g. Microsoft Word and Active Accessibility).{{citation needed|date=January 2015}}', 18 => 'Some programs and applications have voicing technology built in alongside their primary functionality. These programs are termed [[self-voicing]] and can be a form of [[assistive technology]] if they are designed to remove the need to use a screen reader.{{citation needed|date=January 2015}}', 19 => 'Some telephone services allow users to interact with the internet remotely. For example, TeleTender can read web pages over the phone and does not require special programs or devices on the user side.{{citation needed|date=January 2015}}', 20 => 'A relatively new development in the field is web-based applications like Spoken-Web that act as web portals, managing content like news updates, weather, science and business articles for visually-impaired or blind computer users.{{citation needed|date=January 2015}} Other examples are ReadSpeaker or [[BrowseAloud]] that add [[text-to-speech]] functionality to web content.{{citation needed|date=January 2015}} The primary audience for such applications is those who have difficulty reading because of learning disabilities or language barriers.{{citation needed|date=January 2015}} Although functionality remains limited compared to equivalent desktop applications, the major benefit is to increase the accessibility of said websites when viewed on public machines where users do not have permission to install custom software, giving people greater "freedom to roam".{{citation needed|date=January 2015}}', 21 => 'With the development of [[smartphone]]s, the ability to listen to written documents (textual web content, <abbr>[[Portable Document Format|PDF]]</abbr> documents, e-mails etc.) while driving or during a similar activity in the same way that listening to music, will benefit a much broader audience than visually-impaired people.{{citation needed|date=January 2015}} The best-known examples are [[Siri (software)|Siri]] for [[iOS]], and [[Google Now]] and [[Iris (software)|Iris]] for [[Android (operating system)|Android]]. With the release of the [[Samsung Galaxy S III|Galaxy S III]], [[Samsung]] also introduced a similar [[intelligent personal assistant]] called [[S Voice]]. On the [[BlackBerry 10 operating system voice control|BlackBerry 10 operating system]], their [[BlackBerry Z30|Z30 smartphone]] also features spoken interaction features, which are similar to the other [[mobile operating system]]s.{{citation needed| date= October 2013| reason= Similarity unverified.}}', 22 => 'This functionality depends on the quality of the software but also on a logical structure of the text. Use of headings, punctuation, presence of alternate attributes for images, etc. is crucial for a good vocalization. Also a web site may have a nice look because of the use of appropriate two dimensional positioning with CSS but its standard linearization, for example, by suppressing any CSS and Javascript in the browser may not be comprehensible.{{citation needed|date=January 2015}}' ]
Lines added in edit, pre-save transformed (added_lines_pst)
[ 0 => 'In early operating systems, such as MS-DOS, which employed command-line interfaces (<abbr>CLI</abbr>s), the screen display consisted of characters mapping directly to a screen buffer in memory and a cursor position. Input was by keyboard. All this information could therefore be obtained from the system either by hooking the flow of information around the system and reading the screen buffer or by using a standard hardware output socket and communicating the results to the user.', 1 => 'In the 1980s, the Research Centre for the Education of the Visually Handicapped (<abbr>RCEVH</abbr>) at the University of Birmingham developed Screen Reader for the BBC Micro and <!--abbr?-->NEC Portable.', 2 => 'With the arrival of graphical user interfaces (<abbr>GUI</abbr>s), the situation became more complicated. A <abbr>GUI</abbr> has characters and graphics drawn on the screen at particular positions, and therefore there is no purely textual representation of the graphical contents of the display. Screen readers were therefore forced to employ new low-level techniques, gathering messages from the operating system and using these to build up an "off-screen model", a representation of the display in which the required text content is stored.', 3 => 'For example, the operating system might send messages to draw a command button and its caption. These messages are intercepted and used to construct the off-screen model. The user can switch between controls (such as buttons) available on the screen and the captions and control contents will be read aloud and/or shown on refreshable Braille display.', 4 => 'Screen readers can also communicate information on menus, controls, and other visual constructs to permit blind users to interact with these constructs. However, maintaining an off-screen model is a significant technical challenge; hooking the low-level messages and maintaining an accurate model are both difficult tasks.{{citation needed|date=November 2021}}', 5 => 'Operating system and application designers have attempted to address these problems by providing ways for screen readers to access the display contents without having to maintain an off-screen model. These involve the provision of alternative and accessible representations of what is being displayed on the screen accessed through an <abbr>API</abbr>. Existing <abbr>API</abbr>s include:', 6 => '* Android Accessibility Framework', 7 => '* Apple Accessibility <abbr>API</abbr>', 8 => '* AT-SPI', 9 => '* I Accessible', 10 => '* Microsoft Active Accessibility (<abbr>MSAA</abbr>)', 11 => '* Microsoft UI Automation', 12 => '* Java Access Bridge', 13 => 'Screen readers can query the operating system or application for what is currently being displayed and receive updates when the display changes. For example, a screen reader can be told that the current focus is on a button and the button caption to be communicated to the user. This approach is considerably easier for the developers of screen readers, but fails when applications do not comply with the accessibility <abbr>API</abbr>: for example, Microsoft Word does not comply with the <abbr>MSAA</abbr> <abbr>API</abbr>, so screen readers must still maintain an off-screen model for Word or find another way to access its contents.{{citation needed|date=November 2021}} One approach is to use available operating system messages and application object models to supplement accessibility <abbr>API</abbr>s.', 14 => 'Screen readers can be assumed to be able to access all display content that is not intrinsically inaccessible. Web browsers, word processors, icons and windows and email programs are just some of the applications used successfully by screen reader users. However, according to some users,{{Who|date=November 2021}} using a screen reader is considerably more difficult than using a GUI, and many applications have specific problems resulting from the nature of the application (e.g. animations) or failure to comply with accessibility standards for the platform (e.g. Microsoft Word and Active Accessibility).{{citation needed|date=November 2021}}', 15 => 'Some programs and applications have voicing technology built in alongside their primary functionality. These programs are termed self-voicing and can be a form of assistive technology if they are designed to remove the need to use a screen reader.{{citation needed|date=November 2021}}', 16 => 'Some telephone services allow users to interact with the internet remotely. For example, TeleTender can read web pages over the phone and does not require special programs or devices on the user side.{{citation needed|date=November 2021}}', 17 => 'A relatively new development in the field is web-based applications like Spoken-Web that act as web portals, managing content like news updates, weather, science and business articles for visually-impaired or blind computer users.{{citation needed|date=November 2021}} Other examples are ReadSpeaker or BrowseAloud that add text-to-speech functionality to web content.{{citation needed|date=November 2021}} The primary audience for such applications is those who have difficulty reading because of learning disabilities or language barriers.{{citation needed|date=November 2021}} Although functionality remains limited compared to equivalent desktop applications, the major benefit is to increase the accessibility of said websites when viewed on public machines where users do not have permission to install custom software, giving people greater "freedom to roam".{{citation needed|date=November 2021}}', 18 => 'With the development of smartphones, the ability to listen to written documents (textual web content, <abbr>PDF</abbr> documents, e-mails etc.) while driving or during a similar activity in the same way that listening to music, will benefit a much broader audience than visually-impaired people.{{citation needed|date=November 2021}} The best-known examples are Siri for iOS, and Google Now and Iris for Android. With the release of the Galaxy S III, Samsung also introduced a similar intelligent personal assistant called S Voice. On the BlackBerry 10 operating system, their Z30 smartphone also features spoken interaction features, which are similar to the other mobile operating systems.{{citation needed| date= November 2021| reason= Similarity unverified.}}', 19 => 'This functionality depends on the quality of the software but also on a logical structure of the text. Use of headings, punctuation, presence of alternate attributes for images, etc. is crucial for a good vocalization. Also a web site may have a nice look because of the use of appropriate two dimensional positioning with CSS but its standard linearization, for example, by suppressing any CSS and Javascript in the browser may not be comprehensible.{{citation needed|date=November 2021}}' ]
New page wikitext, pre-save transformed (new_pst)
'{{short description|Assistive technology that converts text or images to speech or Braille}} {{More citations needed|date=July 2017}} [[File:Accessible Books Consortium explains - a digital file is not necessarily accessible.webm|thumb|An example of someone using a screen reader showing documents that are inaccessible, readable and accessible.]] {{Use mdy dates|date=July 2017}} A '''screen reader''' is a form of [[assistive technology]] (<abbr>AT</abbr>)<ref>{{cite web|url=https://www.microsoft.com/enable/at/types.aspx|title=Types of Assistive Technology Products|publisher=Microsoft Accessibility|access-date=13 June 2016}}</ref> that renders text and image content as speech or braille output. Screen readers are essential to people who are [[blindness|blind]],<ref name="Braille display"/> and are useful to people who are [[visual impairment|visually impaired]],<ref name="Braille display"/> [[Illiteracy|illiterate]], or have a [[learning disability]].<ref name="Screen1">{{cite web|url=http://www.vadsa.org/ace/reader.htm|title=Screen Readers and how they work with E-Learning|publisher=Virginia.gov|access-date=31 March 2019|archive-url=https://web.archive.org/web/20181113075826/https://www.vadsa.org/ace/reader.htm|archive-date=13 November 2018}}</ref> Screen readers are [[Application software|software applications]] that attempt to convey what people with normal eyesight see on a [[Display device|display]] to their users via non-visual means, like [[text-to-speech]],<ref>{{cite web|url=http://windows.microsoft.com/en-us/windows/hear-text-read-aloud-narrator#1TC=windows-8|title=Hear text read aloud with Narrator|publisher=[[Microsoft Office|Microsoft]]|access-date=13 June 2016}}</ref> sound icons,<ref>{{cite web|url=https://css-tricks.com/accessiblility-basics-turn-your-css-off/|title=Accessibility Basics: How Does Your Page Look To A Screen Reader?|last=Coyier|first=Chris|date=29 October 2007|publisher=CSS-Tricks|access-date=13 June 2016}}</ref> or a [[Refreshable Braille display|Braille device]].<ref name="Braille display">{{cite web|url=http://www.afb.org/info/living-with-vision-loss/using-technology/assistive-technology-videos/screen-reading-technology/1235|title=Screen Readers Technology and Refreshable Braille Displays|publisher=[[American Foundation for the Blind|AFB]]|access-date=13 June 2016}}</ref> They do this by applying a wide variety of techniques that include, for example, interacting with dedicated [[#Accessibility APIs|accessibility APIs]], using various [[operating system]] features (like [[inter-process communication]] and querying [[user interface]] properties), and employing [[hooking]] techniques.<ref name="SR Overview">{{cite web|url=https://www.nomensa.com/blog/2005/what-screen-reader|title=What is a Screen Reader|publisher=[[Nomensa]]|access-date=9 July 2017}}</ref> [[Microsoft Windows]] [[operating systems]] have included the [[Microsoft Narrator]] screen reader since [[Windows 2000]], though separate products such as [[Freedom Scientific]]'s commercially available [[JAWS (screen reader)|JAWS]] screen reader and [[ZoomText]] screen magnifier and the [[free and open source]] screen reader [[NonVisual Desktop Access|NVDA]] by NV Access are more popular for that operating system.<ref>{{cite web|url=https://webaim.org/projects/screenreadersurvey9/|title=Screen Reader User Survey #9|publisher=[[WebAIM]]|access-date=July 1, 2021}}</ref> [[Apple Inc.]]'s [[macOS]], [[iOS]], and [[tvOS]] include [[VoiceOver]] as a built-in screen reader, while [[Google]]'s [[Android (operating system)|Android]] provides the [[Google TalkBack|Talkback screen reader]] and its [[Chrome OS]] can use ChromeVox.<ref>{{cite web|url=http://www.chromevox.com/|title=ChromeVox|publisher=Google|access-date=March 9, 2020}}</ref> Similarly, Android-based devices from Amazon provide the VoiceView screen reader. There are also free and open source screen readers for [[Linux]] and [[Unix-like]] systems, such as Speakup and [[Orca (assistive technology)|Orca]]. == Types == === Command-line (text) === In early operating systems, such as MS-DOS, which employed command-line interfaces (<abbr>CLI</abbr>s), the screen display consisted of characters mapping directly to a screen buffer in memory and a cursor position. Input was by keyboard. All this information could therefore be obtained from the system either by hooking the flow of information around the system and reading the screen buffer or by using a standard hardware output socket and communicating the results to the user. In the 1980s, the Research Centre for the Education of the Visually Handicapped (<abbr>RCEVH</abbr>) at the University of Birmingham developed Screen Reader for the BBC Micro and <!--abbr?-->NEC Portable. === Graphical === ==== Off-screen models ==== With the arrival of graphical user interfaces (<abbr>GUI</abbr>s), the situation became more complicated. A <abbr>GUI</abbr> has characters and graphics drawn on the screen at particular positions, and therefore there is no purely textual representation of the graphical contents of the display. Screen readers were therefore forced to employ new low-level techniques, gathering messages from the operating system and using these to build up an "off-screen model", a representation of the display in which the required text content is stored. For example, the operating system might send messages to draw a command button and its caption. These messages are intercepted and used to construct the off-screen model. The user can switch between controls (such as buttons) available on the screen and the captions and control contents will be read aloud and/or shown on refreshable Braille display. Screen readers can also communicate information on menus, controls, and other visual constructs to permit blind users to interact with these constructs. However, maintaining an off-screen model is a significant technical challenge; hooking the low-level messages and maintaining an accurate model are both difficult tasks.{{citation needed|date=November 2021}} ==== Accessibility APIs ==== Operating system and application designers have attempted to address these problems by providing ways for screen readers to access the display contents without having to maintain an off-screen model. These involve the provision of alternative and accessible representations of what is being displayed on the screen accessed through an <abbr>API</abbr>. Existing <abbr>API</abbr>s include: * Android Accessibility Framework * Apple Accessibility <abbr>API</abbr> * AT-SPI * I Accessible * Microsoft Active Accessibility (<abbr>MSAA</abbr>) * Microsoft UI Automation * Java Access Bridge Screen readers can query the operating system or application for what is currently being displayed and receive updates when the display changes. For example, a screen reader can be told that the current focus is on a button and the button caption to be communicated to the user. This approach is considerably easier for the developers of screen readers, but fails when applications do not comply with the accessibility <abbr>API</abbr>: for example, Microsoft Word does not comply with the <abbr>MSAA</abbr> <abbr>API</abbr>, so screen readers must still maintain an off-screen model for Word or find another way to access its contents.{{citation needed|date=November 2021}} One approach is to use available operating system messages and application object models to supplement accessibility <abbr>API</abbr>s. Screen readers can be assumed to be able to access all display content that is not intrinsically inaccessible. Web browsers, word processors, icons and windows and email programs are just some of the applications used successfully by screen reader users. However, according to some users,{{Who|date=November 2021}} using a screen reader is considerably more difficult than using a GUI, and many applications have specific problems resulting from the nature of the application (e.g. animations) or failure to comply with accessibility standards for the platform (e.g. Microsoft Word and Active Accessibility).{{citation needed|date=November 2021}} === Self-voicing programs and applications === Some programs and applications have voicing technology built in alongside their primary functionality. These programs are termed self-voicing and can be a form of assistive technology if they are designed to remove the need to use a screen reader.{{citation needed|date=November 2021}} === Cloud-based === Some telephone services allow users to interact with the internet remotely. For example, TeleTender can read web pages over the phone and does not require special programs or devices on the user side.{{citation needed|date=November 2021}} === Web-based === A relatively new development in the field is web-based applications like Spoken-Web that act as web portals, managing content like news updates, weather, science and business articles for visually-impaired or blind computer users.{{citation needed|date=November 2021}} Other examples are ReadSpeaker or BrowseAloud that add text-to-speech functionality to web content.{{citation needed|date=November 2021}} The primary audience for such applications is those who have difficulty reading because of learning disabilities or language barriers.{{citation needed|date=November 2021}} Although functionality remains limited compared to equivalent desktop applications, the major benefit is to increase the accessibility of said websites when viewed on public machines where users do not have permission to install custom software, giving people greater "freedom to roam".{{citation needed|date=November 2021}} With the development of smartphones, the ability to listen to written documents (textual web content, <abbr>PDF</abbr> documents, e-mails etc.) while driving or during a similar activity in the same way that listening to music, will benefit a much broader audience than visually-impaired people.{{citation needed|date=November 2021}} The best-known examples are Siri for iOS, and Google Now and Iris for Android. With the release of the Galaxy S III, Samsung also introduced a similar intelligent personal assistant called S Voice. On the BlackBerry 10 operating system, their Z30 smartphone also features spoken interaction features, which are similar to the other mobile operating systems.{{citation needed| date= November 2021| reason= Similarity unverified.}} This functionality depends on the quality of the software but also on a logical structure of the text. Use of headings, punctuation, presence of alternate attributes for images, etc. is crucial for a good vocalization. Also a web site may have a nice look because of the use of appropriate two dimensional positioning with CSS but its standard linearization, for example, by suppressing any CSS and Javascript in the browser may not be comprehensible.{{citation needed|date=November 2021}} == Customization == Most screen readers allow the user to select whether most [[punctuation]] is announced or silently ignored. Some screen readers can be tailored to a particular application through [[script (computing)|'''scripting''']]. One advantage of scripting is that it allows customizations to be shared among users, increasing accessibility for all. <abbr>[[Job Access With Speech|JAWS]]</abbr> enjoys an active script-sharing community, for example.{{citation needed|date=January 2015}} === Verbosity === Verbosity is a feature of screen reading software that supports vision-impaired computer users. Speech verbosity controls enable users to choose how much speech feedback they wish to hear. Specifically, verbosity settings allow users to construct a mental model of web pages displayed on their computer screen. Based on verbosity settings, a screen-reading program informs users of certain formatting changes, such as when a frame or table begins and ends, where graphics have been inserted into the text, or when a list appears in the document.{{citation needed|date=January 2015}} === Language === Some screen readers can read text in more than one [[language]], provided that the language of the material is encoded in its [[Metadata Encoding and Transmission Standard|metadata]].<ref>{{cite web|url=https://developer.yahoo.com/blogs/ydn/yahoo-search-results-now-natural-language-support-7318.html|title=Yahoo! search results now with natural language support|date=March 13, 2008|author=Chris Heilmann|work=[[Yahoo! Developer Network]] Blog|access-date=February 28, 2015|archive-url=https://web.archive.org/web/20090125024422/http://developer.yahoo.net/blog/archives/2008/03/yahoo_search_re.html|archive-date=January 25, 2009|url-status=live}}</ref> Some screen reading programs{{Which|date=January 2015}} also include language verbosity, which automatically detects verbosity settings related to speech output language. For example, if a user navigated to a website based in the United Kingdom, the text would be read with [[British English|an English accent]].{{citation needed|date=January 2015}} == See also == * [[List of screen readers]] * [[Screen magnifier]] * [[Self-voicing]] * [[Speech processing]] * [[Speech recognition]] * [[Speech synthesis]] * [[Vinux]] * [[VoiceOver]] == References == {{Reflist}} == External links == {{Wiktionary}} * [http://www.standards-schmandards.com/projects/fangs Fangs screen reader emulator]{{snds}}an open-source Mozilla [[List of Firefox extensions|Firefox extension]] that simulates how a web page would look in [[JAWS (screen reader)|JAWS]] {{authority control}} [[Category:Assistive technology]] [[Category:Screen readers| ]]'
Whether or not the change was made through a Tor exit node (tor_exit_node)
false
Unix timestamp of change (timestamp)
1635758753