Jump to content

Modality (human–computer interaction)

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by AustinWerner358 (talk | contribs) at 22:45, 21 April 2015. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

In the context of human–computer interaction, a modality is the classification of a single independent channel of sensory input/output between a computer and a human.[1] A system is designated unimodel if it has only one modality implemented, and multimodel if it has more than one.[1] When multiple modalities are available for some tasks or aspects of a task, the system is said to have overlapping modalities. If multiple modalities are available for a task, the system is said to have redundant modalities.[citation needed] Multiple modalities can be used in combination to provide complimentary methods that may be redundant but convey information more effectively.[2] Modalities can be generally defined in two forms: human-computer and computer-human modalities.

Computer–Human Modalities

Any human sense can used as a computer to human modality. The following are examples of modalities and their implementations through which a computer could send information to a human:

The modalities of seeing and hearing are the most commonly employed since they are capable of transmitting more information at a higher speed than other modalities, 250 to 300[3] and 150–160[4] words per minute, respectively. Though uncommonly implemented as computer-human modality, tactition can achieve average 125 wpm [5] through the use of a refreshable Braille display.

Human–Computer Modalities

The computer can be equipped with various types of input devices and sensors to allow it to receive information from the human.

Benefits of Multimodal Systems

There are six types of relations between modalities, and they help define how a combination or fusion of modalities cooperate to convey information more effectively.

See also


  1. ^ a b Karray, Fakhreddine; Alemzadeh, Milad; Saleh, Jamil Abou; Arab, Mo Nours (March 2008). "Human-Computer Interaction: Overview on State of the Art" (PDF). International Journal on Smart Sensing and Intelligent Systems. 1 (1). Retrieved April 21, 2015.
  2. ^ Palanque, Philippe; Paterno, Fabio (2001). Interactive Systems. Design, Specification, and Verification. Springer Science & Business Media. p. 43. ISBN 9783540416630.
  3. ^ Ziefle, M (December 1998). "Effects of display resolution on visual performance". Human factors. 40 (4): 554–68. PMID 9974229.
  4. ^ Williams, J. R. (1998). Guidelines for the use of multimedia in instruction, Proceedings of the Human Factors and Ergonomics Society 42nd Annual Meeting, 1447–1451
  5. ^ "Braille". ACB. American Council of the Blind. Retrieved 21 April 2015.