Perceptual Invariance of Nonlinear Focus+Context Transformations

Abstract
Focus+Context techniques are commonly used in visualization systems to simultaneously provide both the details and the context of a particular dataset. This paper proposes a new methodology to empirically investigate the effect of various Focus+Context transformations on human perception. This methodology is based on the shaker paradigm, which tests performance for a visual task on an image that is rapidly alternated with a transformed version of itself. An important aspect of this technique is that it can determine two different kinds of perceptual cost: (i) the effect on the perception of a static transformed image, and (ii) the effect of the dynamics of the transformation itself. This technique has been successfully applied to determine the extent to which human perception is invariant to scaling and rotation [Rensink 2004]. In this paper, we extend this approach to examine nonlinear fisheye transformations of the type typically used in a Focus+Context system. We show that there exists a no-cost zone where performance is unaffected by an abrupt, noticeable fisheye transformation, and that its extent can be determined. The lack of perceptual cost in regards to these sudden changes contradicts the belief that they are necessarily detrimental to performance, and suggests that smoothly animated transformations between visual states are not always necessary. We show that this technique also can map out low-cost zones where transformations result in only a slight degradation of performance. Finally, we show that rectangular grids have no positive effect on performance, acting only as a form of visual clutter. These results therefore demonstrate that the perceptual costs of nonlinear transformations can be successfully quantified. Interestingly, they show that some kinds of sudden transformation can be experienced with minimal or no perceptual cost. This contradicts the belief that sudden changes are necessarily detrimental to performance, and suggests that smoothly animated transformations between visual states are not always necessary..
Keywords No keywords specified (fix it)
Categories No categories specified
(categorize this paper)
Options
 Save to my reading list
Follow the author(s)
My bibliography
Export citation
Find it on Scholar
Edit this record
Mark as duplicate
Revision history Request removal from index Translate to english
 
Download options
PhilPapers Archive


Upload a copy of this paper     Check publisher's policy on self-archival     Papers currently archived: 10,788
External links
Setup an account with your affiliations in order to access resources via your University's proxy server
Configure custom proxy (use this if your affiliation does not provide a proxy)
Through your library
References found in this work BETA

No references found.

Citations of this work BETA

No citations found.

Similar books and articles
Orlin Vakarelov (2012). The Information Medium. Philosophy and Technology 25 (1):47-65.
Tyler Burge (2003). Perceptual Entitlement. Philosophy and Phenomenological Research 67 (3):503-548.
John Heffner (1976). Some Epistemological Aspects of Recent Work in Visual Perception. PSA: Proceedings of the Biennial Meeting of the Philosophy of Science Association 1976:165 - 174.
Bence Nanay (2012). Perceptual Phenomenology. Philosophical Perspectives 26 (1):235-246.
Analytics

Monthly downloads

Added to index

2010-09-14

Total downloads

2 ( #345,621 of 1,099,035 )

Recent downloads (6 months)

1 ( #287,293 of 1,099,035 )

How can I increase my downloads?

My notes
Sign in to use this feature


Discussion
Start a new thread
Order:
There  are no threads in this forum
Nothing in this forum yet.