Skip to main content

OPINION article

Front. Psychol., 10 March 2020
Sec. Environmental Psychology
This article is part of the Research Topic Environment, Art, and Museums: The Aesthetic Experience in Different Contexts View all 11 articles

Using Visual Aesthetic Sensitivity Measures in Museum Studies

  • 1Department of Psychology, Pace University, New York City, NY, United States
  • 2Université de Paris, LAPEA, Boulogne-Billancourt, France
  • 3LAPEA, Univ. Gustave Eiffel, IFSTTAR, Versailles, France

For over a century, differential psychologists (e.g., Cattell, 1890; Eysenck, 1940), educational psychologists (e.g., Thorndike, 1916; Seashore, 1929) and art theorists (e.g., Graves, 1948; Götz, 1985) have attempted to capture one's ability to form judgments of aesthetic objects that agree with external standards defined by stimulus construction criteria, layperson consensus, and/or expert consensus. In the visual domain, this ability—generally discussed as visual aesthetic sensitivity (Child, 1964) and measured through (notably) the Visual Aesthetic Sensitivity Test (VAST; Götz, 1985), its revision (VAST-R; Myszkowski and Storme, 2017), the Meier Art Tests (MAT; Meier, 1928) and the Design Judgment Test (DJT; Graves, 1948)—has recently regained interest, but has been mainly studied through its relations with individual differences in art expertise, personality, and intelligence among adults (e.g., Furnham and Chamorro-Premuzic, 2004; Myszkowski et al., 2014), and has remained unstudied in museum settings. In this paper, we review the current state of research on the validity of visual aesthetic sensitivity tests, and propose how to best implement them in museum studies.

Elements of Validity of Visual Aesthetic Sensitivity Measures

Most frequently, visual aesthetic sensitivity tests operationalize Child's (1964) definition using “controlled alteration” (Meier, 1928, p. 188), a procedure which consists of deteriorating or creating an altered version of an aesthetic stimulus, and in presenting examinees with the altered and original stimuli, with the task of recognizing which is of better aesthetic quality. The construct validity of tests based on it are however controversial (Gear, 1986; Liu, 1990; Corradi et al., 2019), as it was notably argued that absolute aesthetic standards cannot exist, dismissing any operationalization of Child's definition. Nevertheless, the availability of absolute standards is not a necessary condition for the operationalization of Child's definition (Myszkowski et al., 2020): Aesthetic sensitivity tests rely instead on empirical standards, obtained through expert and/or laypeople consensus. Consequently, they compare an examinees' response with the typical response of experts—as originally suggested by Thorndike (1916)—or use expert agreement to select items—as used in the VAST. While using expert and/or laypeople consensus in lieu of absolute standards seems crude, it is actually common practice whenever correctness is not self-evident: It is for example used in the measurement of emotional intelligence (Mayer et al., 2003) or creativity (Amabile, 1982).

Still, using empirical standards poses the question of measurement (in)variance, especially across cultural backgrounds: Two artworks A and B may be aesthetically ordered as A > B for a group but as B < A for another. Fortunately, on that matter, studies of cultural measurement invariance—especially on the VAST (Iwawaki et al., 1979; Chan et al., 1980; Eysenck et al., 1984)—have provided encouraging results, with positive strong correlations between the item difficulties of the test across different groups differing in gender, age, and nationality (England, Japan, Hong Kong, Germany, and Singapore). More robust analyses (e.g., using differential item functioning), are certainly called for, but there is currently no empirical evidence of problematic measurement variance across cultures. We could speculate that the reason for this is that the controlled alteration method leads to examinees having to judge stimuli that are in the same (sub)category. Indeed, in visual aesthetic sensitivity tests, examinees do not compare Picasso's Guernica with Da Vinci's Mona Lisa—rather, they are asked to compare an original work of art with an almost identical (yet altered) version. Therefore, responding is less a matter of personal/cultural inclination regarding movements and styles, but more a matter of detecting an “out-of-tune” execution. It thus engages more the “ability to perform a set of basic perceptual analyses of the stimulus” (Myszkowski et al., 2014, p. 16) than one's ability to apply culturally relative norms.

Another sign of construct validity can be found in the concurrent validity of visual aesthetic sensitivity tests. This point is also quite controversial (Corradi et al., 2019; Myszkowski et al., 2020), but this is mainly because the nomological network of visual aesthetic sensitivity is yet to be clearly defined. Notably, Eysenck introduced confusion by originally discussing the construct as intelligence in the aesthetic domain (1940) to then speculate that the construct should be independent from intelligence (Frois and Eysenck, 1995)—which is contradicted in a recent meta-analysis (Myszkowski et al., 2018), which showed across 23 studies that its correlation with intelligence is significant and around 0.30. Nevertheless, one can reasonably expect that, as is found empirically, visual aesthetic sensitivity would be positively correlated with intelligence—because common cognitive processes are likely engaged in both measures (Myszkowski et al., 2018), and because it is common to observe relations between sensory perception in other domains and intelligence (e.g., Troche and Rammsayer, 2009)—or with personality traits like openness to aesthetics (Myszkowski et al., 2014)—because individuals with stronger interest in aesthetics may engage in more extensive processing, leading to higher accuracy, as it was for example found (Myszkowski, 2019) that, in these tests, response speed is negatively correlated with accuracy. Therefore, even though the nomological network of visual aesthetic sensitivity is not sufficiently (nor consistently) discussed, the pattern of relations between aesthetic sensitivity and other measures does suggest that visual aesthetic sensitivity measures present evidence of concurrent validity (Myszkowski et al., 2020).

These signs of validity could lead to a wide use of visual aesthetic sensitivity tests in the field where they would seem to belong: In contexts that naturally involve aesthetic judgments, such as museum visits. As they are however absent from museum studies, we will now discuss ways to facilitate their implementation in such contexts.

How to Measure Visual Aesthetic Sensitivity in Museum Contexts

Because several visual aesthetic sensitivity tests are still in use, a first challenge could be to select one. Although these tests have showed satisfactory internal consistency in recent studies—with satisfactory Cronbach's αs (Furnham and Chamorro-Premuzic, 2004; Myszkowski et al., 2014; Summerfeldt et al., 2015)—their unidimensionality—a condition to even investigate internal consistency—and thus also their structural validity are largely unstudied. An exception is the VAST-R, which has been showed to present unidimensionality and structural validity—with a satisfactory fit of unidimensional Item-Response Theory models (Myszkowski and Storme, 2017). In addition, the VAST (and VAST-R) items present better evidence of content validity with the selection of the correct items by unanimity of a panel of 8 art experts (Götz et al., 1979). Finally, evidence of measurement invariance (though limited) is only provided for the VAST(-R) items (as discussed previously). Therefore, based on the current state of research we would suggest to prefer the VAST-R to other tests.

A second issue relates to scoring. While it seems straightforward to use sum/average scoring here, since the items of such tests are pass-fail items and vary greatly in difficulty (Myszkowski and Storme, 2017), one would advise to instead use Item-Response Theory (IRT) scoring. Using IRT in scoring such tests presents several advantages, such as obtaining conditional standard errors, which allows to identify cases that have been unreliably measured, or accounting for the guessing phenomena present in these tests. Still, using IRT remains challenging: It often requires specific training absent from many curricula (Borsboom, 2006) and demands large sample sizes for accurate estimation, which are not easily found in museum studies. Hopefully, regarding the VAST-R (other tests have not yet been studied with IRT), correlations between person estimates from (well-fitting) IRT models and sum/average scores are near perfect (Myszkowski and Storme, 2017). Therefore, even though IRT scoring is preferable, should IRT modeling not be possible, one could still use sum or average scores as an excellent proxy for IRT factor scores.

Related to technological advances, although this point remains unstudied, there is no evidence that these tests perform any differently when taken on-screen vs. in paper-and-pencil form: Both have been used indifferently. While measurement invariance between administration modalities needs empirical investigations, we could speculate that the two are equivalent. Actually, it may be more convenient in museum or virtual museum contexts to use tablets or computers for administration—smartphone screens are likely too small for properly displaying stimuli—and as we later suggest, there are psychometric advantages to using on-screen testing.

The use of computerized assessment first presents the practical advantage of allowing to reduce test length without compromising reliability, which would be desirable in assessing museum visitors. Because IRT models fit the VAST-R well (Myszkowski and Storme, 2017), researchers could use a Computerized Adaptive Testing (CAT) modified VAST-R, in which examinees would only take a subset of items that matches to their ability—re-estimated after each item—stopping assessment when such ability is estimated reliably enough (Green et al., 1984). The use of CAT is now largely facilitated by the availability of more software packages (e.g., Chalmers, 2016), and future studies may examine its usability with aesthetic sensitivity tests.

Further, as response times can be routinely collected when using computerized tests, we may suggest that recent IRT modeling advances in joint response and response time modeling could also allow to use response times as collateral information in the estimation of one's ability. Indeed, recent research (Myszkowski, 2019) suggests that there are strong dependencies between responses and response times (both related to a persons' speed and ability and to an item's difficulty and time intensity), which suggests that response times may be used to, for example, improve the accuracy of one's ability score, especially when fewer items are used (van der Linden et al., 2010). This could allow for even shorter tests, along with the improved detection of aberrant response/response times patterns (Marianti et al., 2014). As accuracy and speed are negatively correlated in the VAST-R, it has been also suggested (Myszkowski, 2019) to consider computing visual aesthetic sensitivity scores (accuracy scores) that are statistically controlled for response speed. This point is especially relevant for museum studies, because it is probably more likely to collect rushed responses from museum visitors than in experimental settings.

Finally, although we proposed that the VAST-R is the test that should currently be preferred, its content—black and white formal abstract paintings by Karl Otto Götz—remains rather narrow, and one may question the generalizability of the results of the test to other art styles and movements. We thus suggest that ad-hoc tests be built on a case-by-case basis using the controlled alteration procedure. One could for example use image modification software to alter artworks from the very exhibit studied and create stimuli pairs. In museum studies contexts, it would in fact probably be easier to identify subject matter experts to ensure content validity. The expert panel would then be asked which stimuli of the pair is of higher aesthetic quality, and one would select items where there is a strong or unanimous agreement (Götz et al., 1979) or keep all items and score as a function of a respondent's agreement with the expert consensus (Thorndike, 1916).

Conclusion

In over a century of research, visual aesthetic sensitivity testing has slowly advanced toward offering test material that finally presents encouraging—although fragile—signs of validity. Both psychometric research in visual aesthetic sensitivity testing and museum research could benefit from the implementation of these tests in museum contexts. For the former, we think that it could lead to clarifying the real-world implications of visual aesthetic sensitivity; for the latter, it could prove an important factor in the understanding of individual differences between museum visitors. While speculatory at this stage, the findings previously discussed could, for example, lead to hypothesize high aesthetic sensitivity individuals to be more engaged, reflective and attentive when visiting museums and viewing artworks, to demand more cognitive stimulation (with, for example, more contextual explanations), to make longer museum visits, to compare artworks more extensively, and to be more critical of exhibited artworks. We could thus anticipate visual aesthetic sensitivity tests to be useful in better understanding the traits of a museum's or an exhibition's audience—in both understanding who the typical visitor is, and in how different the visitors may be in their approach to art—and it thus may be useful in tailoring the museum experience to better anticipate and respond to the visitors' characteristics.

Author Contributions

NM proposed and drafted the paper. NM and FZ participated in its conceptualization and FZ made important modifications to enhance the quality of the paper.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

Amabile, T. M. (1982). Social psychology of creativity: a consensual assessment technique. J. Pers. Soc. Psychol. 43, 997–1013. doi: 10.1037/0022-3514.43.5.997

CrossRef Full Text | Google Scholar

Borsboom, D. (2006). The attack of the psychometricians. Psychometrika 71, 425–440. doi: 10.1007/s11336-006-1447-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Cattell, J. M. (1890). Mental tests and measurements. Mind 15, 373–381. doi: 10.1093/mind/os-XV.59.373

CrossRef Full Text

Chalmers, R. P. (2016). Generating adaptive and non-adaptive test interfaces for multidimensional item response theory applications. J. Stat. Softw. 71, 1–38. doi: 10.18637/jss.v071.i05

CrossRef Full Text | Google Scholar

Chan, J., Eysenck, H. J., and Götz, K. O. (1980). A new visual aesthetic sensitivity test: III. crosscultural comparison between Hong Kong children and adults, and english and japanese samples. Percept. Mot. Ski. 50(3 Pt 2), 1325–1326. doi: 10.2466/pms.1980.50.3c.1325

CrossRef Full Text | Google Scholar

Child, I. L. (1964). Observations on the meaning of some measures of esthetic sensitivity. J. Psychol. 57, 49–64. doi: 10.1080/00223980.1964.9916671

PubMed Abstract | CrossRef Full Text | Google Scholar

Corradi, G., Chuquichambi, E. G., Barrada, J. R., Clemente, A., and Nadal, M. (2019). A new conception of visual aesthetic sensitivity. Br. J. Psychol. doi: 10.1111/bjop.12427. [Epub ahead of print].

PubMed Abstract | CrossRef Full Text | Google Scholar

Eysenck, H. J. (1940). The general factor in aesthetic judgements. Br. J. Psychol. Gen. Sec. 31, 94–102. doi: 10.1111/j.2044-8295.1940.tb00977.x

CrossRef Full Text | Google Scholar

Eysenck, H. J., Götz, K. O., Long, H. Y., Nias, D. K. B., and Ross, M. (1984). A new visual aesthetic sensitivity test: IV. cross-cultural comparisons between a Chinese sample from singapore and an english sample. Pers. Individ. Differ. 5, 599–600. doi: 10.1016/0191-8869(84)90036-9

CrossRef Full Text | Google Scholar

Frois, J. P., and Eysenck, H. J. (1995). The visual aesthetic sensitivity test applied to portuguese children and fine arts students. Creat. Res. J. 8, 277–284. doi: 10.1207/s15326934crj0803_6

CrossRef Full Text | Google Scholar

Furnham, A., and Chamorro-Premuzic, T. (2004). Personality, intelligence, and art. Pers. Individ. Differ. 36, 705–715. doi: 10.1016/S0191-8869(03)00128-4

CrossRef Full Text | Google Scholar

Gear, J. (1986). Eysenck's visual aesthetic sensitivity test (VAST) as an example of the need for explicitness and awareness of context in empirical aesthetics. Poetics 15, 555–564. doi: 10.1016/0304-422X(86)90011-2

CrossRef Full Text | Google Scholar

Götz, K. O. (1985). VAST: Visual Aesthetic Sensitivity Test, 4th Edn. Düsseldorf: Concept Verlag.

Götz, K. O., Borisy, A. R., Lynn, R., and Eysenck, H. J. (1979). A new visual aesthetic sensitivity test: I. construction and psychometric properties. Percept. Mot. Ski. 49, 795–802. doi: 10.2466/pms.1979.49.3.795

CrossRef Full Text | Google Scholar

Graves, M. E. (1948). Design Judgment Test. New York, NY: Psychological Corporation.

Google Scholar

Green, B. F., Bock, R. D., Humphreys, L. G., Linn, R. L., and Reckase, M. D. (1984). Technical guidelines for assessing computerized adaptive tests. J. Educ. Meas. 21, 347–360. doi: 10.1111/j.1745-3984.1984.tb01039.x

CrossRef Full Text | Google Scholar

Iwawaki, S., Eysenck, H. J., and Götz, K. O. (1979). A new visual aesthetic sensitivity test: II. cross-cultural comparison between England and Japan. Percept. Mot. Ski. 49, 859–862. doi: 10.2466/pms.1979.49.3.859

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, F.-J. (1990). Critique of three tests of aesthetic judgment; maitland graves design judgment test; the meier art tests: I, art judgment; and the meier art tests: II, aesthetic perception. Vis. Arts Res. 16, 90–99.

Google Scholar

Marianti, S., Fox, J.-P., Avetisyan, M., Veldkamp, B. P., and Tijmstra, J. (2014). Testing for aberrant behavior in response time modeling. J. Educ. Behav. Stat. 39, 426–451. doi: 10.3102/1076998614559412

CrossRef Full Text | Google Scholar

Mayer, J. D., Salovey, P., Caruso, D. R., and Sitarenios, G. (2003). Measuring emotional intelligence with the MSCEIT V2.0. Emotion 3, 97–105. doi: 10.1037/1528-3542.3.1.97

PubMed Abstract | CrossRef Full Text | Google Scholar

Meier, N. C. (1928). A measure of art talent. Psychol. Monogr. 39, 184–199. doi: 10.1037/h0093346

CrossRef Full Text | Google Scholar

Myszkowski, N. (2019). The first glance is the weakest: “Tasteful” individuals are slower to judge visual art. Pers. Individ. Differ. 141, 188–195. doi: 10.1016/j.paid.2019.01.010

CrossRef Full Text | Google Scholar

Myszkowski, N., Çelik, P., and Storme, M. (2018). A meta-analysis of the relationship between intelligence and visual “taste” measures. Psychol. Aesthet. Creat. Arts 12, 24–33. doi: 10.1037/aca0000099

CrossRef Full Text | Google Scholar

Myszkowski, N., Çelik, P., and Storme, M. (2020). Commentary on corradi et al.'s new conception of aesthetic sensitivity: is the ability conception dead? Br. J. Psychol. doi: 10.1111/bjop.12440. [Epub ahead of print].

PubMed Abstract | CrossRef Full Text | Google Scholar

Myszkowski, N., and Storme, M. (2017). Measuring “Good Taste” with the visual aesthetic sensitivity test-revised (VAST-R). Pers. Individ. Differ. 117, 91–100. doi: 10.1016/j.paid.2017.05.041

CrossRef Full Text | Google Scholar

Myszkowski, N., Storme, M., Zenasni, F., and Lubart, T. (2014). Is visual aesthetic sensitivity independent from intelligence, personality and creativity? Pers. Individ. Differ. 59, 16–20. doi: 10.1016/j.paid.2013.10.021

CrossRef Full Text | Google Scholar

Seashore, C. E. (1929). Meier-seashore art judgment test. Science 69, 380–380. doi: 10.1126/science.69.1788.380

PubMed Abstract | CrossRef Full Text | Google Scholar

Summerfeldt, L. J., Gilbert, S. J., and Reynolds, M. (2015). Incompleteness, aesthetic sensitivity, and the obsessive-compulsive need for symmetry. J. Behav. Ther. Exp. Psychiatry 49, 141–149. doi: 10.1016/j.jbtep.2015.03.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Thorndike, E. L. (1916). Tests of esthetic appreciation. J. Educ. Psychol. 7, 509–522. doi: 10.1037/h0073375

CrossRef Full Text | Google Scholar

Troche, S. J., and Rammsayer, T. H. (2009). The influence of temporal resolution power and working memory capacity on psychometric intelligence. Intelligence 37, 479–486. doi: 10.1016/j.intell.2009.06.001

CrossRef Full Text | Google Scholar

van der Linden, W. J., Klein Entink, R. H., and Fox, J.-P. (2010). IRT parameter estimation with response times as collateral information. Appl. Psychol. Measur. 34, 327–347. doi: 10.1177/0146621609349800

CrossRef Full Text | Google Scholar

Keywords: aesthetic abilities, aesthetic sensitivity, good taste, museum experience, psychological measurement

Citation: Myszkowski N and Zenasni F (2020) Using Visual Aesthetic Sensitivity Measures in Museum Studies. Front. Psychol. 11:414. doi: 10.3389/fpsyg.2020.00414

Received: 07 September 2019; Accepted: 24 February 2020;
Published: 10 March 2020.

Edited by:

Pablo P. L. Tinio, Montclair State University, United States

Reviewed by:

Eva Specker, University of Vienna, Austria
Elizabeth Vallance, Indiana University, United States

Copyright © 2020 Myszkowski and Zenasni. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Nils Myszkowski, bm15c3prb3dza2kmI3gwMDA0MDtwYWNlLmVkdQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.