Research Reports

Effect of Congruence Between Sound and Video on Heart Rate and Self-Reported Measures of Emotion

Sebastián Calderóna, Raúl Rincóna, Andrés Araujob, Carlos Gantiva*b

Europe's Journal of Psychology, 2018, Vol. 14(3), 621–631, https://doi.org/10.5964/ejop.v14i3.1593

Received: 2018-01-25. Accepted: 2018-04-06. Published (VoR): 2018-08-31.

Handling Editors: Vlad Glăveanu, Department of Psychology, Webster University Geneva, Geneva, Switzerland; Rhian Worth, University of South Wales, Newport, United Kingdom

*Corresponding author at: Universidad de San Buenaventura, Bogotá, Colombia, Department of Psychology, Neuroscience and Neuropsychology Laboratory, Address: Cr 8 H # 172-20. E-mail: cgantiva@usbbog.edu.co

This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/3.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Most studies of emotional responses have used unimodal stimuli (e.g., pictures or sounds) or congruent bimodal stimuli (e.g., video clips with sound), but little is known about the emotional response to incongruent bimodal stimuli. The aim of the present study was to evaluate the effect of congruence between auditory and visual bimodal stimuli on heart rate and self-reported measures of emotional dimension, valence and arousal. Subjects listened to pleasant, neutral, and unpleasant sounds, accompanied by videos with and without content congruence, and heart rate was recorded. Dimensions of valence and arousal of each bimodal stimulus were then self-reported. The results showed that heart rate depends of the valence of the sounds but not of the congruence of the bimodal stimuli. The valence and arousal scores changed depending on the congruence of the bimodal stimuli. These results suggest that the congruence of bimodal stimuli affects the subjective perception of emotion.

Keywords: bimodal stimuli, congruence, heart rate, valence, arousal

Emotion is a prerequisite for action that results from the activation of one of two primary motivational systems (i.e., appetitive and defensive) by the perception of a stimulus (Lang, 1995). When the appetitive motivational system is activated, approach behaviors are generated, with the goal of consummation, procreation, or nurturance. Conversely, when the defensive motivational system is activated, defense behaviors to protect the organism from threat are generated. The activation of these two systems modulates reflex responses (e.g., heart rate) that are compatible or incompatible with the required behavior (Alpers, Adolph, & Pauli, 2011). Emotional reactions to external cues are modulated by valence and arousal dimensions. Valence dimension (i.e., pleasure or displeasure) is related to the active motivational system, and arousal dimension is related to the intensity of emotional response (Bradley, Codispoti, Cuthbert, & Lang, 2001). The heart rate response is one of the most used physiological measures in studies of emotion (Barry, 1984; Berntson, Quigley, & Lozano, 2007; Lang, 1995; Obrist, 1981; Turpin, 1986). Heart rate is sensitive to the interaction between valence and arousal (Bradley et al., 2001) and is a reliable indicator of attentional engagement that is generated by the stimuli (Bradley, 2009; Campbell, Wood, & McBride, 1997).

Previous studies have reported the effects of affective stimulus presentation on heart rate. Bradley et al. (2001), found a pattern of initial deceleration of heart rate during the presentation of affective pictures. Higher deceleration occurred in response to unpleasant and pleasant pictures with high arousal compared with neutral pictures. Similar results were found with the presentation of affective sounds (i.e., initial deceleration, followed by acceleration and deceleration). Unpleasant sounds that evoked high arousal generated greater initial heart rate deceleration than pleasant sounds that evoked high arousal and neutral sounds (Bradley & Lang, 2000). These results suggest that initial heart rate deceleration occurs as a result of the response of attention that is generated by the interaction between valence and arousal and parasympathetic dominance of the autonomic nervous system (Bradley & Lang, 2007a).

Studies with motion pictures compared with still pictures have reported greater heart rate deceleration while viewing motion pictures, and movement affected self-reported valence and arousal (Detenber, Simons, & Bennett, 1998; Simons, Detenber, Roedema, & Reiss, 1999). Studies with videos found greater heart rate deceleration in response to unpleasant film clips compared with pleasant and neutral film clips (Bos, Jentgens, Beckers, & Kindt, 2013; Codispoti, Surcinelli, & Baldaro, 2008; Gomez, Zimmermann, Guttormsen-Schär, & Danuser, 2005). However, when the pictures contained a phobic stimulus (e.g., snakes or spiders; Hamm, Cuthbert, Globisch, & Vaitl, 1997) or when the videos contained an explicit threat or a phobic stimulus (Courtney, Dawson, Schell, Iyer, & Parsons, 2010; Palomba, Sarlo, Angrilli, Mini, & Stegagno, 2000), initial heart rate acceleration was observed, consistent with a sympathetic response of the autonomic nervous system.

In everyday life, emotional experience is often influenced by both visual and auditory information. The bimodal integration of these stimuli is important for understanding the emotional experience. Studies of emotional responses to bimodal stimuli have been performed using self-report measures and physiological responses. For example, Cox (2008) used self-report measures to evaluate the interaction between unpleasant sounds and pictures that were either associated or not associated with the sounds. This author found that pictures that were associated with sounds (with the exception of disgusting sounds) increased the level of sound aversion in some cases. Scherer and Larsen (2011) evaluated the effect of cross-modal priming, a phenomenon that occurs when a prime that is presented in one sensory modality (e.g., auditory) influences responses to a target in a different sensory modality (e.g., visual). They used affective sounds (i.e., pleasant and unpleasant) as a prime for affective visual stimuli (i.e., pleasant and unpleasant words) and found significant cross-modal priming effects but only for unpleasant primes (i.e., the participants evaluated the positive words as negative after listening to an unpleasant sound).

Studies that have used electrophysiological measures showed that event-related potential (P100 and P200) amplitudes increased in response to all types of pictures (i.e., pleasant, neutral, and unpleasant) that were accompanied by emotional sounds compared with pictures that were accompanied by neutral sounds (Gerdes et al., 2013). Brouwer, van Wouwe, Mühl, van Erp, and Toet (2013) presented unimodal and bimodal stimuli that were composed of affective pictures and sounds. Their results showed no differences in heart rate, heart rate variability, galvanic skin response, or self-report measures in response to pleasant and unpleasant stimuli, independent of the modality (i.e., unimodal or bimodal). The authors recognized that these results may be attributable to the fact that the bimodal stimuli were not optimally congruent.

Neuroimaging studies have shown that the processing of incongruent bimodal stimuli requires greater cognitive control, reflected by activation of the inferior frontal cortex. The processing of congruent bimodal stimuli was shown to activate temporal regions (Belardinelli et al., 2004; van Atteveldt, Formisano, Goebel, & Blomert, 2007). From a functional perspective, these data imply a greater investment of cognitive resources to evaluate incongruent bimodal stimuli compared with congruent bimodal stimuli (Doehrmann & Naumer, 2008).

Research that evaluates congruence between bimodal stimuli has been limited because only sounds and pictures are typically used (Gerdes, Wieser, & Alpers, 2014), thus generating less of an emotional response than videos (Courtney et al., 2010; Detenber et al., 1998; Simons et al., 1999). Such bimodal stimuli (pictures and sounds) are also less familiar compared with everyday experiences, which often contain movement stimuli. Additionally, the most commonly used affective stimuli are pictures that are contained in the International Affective Picture System (IAPS; Lang, Bradley, & Cuthbert, 2008) and sounds that are contained in the International Affective Digitized Sound System (IADS; Bradley & Lang, 2007b). Finding absolute congruence between these two types of stimuli is very difficult because each stimulus group was built independently (i.e., the IAPS has over 1000 pictures, and the IADS has 167 sounds). Furthermore, the sounds are dynamic, and the pictures are static, which limits evaluations of the effects of congruence between bimodal stimuli.

The objective of the present study was to evaluate the effect of congruence between bimodal stimuli that were composed of emotional sounds in the IADS (Bradley & Lang, 2007b) and videos on heart rate and self-report measures of valence and arousal. Our hypothesis was that congruent bimodal stimuli that are more realistic and captivating than incongruent bimodal stimuli (because they make more sense to the subject) would have a stronger emotional impact, reflected by subjective measures. In contrast, incongruent bimodal stimuli require a higher workload to be processed and thus would generate heart rate acceleration. The results of this study will provide evidence of the mechanisms by which information from the environment is integrated.

Methods

Participants

Thirty right-handed undergraduate students (24 males and six females), ranging in age from 20 to 27 years (M = 22.67 years, SD = 2.26 years), participated in the study. The participants were recruited from a university in Bogota, Colombia. The exclusion criteria were current medical or psychological treatment and visual/auditory problems without correction. All of the participants provided written informed consent. The study was approved by the University of San Buenaventura Review Board.

Stimuli and Procedure

Thirty sounds (10 pleasant with high arousal, 10 neutral with low arousal, and 10 unpleasant with high arousal) were selected from the IADSi according to normative ratings that were proposed by Bradley and Lang (2007b). Significant differences were found between sound categories in the valence dimension (all p < .001). For the arousal dimension, significant differences were found between affective sounds (pleasant and unpleasant) and neutral sounds (all p = .001). Pleasant and unpleasant sounds were not significantly different in the arousal dimension (p > .05). Sixty videos were selected from different web pages (e.g., YouTube or Google videos). Half of the videos had high congruence with the selected IADS sounds, and the other half had no congruence.

To obtain bimodal stimuli, videos with and without content congruence were added to each sound. Bimodal stimuli with congruence had credibility between the sources of the sound and video (e.g., for the sound of a roller coaster going downhill, the congruent video showed a rollercoaster going downhill). In contrast, bimodal stimuli without congruence had no credibility between the sources of the sound and video (e.g., for the sound of a roller coaster going downhill, the incongruent video showed a roller coaster going uphill). Before the experiment, each bimodal stimulus was evaluated by 52 subjects on a scale of 1 to 9 (1 = complete non-congruence, and 9 = complete congruence). Congruent bimodal stimuli had a mean score of 8.03 (SD = .40). Incongruent bimodal stimuli had a mean score of 1.55 (SD = .75). None of the subjects who participated in the initial evaluation of the bimodal stimuli participated in the experiment. Sounds and videos were edited using Adobe Premiere Pro CC software. Each bimodal stimulus pair was presented for 6 s at a resolution of 1280 × 720 pixels and frame rate of 24 frames per second.

The 60 bimodal stimuli were presented on a 19-inch flat-screen monitor that was located approximately 60 cm from the subject and through headphones. The sound level was 65 dB. Two different pseudo-randomized bimodal stimulus presentation orders were prepared (each order was applied to 15 subjects). Each order had the constraint of not presenting the same sound category consecutively and not presenting congruent or incongruent bimodal stimuli consecutively. Each bimodal stimulus pair was presented once. After each bimodal stimulus pair presentation, the emotional experience of each participant (i.e., valence and arousal ratings) was evaluated using digital versions of the Self-Assessment Manikin (SAM; Bradley & Lang, 1994). The intertrial interval (ITI) varied randomly from 10 to 14 s.

Self-Report Measures

Valence and arousal ratings were obtained using digitized versions of the SAM (Bradley & Lang, 1994), which depicts a graphic figure that varies along two dimensions of pleasure and arousal on a 9-point scale. The SAM ranges from smiling and happy (9) to frowning and unhappy (1) when representing the valence dimension. For the arousal dimension, the SAM ranges from excited and wide eyed (9) to relaxed and sleepy (1).

Data Acquisition and Reduction

The bimodal stimuli were presented using E-Prime 2.0 software (Psychology Software Tools, PA, USA). Electrocardiogram (EKG) recordings were continuously collected at a sampling rate of 1000 Hz with a 10-500 Hz frequency band filter using PowerLab 26T equipment (ADInstruments, Australia) using three large Sensormedic electrodes filled with electrolyte paste at lead III. All of the EKG data was manually inspected prior to analysis. Heart rate was derived from the EKG. R-wave–to–R-wave (R-R) intervals were extracted from the EKG using LabChart 7.3 software (ADInstruments, Australia), which were then imported into Matlab 8.0 (MathWorks, MA, USA) for analysis. Heart rate responses were determined by averaging across each half-second during the 6-s bimodal stimulus presentation and subtracting that activity from the activity that was obtained 1 s before bimodal stimulus onset.

Statistical Analysis

Heart rate responses were analyzed using a 3 × 2 × 12 repeated-measures analysis of variance (RM-ANOVA), with Sound (pleasant, neutral, and unpleasant), Video Congruence (with and without), and Time (the 12 half-second time bins throughout the duration of bimodal stimulus presentation) as the repeated measures. SAM ratings were analyzed using a 3 × 2 ANOVA, with Sound and Video Congruence as the repeated measures. Greenhouse-Geisser correction was applied to correct any violation of sphericity. Post hoc analyses of mean values were performed using the paired multiple-comparison Least Significant Difference (LSD) test. The level of significance for all of the analyses was p < .05. The partial eta squared (η2p) is also presented. All of the analyses were performed using SPSS 20.0 software.

Results

Heart Rate Response

The ANOVA revealed a significant main effect of Time (F(11, 319) = 3.689, p = .041, η2p = .11). We identified two deceleration phases: one phase between 0.5 and 1.5 s (with significant differences between 0.5 s and 1 and 1.5 s; both p < .02) and one phase between 3 and 5.5 s (with significant differences between 3 s and 5 and 5.5 s; both p < .02). Significant differences at all time points between 0.5 s and 3.5-6 s were found (all p < .05). A trend analysis was performed to identify the pattern of heart rate responses, showing a significant linear trend (F(1, 29) = 4.752, p = .038, η2p = .14).

A significant Sound × Time interaction was found (F(22, 638) = 2.574, p = .049, η2p = .08), suggesting that the valence of the bimodal stimuli generated differential heart rate response patterns at different time points. The deceleration of heart rate was greater during the presentation of unpleasant bimodal stimuli compared with neutral bimodal stimuli between 1.5 and 2.5 s (all p < .02). Similarly, the deceleration of heart rate during the presentation of pleasant sounds was greater compared with neutral sounds between 2 and 3 s (all p < .02). At 4.5 s, a significant acceleration of heart rate was found during the presentation of unpleasant sounds compared with pleasant sounds (p = .04; Figure 1). A significant linear trend was found for pleasant (F(1, 29) = 6.011, p = .02, η2p = .17) and neutral (F(1, 29) = 9.761, p = .004, η2p = .25) sounds. No significant main effects or interactions were found for Video Congruence (all p > .05).

ejop.v14i3.1593-f1
Figure 1

Heart rate response to sounds. bpm, beats per minute.

Subjective Ratings

Valence

The ANOVA of the valence dimension revealed a significant main effect of Sound (F(2, 58) = 132.41, p < .001, η2p = .82). As expected, significant differences were found between pleasant, neutral, and unpleasant sounds (all p < .001). A significant main effect of Video Congruence was found (F(1, 29) = 15.13, p = .001, η2p = .34). Sounds without video congruence had lower valence scores (p = .001). A significant Sound × Video Congruence interaction was found (F(2, 58) = 11.17, p < .001, η2p = .27). Pleasant sounds were evaluated as less appetitive when they had no video congruence (p < .001; Figure 2).

Arousal

The ANOVA of the arousal dimension revealed a significant main effect of Sound (F(2, 58) = 75.04, p < .001, η2p = .72). As expected, pleasant and unpleasant sounds were perceived as more arousing than neutral sounds (both p < .001). A significant main effect of Video Congruence was found (F(1, 29) = 22.52, p < .001, η2p = .43). Sounds without video congruence were perceived as less arousing (p < .001). A significant Sound × Video Congruence interaction was found (F(2, 58) = 17.22, p < .001, η2p = .37). Pleasant and unpleasant sounds without video congruence were perceived as less arousing than pleasant and unpleasant sounds with video congruence (both p < .002; Figure 2).

ejop.v14i3.1593-f2
Figure 2

Subjective ratings of valence and arousal for the sounds with and without video congruence. Bars indicate the standard error of the mean.

*p < .05.

Discussion

The objective of the present study was to evaluate the effect of congruence between bimodal stimuli (emotional sounds that are contained in the IADS and videos) on heart rate and self-reported measures of valence and arousal. The results suggested that the congruence between the sound and the video did not affect the heart rate response but affected the subjective evaluation of valence and arousal. Pleasant sounds without video congruence were evaluated as less appetitive, and pleasant and unpleasant sounds without video congruence were perceived as less arousing. These results suggest that the absence of congruence between bimodal stimuli changes the subjective emotional experience (i.e., it diminishes the level of arousal of the affective stimuli). In contrast, congruent stimuli are more real and captivating, leading to stronger emotional effects (i.e., greater arousal).

The absence of congruence also affected the valence of pleasant sounds but not the valence of unpleasant sounds. These data suggest a functional bias toward aversive stimulation because greater responsiveness to aversive stimuli than to appetitive stimuli is more adaptive for survival (Bradley & Lang, 2007a; Lang, Davis, & Öhman, 2000), independent of their congruence. Therefore, the valence of unpleasant sounds is more important than the congruence between auditory and visual stimulation. This bias toward aversive stimuli has been found in several studies that used affective pictures (Bradley et al., 2001; Gantiva & Camacho, 2016; Gantiva, Guerra, & Vila, 2011; Vila et al., 2001) and affective sounds (Bradley & Lang, 2000; Fernández-Abascal et al., 2008).

Congruence did not significantly affect heart rate, but a greater initial deceleration was observed in response to pleasant and unpleasant sounds compared with neutral sounds, indicating a greater attention response that was caused by the interaction between valence and arousal and dominance of the parasympathetic autonomic nervous system (Bradley & Lang, 2007a), at least during the first phase of bimodal stimulus presentation. Similar results have been reported by studies that used pictures (Bradley et al., 2001), videos (Bos et al., 2013; Codispoti et al., 2008; Gomez et al., 2005), and affective sounds (Bradley & Lang, 2000).

However, the unpleasant sounds subsequently generated an acceleration of heart rate. This acceleration was not observed when pleasant and neutral sounds were presented, indicating activation of the sympathetic autonomic nervous system that was associated with a motivational and behavioral response (Courtney et al., 2010; Palomba et al., 2000). Nevertheless, pleasant and neutral sounds maintained a relatively constant deceleration of heart rate.

The present results suggest that the congruence of bimodal stimuli increases the subjective emotional response, especially the perception of arousal. However, contrary to our hypothesis, the absence of congruence between bimodal stimuli did not increase heart rate. Similar results were previously found by Brouwer et al. (2013), who used bimodal stimuli that consisted of pictures and sounds. This suggests that the cardiac response is more influenced by the valence of the stimuli than by the degree of congruence, suggesting greater behavioral relevance of valence compared with congruence.

Several previous studies have shown the influence of valence on heart rate, in which aversive stimuli generated greater heart rate deceleration compared with appetitive and neutral stimuli (Bradley et al., 2001; Bradley & Lang, 2000). In fact, sustained heart rate deceleration is the most common initial response to a threatening stimulus (Campbell et al., 1997). According to the defense cascade model (Lang, Bradley, & Cuthbert, 1997), defensive responses begin with a stage of oriented attention (e.g., distant threat) and progress to a stage of defensive action (e.g., imminent attack). Heart rate deceleration is a characteristic physiological response of the first stage.

In conclusion, the present results suggest that incongruent bimodal stimuli affect subjective emotional experiences (i.e., cognitive processing), decreasing arousal perception and appetitive valence in the case of pleasant stimuli. This could happen because the cognitive processing of incongruent bimodal stimuli is more demanding than the processing of congruent stimuli (Gerdes et al., 2014; Gerdes et al., 2013). This implies a greater investment in cognitive processing while interpreting incongruent bimodal stimuli, thus leaving fewer resources for interpreting emotional subjective experiences. Conversely, congruence between stimuli does not require additional investments in cognitive resources, thus generating a greater subjective emotional impact.

In the present study, the cardiac response was unaffected by the congruence of stimuli, but it was affected by valence. This suggests that with regard to the physiological component of the emotional response (e.g., heart rate), the valence of the stimulus prevails over congruence. Nonetheless, congruence later modulates the perception of the emotional response. This corroborates the defense cascade model, in which physiological emotional reactions are a function of the valence of the stimulus (Lang et al., 1997). For example, when a stimulus that indicates a threat is perceived, heart rate initially decelerates, the skin conductance response increases, and the startle reflex is enhanced. These responses are generated to favor a fight or flight response and subsequent survival of the individual (Bradley & Lang, 2007a).

The relevance of our findings should be evaluated while considering several possible limitations. First, the sample did not include a balanced number of men and women. Similar studies should be conducted with broader samples and a balanced distribution of men and women, which would allow comparisons between sexes. Second, our study only revealed an effect of congruence of bimodal stimuli on heart rate responses and self-reported valence and arousal. Future studies should use other physiological measures to increase our understanding of the impact of the congruence of bimodal stimuli on the emotional response.

Notes

i) Pleasant sounds: 360, 215, 201, 817, 311, 815, 353, 717, 808, 365. Neutral sounds: 262, 708, 700, 705, 720, 374, 382, 373, 252, 322. Unpleasant sounds: 424, 712, 134, 276, 709, 711, 281, 420, 115, 719.

Funding

This study was funded by Universidad de San Buenaventura, Bogotá, Colombia (grant number CHS 012-10).

Competing Interests

The authors have declared that no competing interests exist.

Acknowledgments

The authors have no support to report.

References

  • Alpers, G. W., Adolph, D., & Pauli, P. (2011). Emotional scenes and facial expressions elicit different psychophysiological responses. International Journal of Psychophysiology, 80(3), 173-181. https://doi.org/10.1016/j.ijpsycho.2011.01.010

  • Barry, R. J. (1984). The evoked cardiac response under processing load. Physiological Psychology, 12(1), 35-40. https://doi.org/10.3758/BF03332162

  • Belardinelli, M. O., Sestieri, C., Di Matteo, R., Delogu, F., Del Gratta, C., Ferretti, A., . . . Romani, G. L., (2004). Audio-visual crossmodal interactions in environmental perception: An fMRI investigation. Cognitive Processing, 5(3), 167-174. https://doi.org/10.1007/s10339-004-0024-0

  • Berntson, G., Quigley, K., & Lozano, D. (2007). Cardiovascular psychophysiology. In J. Cacioppo, L. Tassinary, & G. Berntson (Eds.), The handbook of psychophysiology (pp. 182-210). New York, NY, USA: Cambridge University Press.

  • Bos, M. G., Jentgens, P., Beckers, T., & Kindt, M. (2013). Psychophysiological response patterns to affective film stimuli. PLoS One, 8(4), Article e62661. https://doi.org/10.1371/journal.pone.0062661

  • Bradley, M. M. (2009). Natural selective attention: Orienting and emotion. Psychophysiology, 46(1), 1-11. https://doi.org/10.1111/j.1469-8986.2008.00702.x

  • Bradley, M. M., Codispoti, M., Cuthbert, B., & Lang, P. J. (2001). Emotion and motivation I: Defensive and appetitive reactions in picture processing. Emotion, 1(3), 276-298. https://doi.org/10.1037/1528-3542.1.3.276

  • Bradley, M. M., & Lang, P. J. (1994). Measuring emotion: The self-assessment manikin and the semantic differential. Journal of Behavior Therapy and Experimental Psychiatry, 25, 49-59. https://doi.org/10.1016/0005-7916(94)90063-9

  • Bradley, M. M., & Lang, P. J. (2000). Affective reactions to acoustic stimuli. Psychophysiology, 37(2), 204-215. https://doi.org/10.1111/1469-8986.3720204

  • Bradley, M. M., & Lang, P. J. (2007a). Emotion and motivation. In J. Cacioppo, L. Tassinary, & G. Berntson (Eds.), The handbook of psychophysiology. New York, NY, USA: Cambridge University Press.

  • Bradley, M. M., & Lang, P. J. (2007b). The International Affective Digitized Sounds (2nd ed.; IADS-2): Affective ratings of sounds and instruction manual (Technical report B-3). Gainesville, FL, USA: University of Florida.

  • Brouwer, A. M., van Wouwe, N., Mühl, C., van Erp, J., & Toet, A. (2013). Perceiving blocks of emotional pictures and sounds: Effects on physiological variables. Frontiers in Human Neuroscience, 7, Article 295. https://doi.org/10.3389/fnhum.2013.00295

  • Campbell, B. A., Wood, G., & McBride, T. (1997). Origins of orienting and defensive responses: An evolutionary perspective. In P. J. Lang, R. F. Simons, & M. T. Balaban (Eds.), Attention and orienting: Sensory and motivational processes (pp. 41–67). Hillsdale, NJ, USA: Erlbaum.

  • Codispoti, M., Surcinelli, P., & Baldaro, B. (2008). Watching emotional movies: Affective reactions and gender differences. International Journal of Psychophysiology, 69(2), 90-95. https://doi.org/10.1016/j.ijpsycho.2008.03.004

  • Courtney, C. G., Dawson, M. E., Schell, A. M., Iyer, A., & Parsons, T. D. (2010). Better than the real thing: Eliciting fear with moving and static computer-generated stimuli. International Journal of Psychophysiology, 78(2), 107-114. https://doi.org/10.1016/j.ijpsycho.2010.06.028

  • Cox, T. J. (2008). The effect of visual stimuli on the horribleness of awful sounds. Applied Acoustics, 69(8), 691-703. https://doi.org/10.1016/j.apacoust.2007.02.010

  • Detenber, B. H., Simons, R. F., & Bennett, G. G., Jr. (1998). Roll ‘em! The effects of picture motion on emotional responses. Journal of Broadcasting & Electronic Media, 42(1), 113-127. https://doi.org/10.1080/08838159809364437

  • Doehrmann, O., & Naumer, M. J. (2008). Semantics and the multisensory brain: How meaning modulates processes of audio-visual integration. Brain Research, 1242, 136-150. https://doi.org/10.1016/j.brainres.2008.03.071

  • Fernández-Abascal, E. G., Guerra, P., Martinez, F., Dominguez, F. J., Muñoz, M. A., Egea, D. A., . . . Vila, J., (2008). El Sistema Internacional de Sonidos Afectivos (IADS): Adaptación española. Psicothema, 20(1), 104-113.

  • Gantiva, C., & Camacho, K. (2016). Características de la respuesta emocional generada por las palabras: Un estudio experimental desde la emoción y la motivación. Psychologia, 10(2), 55-62. https://doi.org/10.21500/19002386.2550

  • Gantiva, C., Guerra, P., & Vila, J. (2011). Validación colombiana del sistema internacional de imágenes afectivas: Evidencias del origen transcultural de la emoción. Acta Colombiana de Psicologia, 14(2), 103-111.

  • Gerdes, A. B., Wieser, M. J., & Alpers, G. W. (2014). Emotional pictures and sounds: A review of multimodal interactions of emotion cues in multiple domains. Frontiers in Psychology, 5, Article 1351. https://doi.org/10.3389/fpsyg.2014.01351

  • Gerdes, A. B., Wieser, M. J., Bublatzky, F., Kusay, A., Plichta, M. M., & Alpers, G. W. (2013). Emotional sounds modulate early neural processing of emotional pictures. Frontiers in Psychology, 4, Article 741. https://doi.org/10.3389/fpsyg.2013.00741

  • Gomez, P., Zimmermann, P., Guttormsen-Schär, S., & Danuser, B. (2005). Respiratory responses associated with affective processing of film stimuli. Biological Psychology, 68(3), 223-235. https://doi.org/10.1016/j.biopsycho.2004.06.003

  • Hamm, A. O., Cuthbert, B. N., Globisch, J., & Vaitl, D. (1997). Fear and the startle reflex: Blink modulation and autonomic response patterns in animal and mutilation fearful subjects. Psychophysiology, 34(1), 97-107. https://doi.org/10.1111/j.1469-8986.1997.tb02420.x

  • Lang, P. J. (1995). The emotion probe: Studies of motivation and attention. The American Psychologist, 50, 372-385. https://doi.org/10.1037/0003-066X.50.5.372

  • Lang, P. J., Bradley, M. M., & Cuthbert, B. N. (1997). Motivated attention: Affect, activation, and action. In P. J. Lang, R. F. Simons, & M. T. Balaban (Eds.), Attention and orienting: Sensory and motivational processes (pp. 97–135). Hillsdale, NJ, USA: Erlbaum.

  • Lang, P. J., Bradley, M. M., & Cuthbert, B. N. (2008). International Affective Picture System (IAPS): Affective ratings of pictures and instruction manual (Technical report A-8). Gainesville, FL, USA: University of Florida.

  • Lang, P. J., Davis, M., & Öhman, A. (2000). Fear and anxiety: Animal models and human cognitive psychophysiology. Journal of Affective Disorders, 61(3), 137-159. https://doi.org/10.1016/S0165-0327(00)00343-8

  • Obrist, P. A. (1981). Cardiovascular psychophysiology: A perspective. New York, NY, USA: Plenum Press.

  • Palomba, D., Sarlo, M., Angrilli, A., Mini, A., & Stegagno, L. (2000). Cardiac responses associated with affective processing of unpleasant film stimuli. International Journal of Psychophysiology, 36(1), 45-57. https://doi.org/10.1016/S0167-8760(99)00099-9

  • Scherer, L. D., & Larsen, R. J. (2011). Cross-modal evaluative priming: Emotional sounds influence the processing of emotion words. Emotion, 11(1), 203-208. https://doi.org/10.1037/a0022588

  • Simons, R. F., Detenber, B. H., Roedema, T. M., & Reiss, J. E. (1999). Emotion processing in three systems: The medium and the message. Psychophysiology, 36(5), 619-627. https://doi.org/10.1111/1469-8986.3650619

  • Turpin, G. (1986). Effects of stimulus intensity on autonomic responding: The problem of differentiating orienting and defense reflexes. Psychophysiology, 23, 1-14. https://doi.org/10.1111/j.1469-8986.1986.tb00583.x

  • van Atteveldt, N. M., Formisano, E., Goebel, R., & Blomert, L. (2007). Top–down task effects overrule automatic multisensory responses to letter–sound pairs in auditory association cortex. NeuroImage, 36(4), 1345-1360. https://doi.org/10.1016/j.neuroimage.2007.03.065

  • Vila, J., Sánchez, M., Ramírez, I., Fernández, M., Cobos, P., Rodríguez, S., . . . Moltó, J., (2001). El sistema internacional de imágenes afectivas (IAPS): Adaptación española. Segunda parte. Revista de Psicología General y Aplicada, 54(4), 635-657.

About the Authors

Sebastián Calderón: earned his B.A. in Engineering at the Universidad de San Buenaventura, Bogotá, Colombia. His research interests deal with the effects of sound in emotions.

Raúl Rincón: M.A. is an Associate Professor at the Department of Engineering, Universidad de San Buenaventura, Bogotá, Colombia. His research interests deal with the effects of sound in emotions.

Andrés Araujo: earned his B.A. in Psychology at the Universidad de San Buenaventura, Bogotá, Colombia. His current research focuses on psychophysiology of emotions and psychophysical health.

Carlos Gantiva: Ph.D., is an Full Professor at the Department of Psychology and the Neuroscience and Neuropsychology Laboratory, Universidad de San Buenaventura, Bogotá, Colombia. His current research focuses on psychophysiology of emotions, addictions and social neuroscience.