Auditory-Visual Shift Study

OpenNeuro/NEMAR Dataset: ds002893 Files: 370 Dataset size: 7.7 GB
Channels: 33 EEG,2 EOG,1 Misc
Participants: 49
Event files: 52 View events summary
HED annotation: Yes View HED tags summary

Download Download this dataset as a zip file (2.1 GB).
Compute Process this dataset using the Neuroscience Gateway (NSG).
Discuss Read & contribute to a discussion of this dataset. (1 comment)
OpenNeuro Browse the OpenNeuro entry for this dataset.
Citations Browse 1 citing paper.
README

Audio-Visual Attention Shift Experiment

Project name: [Sensory processing in aging]

Years the project ran: 2007-2008

Brief overview of experiment task: The purpose of this Auditory-Visual Attention Shift study was to explore the effects of aging on selective attending and responding to auditory and visual stimulus differences using an interleaved dual-oddball audio-visual task design. EEG and EOG channels were acquired.

Data collection. Scalp EEG data were collected from 33 scalp electrode channels, each referred to a right mastoid electrode, within an analogue passband of 0.1 to 60 Hz.

Contact person: Scott Makeig smakeig@ucsd.edu, ORCID: 0000-0002-9048-8438.

Access information: Contributed to OpenNeuro.org and NEMAR.org in BIDS format following annotation using HED 8.0.0 in April, 2022.

Independent variables: Stimulus stream (visual, auditory, cue); stimulus stream identity (target, standard); task condition (FA, FV, SH)

Dependent variables: Participant response (correct/incorrect). Button press response attributes (task time window and post-target latency).

Participant pool: The dataset includes data collected from 19 younger adult subjects (8 male, 11 female, ages 20?40 years) and 30 older adult subjects (11 male, 19 female, ages 49-73 years). The subjects were cognitively intact and had normal or adjusted to normal hearing and vision.

Initial setup: EEG data were collected from 33 EEG channels using the 10-20 placement and referenced to the right mastoid. The left mastoid and two EOG channels were also included in the collection. The data was acquired at a sampling rate of 250 Hz with an analog pass band of 0.01 to 60 Hz (SA Instrumentation, San Diego). Input impedances were brought under 5 kilo-ohms by careful scalp preparation.

Task conditions:

  • Focus Visual (FV): participants pressed the response button only in response to target visual stimuli.
  • Focus Auditory (FA): participants pressed the same button only in response to target auditory stimuli.
  • Shift Focus (SF): participants shifted between performing the FV and FA tasks as cued by the preceding (Look/Hear) cue stimulus.

Task organization: The stimuli were presented in blocks of 264 for a duration of 2.64. In each block there were 12 "Hear" and 12 "Look" cues. A total of 20 blocks were presented for each session. Each experiment began with two non-shift blocks (one each of auditory focus FA and visual focus FV counter-balanced across sessions). These were followed by 12 SF shift blocks. Finally an auditory focus FA group (3 blocks) and a visual focus FV group (3 blocks) were presented. The order of these groups was counter-balanced across experiments. Brief rest periods occur between task blocks. The task condition in the next block was given verbally to the participant during the pre-block rest period.

Task details: Participants respond by finger button press selectively to auditory (brief tones) and visual (colored squares) stimuli constituting distinct, interleaved auditory and visual oddball stimulus streams whose stimuli are presented in randomly interleaved order with stimulus-onset asynchronies (SOAs) varying randomly between 200 and 800 ms.

  • Visual stimuli: were (infrequent, 10%) dark blue target or (frequent, 90%) light blue standard 8.4-cm2 squares presented for 100 ms.
  • Auditory stimuli: were (infrequent, 10%) 550-Hz target or (frequent, 90%) 500-Hz tones with 100 msec duration and 63 dB SPL intensity.
  • Task cue stimuli: interspersed in the stimulus sequence at mean 5-sec intervals and consisting of the simultaneous spoken and printed display of one of the words Look or Hear each presented for 200 msec.

Additional data acquired: Participants had no history of major neurological, psychiatric, or medical disorders. All had normal or adjusted to normal vision and hearing (none wore hearing aids). Verbal and performance IQ were assessed using the WASI-III (Wechsler, 1997). There were no significant differences between the groups in IQ measures or years of education. Participants in the Older group received a battery of neuropsychological tests to assure normal cognitive functioning, including the Mini Mental State Exam (MMSE) (Folstein et al., 1975), Dementia Rating Scale (DRS) (Mattis, 1988), Wechsler Memory Scale.

Experiment location: Department of Psychiatry laboratory of Jeanne Townsend, University of California San Diego, La Jolla CA (USA).

Note 1: ERP measure results for the FA and FV conditions only were presented in Ceponiene, R., Westerfield, M., Torki, M. and Townsend, J., 2008. Modality-specificity of sensory aging in vision and audition: evidence from event-related potentials.?Brain research,?1215, pp.53-68. Some unpublished results by Christian Kothe and Scott Makeig on the SH condition may be available from the authors christiankothe@gmail.com smakeig@ucsd.edu.

Note 2: The code subdirectory has several auxilliary files that were produced during the curation process. The curation was done using a series of Jupyter notebooks that are available as run in the code/curation_notebooks subdirectory.

During the running of these curation notebooks information about the status was logged using the HEDLogger. The output of the logging process is in code/curation_logs.

Updated versions of the curation notebooks can be found at: https://github.com/hed-standard/hed-examples/tree/main/hedcode/jupyter_notebooks/dataset_specific_processing/attention_shift.


HED Event descriptors
word cloud:

BIDS Version: 1.7.0 HED Version: 8.0.0 Version: 2.0.0

On Brain life: True Published date: 2020-06-29 23:47:19

Tasks: AuditoryVisualShift

Available modalities: EEG

Format(s): .fdt, .set

Sessions: 1 Scans/session: 2 Ages (yrs): 20 - 78 License: CC0

Dataset DOI: doi:10.18112/openneuro.ds002893.v2.0.0

Uploaded by Dung Truong on 2020-06-15 06:04:46

Last Updated 2022-06-18 20:41:02

Authors
Marissa Westerfield (data, curation), Scott Makeig (data, curation), Dung Truong (curation), Kay Robbins (curation), Arno Delorme (curation)

Acknowledgements

How to Acknowledge
Cite this paper: https://pubmed.ncbi.nlm.nih.gov/18482717/ and consider including the following message: 'This data was obtained from the OpenNeuro database. Its accession number is ds002893'

Funding
  • National Institutes of Health National Institute on Aging (R01-AG18030).
  • References and Links
  • Ceponiene R., Westerfield M., Torki M., and Townsend J.(2008). Modality-specificity of sensory aging in vision and audition: Evidence from event-related potentials, Brain Research, vol. 1215, 53-68. DOI: 10.1016/j.brainres.2008.02.010.
  • Ethics Approvals