labvanced logoLabVanced
  • Research
    • Publications
    • Researcher Interviews
    • Use Cases
      • Developmental Psychology
      • Linguistics
      • Clinical & Digital Health
      • Educational Psychology
      • Cognitive & Neuro
      • Social & Personality
      • Arts Research
      • Sports & Movement
      • Marketing & Consumer Behavior
      • Economics
      • HCI / UX
      • Commercial / Industry Use
    • Labvanced Blog
  • Technology
    • Feature Overview
    • Desktop App
    • Phone App
    • Precise Timing
    • Experimental Control
    • Eye Tracking
    • Multi User Studies
    • More ...
      • Questionnaires
      • Artificial Intelligence (AI) Integration
      • Mouse Tracking
      • Data Privacy & Security
      • Text Transcription
      • Heart Rate Detection (rPPG)
  • Learn
    • Guide
    • Videos
    • Walkthroughs
    • FAQ
    • Release Notes
    • Documents
    • Classroom
  • Experiments
    • Cognitive Tests
    • Sample Studies
    • Public Experiment Library
  • Pricing
    • Pricing Overview
    • License Configurator
    • Single License
    • Research Group
    • Departments & Consortia
  • About
    • About Us
    • Contact
    • Downloads
    • Careers
    • Impressum
    • Disclaimer
    • Privacy & Security
    • Terms & Conditions
  • Appgo to app icon
  • Logingo to app icon
Research
Publications
Tasks
Researcher Interviews
Use Cases
Labvanced Blog
  • 中國人
  • Deutsch
  • Français
  • Español
  • English
  • 日本語
Publications
Tasks
Researcher Interviews
Use Cases
Labvanced Blog
  • 中國人
  • Deutsch
  • Français
  • Español
  • English
  • 日本語
  • Publications
  • Tasks
  • Researcher Interviews
    • Comparing Online Webcam- and Laboratory-Based Eye-Tracking for the Assessment of Infant Audio-Visual Synchrony Perception
    • Finding Goldilocks Influencers- How Follower Count Drives Social Media Engagement
    • The semantic interference in 9- to 36-month old infants: An at-home eye-tracking study on infants' lexical abilities
    • Song Is More Memorable Than Speech Prosody - Discrete Pitches Aid Auditory Working Memory
    • Orthographic Relatedness and Transposed-word Effects in the Grammatical Decision Task
    • Emotion Modulation through Music after Sadness Induction
    • Children Learning Non-Adjacent Dependencies
    • Personality Hear in Noise
    • Are All Eyes the Same?
    • Verbal Priming in Infants
  • Use Cases
    • Research Areas

      • Developmental Psychology
      • Linguistics
      • Clinical & Digital Health
      • Educational Psychology
      • Cognitive & Neuro
      • Social & Personality Psychology
      • Arts, Music & Digital Media
      • Sports & Movement Psychology
      • Marketing & Consumer Behavior
      • Economics
      • HCI / UX
      • Commercial / Industry Use
    • Researchers

      • Students
      • Researchers
      • Groups
  • Blog
    • Generative AI in Educational Psychology Research
    • Best Practices for Webcam Eye Tracking Research with Infants and Toddlers
    • Types of Stimuli in Delay Discounting Tasks
    • Delay Discounting Task
    • Mindfulness-based Self-Efficacy Scale - Revised (MSES-R)
    • Team Dynamics and Research
    • Stroop Task | History, Task Description, Data and Psychology
    • BKB Sentence Test | Procedure & Research
    • Dyadic Effect in Psychology | Overview & Research
    • Navon Task: Task Setup, Research & More
    • Peer-reviewed Webcam Eye Tracking
    • Corsi Block-Tapping Test: From cubes to online design
    • Dot-Probe Task | Full Guide
    • Types of Memory: Concepts to Research
    • Ultimatum Game
    • The Visual Search Task
    • Attention Tasks in Psychology Research
    • Generalized Anxiety Disorder Scale-7 (GAD-7)
    • Decision Making Tasks in Psychology
    • The Obsessive Compulsive Inventory – Revised (OCI-R)
    • Assessing Executive Function Skills | Tasks & Batteries
    • The Flourishing Scale (FS) Questionnaire
    • Labvanced and the Spirit of Open Science
    • The Psychology of the Incubation Effect
    • The Bouba-Kiki Effect and Task
    • Lexical Decision Task: Accessing the Mental Lexicon
    • Image Description Task and Game with a Chatbox
    • Smooth Data Collection Process | 6 Tips for Research
    • Music Research with Labvanced
    • 7 Classic Cognitive Tasks & Examples
    • Mental Rotation Test | A Spatial Processing Task
    • XY Coordinates in Labvanced
    • 5 Famous Social Psychology Experiments
    • The Power of Remote & Infant-friendly ET
    • The Wisconsin Card Sorting Test
    • 13 Head Tracking Use Cases for Research
    • 5 Tips for Improving Your Perception Skills
    • Introduction to the Preferential Looking Paradigm
    • Researching Cognition in Migraine & Headache Patients
    • The Landing Page - More Info Than You Think!
    • Headphone Checks - Then and Now
    • Sample Studies - Helpful Templates and Demos!
    • What Is Neuroplasticity?
    • 15 Famous Developmental Theories
    • Visual Attention and Eye Tracking
    • What Is Eye Tracking Technology?
    • Eye Tracking in Applied Linguistics Research
    • 10 Popular Linguistic Experiments
    • The Placebo Effect
    • 6 Key Concepts of Experimental Design
    • Conditioned Play Audiometry
    • Ebbinghaus Illusion
Bouba-kiki task preview for online and inlab use.

The Bouba–Kiki Task

The Bouba–Kiki Task measures sound–shape correspondence by examining how people associate unfamiliar words with visual forms. It is widely used in cognitive psychology and linguistics research to study perception, language processing, and multisensory integration.

Table of Contents

  • Task Format
  • Data Collected
  • Technology
  • Customization
  • Recommended Use

Task Format | Bouba-Kiki Task Online & In-Lab

In the Bouba–Kiki Task, participants are presented with a novel word along with two abstract shapes and are asked to decide which shape best corresponds to the word. The task is designed to capture intuitive sound–shape associations rather than learned semantic meaning. The session begins with a short practice block to familiarize participants with the response format before the main experimental trials.


Test it out!

Two versions of the bouba-kiki task are available, each optimized for the type of device and input method being used:

Desktop Version

In the desktop version, a word is displayed at the top of the screen with two abstract shapes shown side by side at the bottom. Participants respond using the keyboard by pressing A to select the left shape or L to select the right shape. Participants are instructed to keep their fingers on the respective keys to respond quickly. A practice block precedes the main task.


Mobile Version

In the mobile version, the word and shapes are displayed in the same layout. Participants respond by tapping directly on the shape they believe best matches the presented word. This version is optimized for touchscreen interaction and includes practice trials before the main task begins.

Bouba-Kiki Task Data Collected & Metrics

The Bouba–Kiki Task captures a range of behavioral measurements that reveal how individuals associate sounds with visual shapes, reflecting crossmodal correspondence and sound symbolism effects. The variables recorded enable researchers to assess choice patterns, response times, accuracy where applicable, and trial-specific characteristics that may influence perceptual matching. These measures help quantify consistency of sound–shape associations and the strength of crossmodal integration. All variables can be viewed and customized within the task’s Variables Tab.

Below are examples of variables collected in the Labvanced version of the Bouba-Kiki Task:

Variable NameDescription
Task_NameRecords the name of the specific task of the trial
AnswerRecords the response of the participant (If computer version: key pressed by the participant in that trial (A/L). If mobile version: Image clicked on (Target shape/ Non-target shape)
Choice_shapeThe shape chosen by the participant in a trial
non-target image nameThe filename of the non-target image presented in the trial
reaction timeTime taken by the participant to respond (eg. press A/L) in a trial
target image nameThe filename of the target image presented in the trial.
text_presentedThe word presented in a trial
bouba-bouba_avg_RTAverage RT for bouba-word + bouba-shape matches
kiki-kiki_avg_RTAverage RT for kiki-word + kiki-shape matches

Data preview of behavioral measures and responses collected using the Bouba-Kiki Task template for online research.

Data table showing individual trial level outputs from the online Bouba-Kiki Task in Labvanced.


Bouba-Kiki Task

In this experiment, two shapes are presented and a word. The participant must select one of the two shapes as being the best match to the presented word.



Technology Driving the Bouba-Kiki Task for Online & In-Lab Research

The Bouba–Kiki Task requires flexible stimulus handling, multimodal presentation, and reliable response recording. Labvanced provides the technical infrastructure needed to support these requirements across diverse study designs.

  • Multimodal Stimulus Presentation: Visual shapes, written text, and audio stimuli can be combined within the same trial. This allows researchers to present words visually, auditorily, or both, depending on the experimental design.

  • Flexible Input Collection Across Devices: Responses can be collected through keyboard input on desktop devices or direct touch selection on mobile devices. The same task logic can be reused while adapting the interface to the available input method.

  • Desktop App Support for Controlled Studies: For lab based experiments requiring precise timing or integration with external hardware, the task can be deployed using the Labvanced desktop app. This supports compatibility with EEG or other LSL based systems.

  • Remote and Longitudinal Deployment: The task can be delivered remotely and repeated across multiple sessions, making it suitable for longitudinal studies examining stability or change in perceptual associations over time.

  • Optional Webcam Eye Tracking Integration: The task can be combined with webcam based eye tracking to analyze visual attention, fixation patterns, and decision processes during shape selection.

A person looking at a webcam eye tracking task created with Labvanced during a bouba-kiki task

Webcam Eye Tracking

Capture gaze patterns and visual attention with built-in, code-free and peer-reviewed webcam eye-tracking.

An icon of a clock symbolizing accurate reaction times from this bouba-kiki task template

Timing Precision

Capture reaction times, task performance, and more with millisecond accuracy for time-sensitive tasks.

An image of the desktop app of Labvanced used to modify the bouba-kiki task template

Desktop App

Run in-lab studies using the Desktop App, compatible with EEG and other LSL-connected lab hardware.

Customization of the Bouba-Kiki Task

There are many ways to adapt the Bouba-kiki Task template to meet specific research questions. Below are a few themes researchers commonly ask when it comes to modifying this task.


Shape and Stimulus Design

Abstract shapes are implemented as visual objects that can be replaced or edited directly in the editor. Researchers can modify shape complexity, border, size, or background color through the Object Properties panel.


Word and Audio Configuration

Novel words can be presented as text objects or audio files. Audio stimuli can be uploaded and linked to trials, allowing precise control over pronunciation, timing, and repetition. Audios could be added using the audio option from the media objects.


Condition Logic and Trial Variations

Different phonetic contrasts or shape pairings can be defined through trial conditions. These condition values are read during runtime to determine which stimuli are presented on each trial. You could always make the best out of the Trials & Conditions panel.


Response Handling and Trial Flow

Participant selections are captured through clicks, taps, or key presses and evaluated using Event logic. Researchers can modify how responses during the Bouba-Kiki task (regardless of whether it's administered online or in-lab) are logged, how missed responses are handled, and how trials progress, all through the event system. Simply modify the triggers or actions specific to your needs.


If you need help customizing this task, please feel welcome to write to us and ask:


Contact Us

Recommended Use and Applications of the Bouba-Kiki Task

The Bouba–Kiki Task is widely used across research domains, in online and in-lab settings, that investigate perception, language, and multisensory processing.

  • Perception and Multisensory Integration Research: Commonly used to study how auditory and visual features are integrated and how sensory correspondences emerge.

  • Language and Phonetic Processing Studies: Applied to examine how phonetic features such as consonant sharpness or vowel quality influence perceptual associations.

  • Developmental Research: Used with children and adults to investigate when sound–shape correspondences emerge and how they change across development.

  • Cross Cultural and Linguistic Studies: Frequently employed to compare perceptual mappings across languages and cultural contexts.

  • Neurocognitive and Brain Function Research: Included in studies exploring neural mechanisms underlying crossmodal perception and symbolic mapping.


Sign Up


References

  • Passi, A., & Arun, S. P. (2024). The Bouba–Kiki effect is predicted by sound properties but not speech properties. Attention, Perception, & Psychophysics, 86(3), 976-990.

  • Piller, S., Senna, I., & Ernst, M. O. (2023). Visual experience shapes the Bouba-Kiki effect and the size-weight illusion upon sight restoration from congenital blindness. Scientific Reports, 13(1), 11435.