• How it works
  • Recruit participants

Visual search task

At the end of this guide you will be able to….

  • Run a visual search task in Testable starting from our ready-made template
  • Customise key parameters to adapt this task to your needs
  • Collect data and interpret the results
If you never worked with Testable before, you can check out our 10-minute introduction video here

What is the visual search task and how can you use it in your own research?

We do not see our world as a large collection of individual lines, colours and shapes. Instead we perceive a coherent visual scene with complex objects that we understand. But how do our brains accomplish this? Do we first perceive all the basic features and then bind them as a whole? Or do we rather perceive the whole before we can actively search for more detail if needed? In 1980, Treisman and Gelade developed the visual search task (VST) as a simple way to test both theories.

In this task, participants need to find a target in a scene of distractors. Usually, the target has a particular shape and colour (e.g red triangle 🔺 ) and needs to be spotted among distractors. Distractors could have a different color (blue triangles) a different shape (red squares) or both (blue squares).

Screenshot of the visual search task in Testable

The hallmark observation from the VST is the ‘pop-out’ effect. Targets ‘pop-out’ from a cluttered visual scene as long as they can be identified unambiguously by a single feature. You can instantly spot a target, regardless of the number of distractors, if it is the only red object. This is the reason why signal colours are so effective. Think of bright yellow safety vest, which can capture our attention even in the most crowded visual scene.

The process by which we can recognise a single stand-out feature in our visual scene is called feature search. The insight about the superiority of feature search has been invaluable to marketeers. They can use experiments similar to the VST to develop product packaging and branding that ‘pops out’ from its competitors.

When the target shares features with the distractors, the number of distractors matters. If the target is red triangle, shared-feature distractors might be blue triangles or red squares. In this case it is almost as if we need to scan each object individually for the right feature combination. This effortful mental process is called conjunction search.

This combination of findings suggests that we have automatic and rapid access to “lower level” visual features. This allows us to process some aspects of our entire visual scene at once. More complex, “higher level” objects require the binding together of multiple features. This solicits our selective attention and can only be processed one at a time. Although we are aware of our visual world as a meaningful ‘whole’ it seems that our brains need to puzzle it together from its constituent parts.

The visual search task allows us to test how our attention is detects features in a visual scene.

How does the testable visual search task work.

At the beginning of the task participants are shown a target object that they need to look out for. The target is surrounded by other, similar shapes that act as distractors. It is much like a simplified game of “Where’s Wally”. In our example the target is a red triangle (🔺) surrounded by scattered distractors. These could have a different color (blue triangles), a different shape (red squares) or both (blue squares).

It is the participant’s task to use the keyboard to indicate if the target is present (‘y’) or absent (‘n’) in the scene. Typically, three aspects of the scene are varied from trial to trial:

  • Target present or absent: The triangle could either be or not among the objects in the scene or not, each scenario 50% likely in this version of the experiment.
  • Size of the scene: The total number of items in the visual scene (including the target, if present) can either be 4, 8, 16 or 32.
  • Feature search: Distractors all share one property different from the target → e.g. all are blue, or all a squares.
  • Conjunction search: Distractors can have any combination of properties (except for the target) → red squares and blue triangles could exist together in the same scene.

This results in 16 unique trial conditions, which each have 5 visual scene examples associated with them. This amounts to a total of 80 trials in our demo experiment and should be sufficient to observe a meaningful effect.

In feature search trials it is enough to spot one of the target’s characteristics (either red colour, or triangle shape) to distinguish it from the distractors. Conversely, in conjunction trials, one must look for the combination of red color and triangle shape, as there are other red and triangular distractors around.

Fore each trial we measure if the participant correctly indicated whether the target was among the distractors (accuracy), and how quickly they did so (response time).

visual search experiment

Run this experiment in Testable from our ready-made template

We have created a template for the VST in Testable for you that you can access from our Library. It is set up and ready to go and you can start collecting data straight away by sending the experiment link to your participant. Experiments in Testable will run in every browser, which makes it very easy to collect data both in the lab as well as online.

How to customise key parameters and play around with different options

Experiments in Testable are fully customisable and you will not need to write a single line of code to edit them. The heart of each experiment is what we call the trial file . The trial file contains all information that Testable needs to run the experiment in a simple spreadsheet, that you can edit with any spreadsheet editor you like, such as Google Sheets, Excel or Testable’s in-built preview editor.

To change any part of your experiment, you only need to change the values in the trial file.

Once you have made your changes, you can save and upload the modified trial file to your experiment’s trial file section.

Selection of columns from the Testable experiment trial file of the Visual Search Task

Collect data by sending the experiment link to your participants

After importing this template to your library, you can collect data for your experiment by sharing the unique experiment link (i.e. tstbl.co/xxx-xxx) with your participants. Once participants complete the experiment, their results will appear in the ‘Results’ section of your experiment.

Working with results from the visual search Task

This version of the VST automatically measures a participants’ response time ( RT column) and accuracy ( correct column) for indicating whether the target was present or not in a trial.

In the trial file we have defined three columns called condition1 , condition2 and condition3 that each code one dimension that the stimuli were varied on in this task. condition1 tells us if it was present or not, condition2 tells us if feature search (target can be identified by a single feature) or conjunction search (target can only be identified using a combination of features) and condition3 codes the number of objects in a scene (4-32). We added these columns manually added to the trial file and they have no bearing on the logic of the experiment. But they do come handy for our analysis, as we can now easily group the response times and accuracies for each level of experimental conditions and calculate their mean values.

There are two key effects that you should observe when running the VST:

  • Conjunction search trials should have significantly longer response times compared to feature search trials.
  • Response times are longer for trials with more objects, but only for conjunction search and not for feature search. There, the number of objects shown does not affect response times.

In technical language you would be looking for a main effect of search type, which means that overall conjunction search takes longer than feature search. In addition, you are looking for an interaction between search type and array size. That means that the response times you can expect for the different array sizes depends on the search type required for that trial. This interaction effect allows us to conclude that single features can ‘pop out’ in crowded a visual scene. But once we need to combine features to identify an object, we need to process objects one by one using conscious attention. This takes longer and is more effortful.

visual search experiment

Once you have collected data from multiple participants, you can also use Testable’s ‘wide format’ feature, that allows you to automatically collate all individual result files into a single file. In wide format results every participant’s data shows up as one row in the data file. This makes it easily compatible with statistical analysis packages like R or SPSS where you can assess the statistical significance of any differences you may find.

Reference list:

Treisman, A. M., & Gelade, G. (1980). A feature-integration theory of attention. Cognitive Psychology, 12(1), 97–136. doi:10.1016/0010-0285(80)90005-5

Jansson, C., Marlow, N., & Bristow, M. (2004). The influence of colour on visual search times in cluttered environments. Journal of Marketing Communications, 10(3), 183-193.

visual search experiment

  • Access Account
  • Set Up Account
  • Assignments
  • Privacy Policy
  • Purchase Info
  • Requirements
  • Password/ID Help
  • Student Manual
  • Tech Support
  • Create Group
  • Password Help
  • Log out of CogLab

Visual Search

Estimated time to complete lab: 20 minutes

This experiment explores aspects of attention in a visual-search task. It is a classic experiment that makes strikingly clear the time needed to bring attention to bear on different regions of visual space. The basic idea is to ask people to search a visual image for a particular item and to respond as quickly as possible once they find the item, or to respond as quickly as possible when they are certain the item is not in the image. This type of experiment was used to develop a popular theory of attention (Treisman & Gelade, 1980).

Searches are divided into two types: those that require selective use of attention and those that do not. In the latter, the target item seems to pop-out of the display and the participant can respond quickly. Notably, this pop-out effect allows the participant to respond quickly even when the number of other (distractor) items is increased. In the other type of display, it seems that the participant is forced to study each item individually until the target item is found. In these cases, the target item does not pop-out and search time increases with the number of distractor items. Controlling whether attention is needed or not is accomplished by the type of target and distractor items.

In the experiment below, the target is always a green circle. For the feature condition, the distractors are always blue squares. As you will see, the green circle seems to pop out of the image which allows you to quickly identify the location of the target. In the conjunctive condition, the distractor items are made more complex. Some of the distractors are green squares while others are blue circles. Because some of the distractors are green, the green target circle no longer pops out and you must search through all the items to find the one that is both green and a circle. This type of search is called a conjunctive search because the target is a conjunction, or mix, of features in the distractors.

Instructions

If you have logged in, you'll see a black rectangle below. Make sure that you can see the full area before you begin the lab.

In this lab, a fixation point is shown and then disappears. After a random amount of time, circles and squares of various colors will appear on the screen. Your task is to determine if there is a green circle among the shapes. Your goal is to be both as fast and as accurate as you can be. You should ignore all other stimuli and indicate whether there is a green circle.

Trials on which you make an error will be repeated. There are a minimum of 80 trials.

At the end of the experiment, you will be asked if you want to save your data to a set of global data. After you answer the question, a new Web page window will appear that includes a debriefing, your data, your group's data, and the global data.

Tablet Specific Details

If you are using a tablet, tap the Start Next Trial button to start a trial. If you see a green circle, tap the Present button with your right index finger. If there aren't any green circles, tap the Absent button with your left index finger.

Computer Specific Details

If you are using a computer, press the n -key to start a trial, then move your right index finger over to the m -key. If you see a green circle, press the n -key with your right index finger. If there aren't any green circles, press the z -key with your left index finger.

Remember, on every trial, your task is the same: Determine if a green circle is present and respond as quickly (and as accurately) as you can.

To use CogLab, you must first enable JavaScript on your browser.

Trials to go: 5

Would you like to add your data to the global data set?

This is optional. Whatever you choose, CogLab will save your individual data and record that you have completed the lab.

Your summary data

Your trial-by-trial data, global data.

visual search experiment

IMAGES

  1. The visual search task. Experiment 1 trials included up to 3 targets

    visual search experiment

  2. Visual Search Task

    visual search experiment

  3. Visual Search

    visual search experiment

  4. PPT

    visual search experiment

  5. Visual Search Task

    visual search experiment

  6. Visual Search Task

    visual search experiment

VIDEO

  1. Sofala Explains It All

  2. A.I. Experiment: Mood Board Search

  3. How to Use Visual Search in Pinterest App

  4. Content Based Image Search: InstructBLIP + Sentence Transformers + FAISS

  5. How to use Google Lens to do a visual search. #shorts

  6. TWL user experience.Visual search AI Keywords