Explore the list of questions in the menu to the left. Here you can troubleshoot solutions to commonly asked questions.
If you have general questions about using Gorilla try these How To Guides:
This is often because two nodes (questionnaire widgets) have the same key. If this is the case, an error message, like the one below, will be displayed.
Error Message 'Two of your nodes have the same key: response-2'
Each widget in the questionnaire has to have a unique key. You can leave them as they are when you add the widget (response-1, response-2, response-n), or you can rename them to be more meaningful.
There are generally three reasons why you might not be able to edit a task:
Details of what to do in each case:
Beta Zones are new zones that have been released for public use, but that we intend to make improvements to before they are considered ‘finished’. We cannot guarantee that they will work without fault, and suggest that these be used in addition to our other zones, rather than instead of.
Closed-Beta zones are available ‘on request’ only. These zones are an early implementation of the feature, and have not been thoroughly tested on across all browsers. Access is given on the understanding that you will provide us with feedback on any bugs as soon as possible. Please also let us know if there’s something you’d like to be added to the zone.
Closed-Beta Zones may have limited support information available, so along with use of the Closed Beta Zone, you will be given access to a project containing a quick tutorial or example of the zone so you can see how this feature is set up, and, if desired, clone the example for your own use.
To get access to a Closed-Beta zone, contact email@example.com
There are three common reasons why you might not be able to upload a spreadsheet:
There are two different causes for spreadsheet columns appearing to be missing. Both of which typically occur on your Task's spreadsheet page after a new spreadsheet has been uploaded.
Note: This issue usually only arises when using multiple spreadsheets within your task.
To debug this issue first identify what 'type' of spreadsheet column has gone missing:
Here's what to do in each case:
1. Spreadsheet Metadata Column is missing:
Please refer to this dedicated troubleshooting page: Metadata column is missing.
2. Spreadsheet Source Column is missing:
You uploaded a new spreadsheet to your task and one or more of the columns associated with controlling your spreadsheet does not appear.
Explanation: Gorilla helpfully removes columns which it detects as not in use: This includes empty columns not referenced within the task structure (i.e. columns not linked to a specific zone via setting the source to spreadsheet).
This helps keep both the task spreadsheet and the subsequent metrics produced from the task as clean as possible.
1) The spreadsheet being uploaded did not contain all the columns. 2) The spreadsheet column is not linked to any zone in your Task Structure. 3) The spreadsheet column is used within your task but it's source is set to Spreadsheet-manipulation.
Please refer to this dedicated troubleshooting page: Metadata column is missing for help with this issue.
There are two different causes for metadata columns appearing to be missing. Both of which typically occur on your Task's spreadsheet page after a new spreadsheet has been uploaded.
1) The spreadsheet being uploaded did not contain all the columns. 2) The uploaded metadata column is empty and is not present in all of the spreadsheets uploaded to the task.
Explanation: Gorilla helpfully removes all columns which it detects as not in use. This includes empty columns and those not referenced within the task structure. Where an 'empty column' means there are no data entries in any row beneath the column heading.
This helps keep both the task spreadsheet and the subsequent metrics produced from the task as clean as possible.
Metadata columns with data inside will be visible so long as there is at least one spreadsheet - within all the spreadsheets uploaded to the task - which contains at least one row of data.
Usually this is not problematic but there are instances where empty metadata columns are required. You can read more about this and using Metadata Columns Spreadsheet Walkthrough here.
As spreadsheet columns are visible to all spreadsheets - regardless of whether they are in use by an individual particular spreadsheet or not. These metadata columns will now always appear in each of your spreadsheet without disappearing if they are empty. This is particularly useful when working with spreadsheet-manipulations and/or scripts.
'I've uploaded a new spreadsheet to my Task but extra columns are being displayed on my Spreadsheet Tab page and I can't get rid of them!'
Sometimes you may notice additional spreadsheet columns in your spreadsheet and/or metrics. Generally these are not problematic to the running of the task and can be either safely ignored or removed.
There are three common causes for extra spreadsheet columns appearing:
1) These 'extra' columns are being used by another spreadsheet within your task. 2) The 'extra' column is linked somewhere within your task structure. 3) For some zones, if a zone is set to spreadsheet and changed to another zone type which no longer requires the setting the column may remain in your spreadsheet, even if no longer in use.
Regardless of whether a particular column is in use by an individual spreadsheet. Non empty columns which belong to any spreadsheet uploaded to a task, will appear in all spreadsheets for that task.
For example: If spreadsheet-1, contains columns 'A' and 'B' and spreadsheet-2 contains columns 'C' and 'D'. So long as there is at least one row of data under each column heading: All columns, 'A', 'B', 'C' and 'D' will appear in both spreadsheet-1 and spreadsheet-2.
If you wish to remove a column:
The column should now be removed.
If a spreadsheet column is referenced in your task structure but is not included in your uploaded spreadsheet the column name will appear as a, green headed, empty column on your Task Spreadsheet page.
If you wish to remove this column:
Alternatively - if you intended to use this column but it has been misnamed: Rename the Zone Content so that it exactly matches the spreadsheet column heading you wish to use instead
You can safely ignore this spreadsheet column and remove any references to this column from your metrics. It will not affect the running of your task.
If you wish to remove this column:
The column should now be removed.
I've set up my display, but when I preview my task, this display is being skipped over and isn't shown.
There are two common causes for this issue:
Here's what to do in each case:
1. You have set up the Display in the Task Structure, but you haven't entered the display name correctly into the spreadsheet.
1) Download the spreadsheet you are running your Task from. 2) Highlight and Copy the name of the Display you are having problems with. 3) Paste the name into the 'Display' Column of your spreadsheet & double check there are no trailing spaces at either end. 4) Save your amended spreadsheet and upload it to Gorilla.
2. Your Display name contains spaces, special characters or punctuation.
Display names are Casesensitive. Thus, the most common reason for a Display not appearing is 'stray spaces'.
Its very easy to accidentally put an extra space into your Display names (or when entering these names into the spreadsheet). Most commonly this occurs at the beginning or end of the name when entering the display into the spreadsheet.
1) Remove all spaces from your Task's Display names. Instead use camelCase or PascalCase as the format of your display names. 2) Remove all special characters & Punctuation from your Task's Display names. 3) Download your existing spreadsheet and replace your old display names with your new ones. 4) Save your amended spreadsheet and upload it to Gorilla.
3. My Display still isn't showing?
I've removed all spaces & special characters from my display names and reuploaded a new spreadsheet but it's still not showing this Display.
Get in contact with us via our Support Contact Form. Provide us with a link to your task and we will happily take a look at your issue directly.
See the video below to learn 3 ways to get your task moving forward when it appears to be 'stuck' on a screen.
Length (mins): 5:47
There are three common reasons why your task preview may be different from what you expected:
I get a blank screen on previewing my Task/ part-way through previewing my Task
I get a blank screen and the error message 'No source for the media could be found! Try refreshing the page. If the problem persists, contact the researcher for the experiment'.
When I preview my Task it doesn't look how I am expecting it to (but my displays are working as I expect them to).
If your video stops playing part way through, this is typically because your video file is corrupted in some way.
On Chrome, an error message will appear when this occurs, and video play error information will appear in your metrics. This means that if you launch an experiment containing a corrupted video, only participants on Chrome will show any error in the data.
In this case, you need to use a different video file.
Is your audio set to play automatically instead of on a click or key press?
If so, this problem is likely caused by Autoplay issues. See our guide here.
If your audio is set to manual play, please contact us via our support form and include information about the device and browser this problem occured on.
There are two common reasons you might not be able to preview an experiment:
As soon as a participant starts an experiment (clicks on the recruitment link or is sent a recruitment email) they are marked as 'Live'. They will remain 'Live' until they reach a finish node, at which point they are marked as 'Completed'. It is also at this point that they are automatically 'included' and their metrics become available in your data. For Pay-per-Participant users, this is also the point where a participant token will be consumed.
When a participant is Live, and has only recently joined the experiment, its reasonable to assume that they are still taking part and just need a little more time to finish. If a participant has been Live for a very long time the most likely scenario is that, unfortunately, they have chosen to leave the experiment. This could have happened very early on: perhaps they chose not to consent. Or, very late in the experiment: they may have gotten interrupted by something else and had to leave the experiment.
To help you work out where a participant is in an experiment, its really helpful to include Checkpoint Nodes.
These are available in the experiment tree and we recommend putting then in between key components of your task. The most recent Checkpoint node that a participant passed through is recorded in the Participant tab of your experiment. This way you can work out how far through your experiment a Live participant is and how likely it is that they are still working on your experiment, or have simply left. For Pay-per-Participant users, this is allows you to make an informed choice on whether or not to include a live participants data, consuming a token, or reject them and have the token returned. You can then choose to include only those participants who have progressed far enough through your experiment that their data is worthwhile including.
If your participants are being turned away with the message that the Experiment is full, there are two possibilities:
If your Experiment is ‘Full’ this means that your Recruitment Target has been reached. In other words, that number of participants have entered your Experiment. This means that some of your participants have not completed the Experiment and are still ‘Live’.
If your participants have been Live for a large amount of time, you may need to manually reject them in order to allow more participants to enter the experiment and complete your recruitment target.
For information on how to do this, see the How To: Participants guide.
For a full guide to Media Error Messages check out our dedicated Help Page: Troubleshooting: Media Error Messages
There are three places where you may encounter a 'Media Error Message'.
1) On your Task Stimuli page when uploading an image, audio or video stimuli. 2) When previewing your Task either in Task preview or Experiment Preview. 3) As feedback from a participant who has taken part in your Live experiment.
Gorilla's Media Error messages have been added to help you create a task which will work consitently and robustly across all browsers and devices.
Media Error messages will be triggered if Gorilla notices that there is something wrong with your stimuli which could result in your task performing sub-optimally or, in the worst case, prevent your task from working at all.
1) The most common Media Error message which occurs will be informing you that the stimuli you have just uploaded is not in a supported format.
When you preview your Task you may find that your task and stimuli seem to work just fine. However, depending upon the stimuli type you may find that performance will differ between different browsers and different devices.
To avoid any risk and disapointment that your task will not display correctly for some participants we highly recommend you convert your stimuli to one of the supported web-compatible file types.
If you are sure of which file types are supported you can find out here: Supported File Types. Alternatively, you can find out which file types are supported on a particular Zone by reviewing the Zone's dedicated Tooling Reference Guide page.
2) The most common Media Error message which occurs when previewing a Task is that the media cannot be found.
If these steps do not fix your error please refer to our dedicated Media Error Messages help page.
3) As feedback from a participant who has taken part in your Live experiment.
Use our dedicated Media Error Messages help page to identify the message you recieved and find the appropriate solution.
The majority of reports from participants about Media Error messages can be avoided by previewing your task and experiment fully before releasing your experiment Live.
Once this is successful a small pilot across all browsers you intend to use during your experiment is very highly recommended.
Piloting your full study across all browsers and devices you intend to use will usually pick up any additional media errors which may occur and allow you to prevent them from reoccuring when you come to launch your full experiment.
If your participant receives a 'Please switch to landscape mode' message when completing your task on mobile and this doesn't go away when the phone is rotated, there are two common explanations.
Your participant has auto-rotate locked on their phone. This means the phone, and therefore Gorilla, will not switch to landscape mode when the participant turns their phone. To continue with the task, your participant needs to turn this off. We recommend that when recruiting via mobile, you ask participants to turn auto-rotate lock off before they begin the task.
Your participant has opened the experiment link through an app, such as Facebook or Instagram. This may open the link inside the app rather than within a full browser, and this may not be able to pick up on phone rotation. We strongly recommend that when recruiting via mobile, you ask participants to copy-paste the link into their browser.
These are the most common issues related to viewing metrics that are reported:
As of August 2018, this problem should not occur because Gorilla is now able to interpret special characters in a non-universal format. However, if this does occur, follow the steps below.
This problem is typically encountered when uploading a spreadsheet for your Task in a csv format.
Explanation: This problem occurs because, in many programs like Excel, csv files are not saved using UTF-8 encoding by default. This means that these files don’t save special characters in a way that can be universally understood and recognised. When they are uploaded to Gorilla, because the special characters aren’t readable, they are instead replaced with the diamond with a question mark symbol which looks like this: �.
Solution: The way to resolve this is to save your csv file as a XLSX or ODS file, or as an 'CSV UTF-8 (Comma delimited)' compliant one.
Older versions of EXCEL (2013 and earlier) don't directly offer a 'CSV UTF-8' file format. If you still wish to save this as a CSV, follow the steps below:
If you are using an older version of EXCEL and trying to resave a csv file to use UTF-8 encoding, the above steps do not work reliably! In this case:
This will then convert your csv file to use UTF-8 encoding.
There are two main reasons for this issue:
Here are the solutions for each case:
For those who have downloaded the data file prior to August 2018:
The solution for this issue is to change your computers local 'Delimiter' settings. Please refer to the instructions below, appropriate for your operating system.
Step 1: Click the Start button, and then click Control Panel
Step 2: Click on 'Clock, Language and Region' menu section.
Step 3: In this sub-menu, click 'Region' to display the Region 'dialog-box'
Step 4: Click Additional Settings button, to open the Customize Format 'dialog-box'.
The instructions for windows 10 are identical to Windows-7 with one exception:
Step1: Locate the Control Panel by typing in 'Control Panel' in the Start Menu, search bar.
Some metrics appear in a different order in my data between participants
The most common occurance of this issue is found in Questionnaire metrics.
Downloaded metrics are, among other things, ordered by UTC timestamp, which is the time the metric is recieved by our database. For Questionnaires, because all of the responses in a Questionnaire are collected and uploaded simultaneously, individual responses can sometimes arrive at the database at slightly different times. This can result a participants responses appearing a different order to other participants. The same can occur for tasks, where metrics that are uploaded very close together can sometimes appear in a different order than would be expected.
In both cases, there is no change necessary nor cause for concern.
If you wish you can reorder your metrics within Excel based on Local Timestamp (which is the time the metric was initially recorded on the respondents device) rather than UTC as this may mitigate the issue.
One of my participants reaction time metrics are much shorter/longer than I expect should be possible
In order to accurately measure and record a participants reaction times Gorilla must be the active window, at all times, while the participant is undertaking your Task. If the participant navigates away from Gorilla - by switching to a different tab/browser or opening and using a different programme - while undertaking your Task this can lead to inaccurate recording of their reaction times.
This behaviour will be clearly discernible to you in your metrics as a distinct set of reaction times: typically a reaction time much longer than usual followed by a series of shorter reaction times (See the image example below).
Participant behaviour such as this is an example of 'divided attention'. It indicates that the participant is 'distracted' and is not paying full attention to your Task. As such you will probably want to take this into account in your analysis or else redesign your task to mitigate such behaviour - see our suggestions below.
You will likely only encounter this behaviour if your task has lots of time-limited screens back-to-back. i.e. Screens which are set to auto-advance the participant after a set time and which do not require participant input/response in order to advance.
Example of typical 'distracted-participant' reaction time pattern:
While you cannot force participants to stay focused on your task there are improvements and changes you can implement in your tasks to reduce the likelihood of such participant behaviour.
Here are some suggestions:
As of October 2017, improvements to the system for recording metrics means that duplicates are no longer loaded into your results. You should no longer see duplicate metrics appearing in your data download for participant data gathered from October 2017 onwards. Note that we still carry out the safety check described below to make sure that you collected data is stored successfully.
If you believe you are still experiencing this problem please get in touch with us via our contact form
When the browser is uploading metrics to the server it expects to recieve a message back saying the metric was uploaded successfully. If it doesn't recieve this message, it will retry after a timeout as a failsafe. Sometimes, the server is just being a bit slow or the participants internet connection is unreliaable, and so while the first metric did get stored just fine, the browser thinks that it might not have gotten there, and so tries again. This then results in two entries of the same metric. In these cases, we think it's most scientifically appropriate to just disregard the later values.
The participant hasn't see the trial twice, it is simply that the metrics have been uploaded to the server twice. The metrics are identical, so just delete one of the rows.
As of October 2017, a rare edge case was found where a participant with a poor internet connection could refresh on the final screen of a task, while loading the next part of the experiment, and receive the final screen again. This edge case has now been resolved and should no longer occur.
As of August 2017 we have improved how Gorilla records a participant's progress through a task. As a result it should no longer be possible for participants to see a trial twice; due to their having lost connection (e.g. poor internet connection) with the server. As a result the metrics data should not contain repeated trial metrics.
If you believe you have found repeated trials in your data where you are not expecting them, please get in touch with us via our contact form
For Data collected before August 2017:
If a participant's connection fails during the experiment, they can fail to synchronise their current progress through the task. When they then refresh the page, they go back to the last
point at which their progress was synchronised, which may be earlier in the task.
Typically, at the start of the last trial. This can lead to some trials appearing twice or metrics appearing to be out-of-order.
In this situation the participant has seen the trial twice, because the participants connection failed, their progress wasn't saved to the server as the server has no way to know where the participant was. In these cases, we think it is best to use the responses from the first exposure to the trial.
There are a couple of reasons why it can look like you have missing metrics:
1) Have you checked to make sure all the necessary participants are 'included'?
When a participant is 'included' their data is added to your available metrics download. A participant is included automatically when they reach a finish node and Complete your experiment. However, by default, participants who are still 'live' i.e. still working through your experiment, are not included and their data won't appear in your download. If you want to include the data from participants who are still live, go to the Participants tab on your experiment, click the 'Actions' button on a participant and select 'Include.' Alternatively, you can use the 'Force Include All' option to include everyone. Note for Pay-per-Participant users: remember that including a participant consumes a participant token and this process is irreversible. Make sure you only include participants you really want the data for and purchase more tokens if you need to.
2) Are you looking at the right version of your experiments data?
Participant data is associated with a version of an experiment. If you made any changes to your experiment during data collection, your data will be split across the different versions of your experiment. For example, your overall participant count may be 40 but 20 of them were collected in version 2 of your experiment and another 20 were in version 3. To gather the data for all fourty participants, you would need to download the data from both version 2 and version 3. To change the version, you are currently downloading from, on the Data tab, select the appropriate version from the Version Picker. To see what versions of your experiment your participants saw, go to the Participants tab and review the contents of the Version column.
3) Are you sure the participant completed all parts of your experiment?
Not every participant will complete the whole of your experiment! At any time, a participant has the right to withdraw and you will receive no notification of this, other than a sudden stop in the metrics and the participant status remaining 'Live'. If a participant’s data seems to stop part way through a task, check to see if data from that participant appears in any later stages of your experiment. Try to find that participants unique private or publicID in later questionnaires/tasks. If you can't find them there then they most likely decided to leave! All experiments will experience some form of attrition and there is little that can be done to prevent this, unfortunately.
4) Did you use any Experiment Requirement settings?
If you choose to restrict participation in your experiment via any of the requirements set on your experiments recruitment page. i.e. Restrictions based on Device Types, Browser Types, Location or Connection Speed. Participants who enter your experiment but whom fail to meet your specified requirements will not see any of your experiment and will therefore have no metrics recorded.
If you wish to calculate the number of participants being rejected via your selected experiment requirements use a checkpoint node directly after the start node(s) of your experiment. For pay-per-participant accounts holders, we recommend using this method if you wish to determine which 'live' participants you may wish to reject and which you wish to include.
5) Have you tried all the solutions described above?
If you have reviewed and tested all of the solutions listed on this page and they have not resolved your issue, contact us via our contact form and we'll look into it for you as a matter of priority.
In the first instance, try using Gorilla in Chrome. We are committed to supporting all browsers, but some bugs may get through our testing.
If you find a bug (which then doesn't happen in Chrome), we would be immensely grateful if you could fill out the Support Contact Form with details of the bug and what browser you were using as this will help us address the issue! We won't be able to fix it right away, so use Chrome, but it will get it fixed as soon as possible.