Study: How much research participants want to be paid

By Ian Floyd and The Decision Lab9 min. readApr 24, 2025

Survey results next to gift cards and cash.

Good research is almost always incentivized — people want payment for taking part in studies. But how much should you research participants?

Deciding on the right incentive amount is a delicate problem: pay too little, and you won't get enough participants. Pay too much, and you blow up your budget. We did some research to find out the right amount of payment to offer to keep both your participants and your budget happy.


Calculate an incentive in seconds

Just looking for an answer? Try our research incentive calculator and learn how much to pay your research participants.

A graphic explaining how the Tremendous Research Incentive Calculator works. Simply tell us about your study, and you'll learn how much money will satisfy participants.

We’re not the first to try to figure out how much to pay participants. Other companies have created incentive calculators based on historical data. They’re valuable and insightful. But their work is retrospective in nature. It’s an analysis of the status quo — how thousands of researchers have paid participants for various types of research in the past.

We took a different approach. In partnership with The Decision Lab, we used conjoint analyses on over 3,000 data points to measure what reimbursements participants actually want. You can read the full methodology at the end.

For our research, we considered a number of sub-questions:

  • How much should you pay for a survey versus an interview? 

  • How much should you increase the incentive by if it’s a 5-minute commitment, a 15-minute one, or something much longer?

  • Does it matter if the research is about a sensitive topic compared to something surface-level?

  • To what degree does the preferences of your target population matter?

  • How do participants value (or devalue) different incentive types, like cash, Amazon gift cards, Visa prepaid cards, and charitable donations?

With this approach, we’ve been able to measure how the individual variables interact and impact how much you should offer. 

With this guide, we’ll help you get the most bang for your buck — while still doing right by the people at the heart of your study. 

Table of contents

How to approach research incentives, at a glance

We’ve got a super-short answer, a shorter answer, and the long answer.

Super-short answer: Answer a few questions about your study and the Tremendous Research Incentive Calculator will suggest how much to offer.

Short answer: 

  • As a base, you should pay participants $1.76 per minute. As duration increases, the cost-per-minute gets cheaper. (We explain why below.)

  • Interviews cost more than surveys. For a 30-minute interview you should pay about $57, compared to $48 for a survey.

  • Students are comfortable with a lower rate than those in the workforce. And, people earning $200,000 or more expect much higher compensation for their time.

And it also matters how you send the incentive. 

In a separate survey of Tremendous clients, we found that 49% incentivized research using Amazon.com gift cards prior to switching to Tremendous. Yet, in our research with The Decision Lab, we discovered that most participants don’t want to be boxed into a single retailer (even if that retailer is Amazon).

Some highlights from our research

  • Research participants primarily want to be paid in cash transfers. You have to offer more to achieve the same perceived value when using other payout options, such as Visa or Amazon gift cards.

    • However, you need to balance what recipients most want against your own operational needs.

  • People who earn more than $50,000 a year always welcome a gift card when they’re given the ability to select one from a list of retailers.

  • People dealing with unemployment want over $35 more if they’re paid with a gift card (compared to cash transfers).

If you are an occasional researcher, this might be enough information for you to go on. If figuring out how much (and how) to pay subjects is a meaningful part of your work, you might want to keep reading.

Step 1: Figure out how long your study takes

Time is the most important variable to prospective participants. So the first step to finding the optimal incentive is to figure out how much time you’re asking of participants.

A graph showing that people want $70 for a 60 minute task.

For the sake of simplicity, we’ll share some heuristics on how much people want to be paid for tasks with a few different lengths:

  • 15-minute task: People want to be paid about $26.

  • 30-minute task: The amount jumps to $48.

  • 60-minute task: People expect a little more than $69.

Keep in mind, though, that you're seeing these numbers in isolation. Once you start factoring in the study type, participant demographics, and other variables, these amounts change.

A graph showing that cost per minute decreases as study length increases.

One thing that may surprise you is that the longer your experiment runs, the cheaper each additional minute becomes. For example:

  • If you increase the length of your study from 5 minutes to 15 minutes, each additional minute costs $1.76.

  • Now, say you increase the length of your study from 15 to 30 minutes. Each additional minute now costs $1.44. 

  • Finally, say you decide to stretch your interview from 30 minutes to 1 hour. Each additional minute will cost only $0.71. 

These findings have important implications for how researchers should approach their experimental designs. 

While some researchers worry that increasing study length will impact participant motivation, it may be relatively inexpensive to include additional questions while still adequately compensating respondents.

Step 2: Choose between quantitative versus qualitative research

By using the discrete choice methodology, we presented participants with two options and asked them to choose which was better. This allowed us to analyze different variables and assign a dollar value to their choices.

Discrete choice set sample

Research project AResearch project B
Understanding consumer breakfast cereal preferencesInvestigating household budgeting and debt management behaviors
30-minute online interview60-minute online interview
Receive a $46 cash transfer (via Paypal, Venmo, or bank transfer) as compensationReceive a $96 mailed check as compensation
Get half your compensation upon signing up and the other half after completing the taskGet full compensation after completing the task

With this information, we explored whether people prefer taking a survey vs. participating in an interview.

Turns out, people prefer studies that do not require talking to an actual person. More specifically, our participants showed a preference for online surveys over online interviews

  • On average, online interviews need to be compensated at a flat rate of $8.66 more than surveys to be an equally attractive option.

  • People dealing with unemployment need to be compensated $20.57 more to do an interview compared to a survey. For employed people, the difference is only $7.62.

A graph showing that unemployed people need $12.95 more than employed people to find an interview as attractive as a survey.

Step 3: Consider if your study addresses sensitive subjects

Researchers also need to think about the sensitivity of the research subject matter

In our study, we probed how different topics affected participant motivation levels. We asked our participants to choose between two research projects: one about their breakfast cereal preferences and one about household budgeting and debt management. 

For interviews and surveys that take longer than 30 minutes, we found that participants want to be compensated $3.40 more if the experiment deals with a sensitive or personal subject matter (e.g., household debt) as opposed to surface-level information (e.g., cereal preferences).

Asking someone to talk about a topic like household debt for an hour will require about $9 extra in compensation compared to asking someone about a more casual subject for the same period. 

A graph showing that medium-income participants are more sensitive to the type of research they choose to participate in.

Middle-income earners (making between $50,000 and $100,000) showed the strongest sensitivity to research topic. This group would need to be compensated on average $8.37 more to take a long survey about a sensitive subject compared to someone who makes less than $50,000.

The causality of this finding is unclear. Is debt a particularly touchy subject for middle-income earners, or is cereal a particularly exciting one?

Step 4: Determine your target population

Who you target for your research matters, but some characteristics matter more than others.

Participants' income level matters

Intuitively, we expected participants’ income to affect their perception of research incentives. You might think that high earners would assign greater value to their own time, and you’d be right.

A graph showing that students expect 20% less on the dollar than individuals who are employed.

In our sample, respondents at the highest end of the wealth spectrum did have noticeably higher expectations for research compensation. Specifically, those earning $200,000 or more needed to be paid 46% more than low-income earners (under $50,000) for an incentive to have the same appeal.

On the other hand, there's one group willing to be paid less: students.

A lot of university research relies on students for their studies, which makes sense considering they're the most cost-effective group. Students expect 20% less on the dollar compared to other participants.

Step 5: Pick an incentive that appeals to research participants

We tested a variety of different incentive types to see which are the most appealing to research participants. 

The table below explains what each incentive type is. It also displays the shorthand we’ll use for the remainder of this guide. 

Incentive types & shorthand

The incentives we testedShorthand
Cash transfer (via Paypal, Venmo, or bank transfer)Cash transfer
Digital Visa prepaid cardVisa prepaid card
Digital gift card of your choice (choose from 1,000+ options)Gift card or choice of gift card (This is a gift card to a single retailer that the recipient chooses from Tremendous' options.)
Digital Amazon.com gift cardAmazon gift card (These are the most commonly selected gift cards from the Tremendous catalog of options. This is considered a proxy for gift cards to any single retailer.)

How incentives are delivered to participants can impact motivation levels. Overall, cash transfers are always preferred, especially if you’re hoping to engage a lower-income audience.

A graph showing that Visa prepaid cards, gift cards of the recipients choice, Amazon gift cards, and mailed checks are worth less than money transfers.

Specifically, cash offers at least $9.50 of additional value over any other incentive type we studied. This holds true for all study lengths.

  • On the flip side, mailed checks need to value $13.37 more for participants to consider them equal to a cash transfer. 

  • After cash, Visa prepaid cards are the next best option, presumably due to their flexibility. 

  • Gift cards the participant can choose themself were valued slightly more than gift cards for Amazon.


☝️ So why not just offer cash transfers?

While participants prefer options like bank transfers, PayPal and Venmo, they’re either a pain to set up or expensive to offer.

Sending money to a large number of people directly from a bank or PayPal account is tedious, fraught with errors, and risks tripping fraud alerts. Plus, international transfer options are limited.

Many researchers find that the happy medium is to offer a payout redeemable for a wide assortment of prepaid and gift cards — which, if you use Tremendous, is simple, free, and covers over 200 countries.


We also found some interactions between participant characteristics and the incentives they preferred. Namely, if your target audience earns above $50,000 per year, a choice of gift cards is an equally attractive option to cash and Visa prepaid cards. 

If your participants include unemployed people, avoid retailer-specific gift cards, even to large retailers like Amazon. A gift card would need to be loaded with around $37 more than a cash transfer to be valued equally (or $26 more than a Visa prepaid card).

A graph showing that people who earn under $50k and unemployed people don't want gift cards.

Step 6: Ignore what doesn’t matter 

It’s just as valuable to know what you don’t need to care about. Below is one such consideration, which didn’t yield any significant findings in our analyses. 

Payment timing doesn’t matter

It matters how much you pay your participants, and how you pay them — but when you pay them doesn’t seem to make a difference for incentivization. 

We tested a number of payment schemes, including upfront payment, payment after the study was completed, and a blend (with partial payment before and the remainder after). 

No significant differences emerged between these options. 

Step 7: Piece together research incentives your participant will love

When building an incentive strategy for your research, keep these rules in mind:

1. Study length has the greatest influence on how participants perceive research incentives

  • The relationship between study length and incentive expectations is not linear.

  • The longer your experiment runs, the cheaper each additional minute becomes. 

  • If you were to increase the length of your study from 5 to 15 minutes, each of those additional 10 minutes will cost you $1.76. 

  • But if you were to lengthen a 30-minute study to 60 minutes, each extra minute would cost only $0.71.

2. Participants require more compensation for research modalities that involve human interaction

  • Online interviews cost more than online surveys.

3. Studies revolving around sensitive subject matter are more expensive if they’re on the longer side

  • Participants expect higher compensation if you want them to talk about touchy subjects for an extended period of time.

4. Cash transfers are always preferred and offer at least $9.50 in additional value over any other incentive type we studied

  • Participants were least interested in mailed checks. They needed the checks to be $13.37 greater to perceive them as equal in value to cash transfers.

5. Payment scheme (whether participants were compensated before the study, after, or a blend of the two) did not affect motivation

  • But keep in mind that people in your target population might feel differently about this if they don't regularly participate in research.

6. Participant employment status and income level showed some differences in compensation expectations

  • Very high earners (over $200,000 per year) expected larger monetary incentives compared to other participants.

  • Students were the least expensive group, requiring 20% less than middle-income and unemployed groups to achieve the same utility. 

Step 8: Use our research incentive calculator

We built a research incentive calculator based on these insights. Input the characteristics of your study, and let it recommend the optimal research incentive. 

A graphic explaining how the Tremendous Research Incentive Calculator works. Simply tell us about your study, and you'll learn how much money will satisfy participants.

Send research incentives using Tremendous

Ready to send your next batch of participant rewards? If you don’t already, use Tremendous for research incentives — it’s fast, free, global, and it integrates with your favorite research platforms.

Sign up and start sending immediately. You can also strategize on how best to set up your program by booking a demo with our sales team.

Thousands of brands trust Tremendous to send research incentives.


Watch the webinar


About the research

Tremendous is investing in research about incentives, so we can help you make more informed decisions. To achieve this goal, we’re also being transparent about the methodology and limitations.

Benchmarking

We compared our calculator’s outputs to those of existing incentive calculators, to see how these different methodologies stacked up. Here’s what we found: 

  • Overall, our calculator was aligned with the upper range of the comparison calculator we chose (which provided minimum, average, and maximum incentive amounts). 

  • On average, our recommendations were $5 more than the maximum incentive recommended by the comparison calculator. This was true across study formats (surveys vs. interviews) and various study lengths.

A graph showing that our recommendations are $5 more expensive than comparative calculator's recommendations.

One interesting discrepancy was noted for interview incentives

  • Our calculator starts off at a lower rate for short interviews. However, as interview length increases, we recommend relatively large increases in compensation. 

  • On the other hand, the comparison calculator recommends a larger baseline incentive for short interviews, with smaller increases for longer interviews.

  • Given that the comparison calculator is based on observational data, this suggests that researchers may be overestimating how much money it takes to get someone through the door for an interview, and underestimating how much people value the additional time taken for longer interviews.

To further test how our calculator’s outputs compare with the current state of research incentives, we compared our recommendations with pay practices of over 150 real-world solicitations for research participants. These calls were published via two large online recruitment platforms.

We found a striking difference in the current state of research incentives across the two platforms. Specifically:

  • Across surveys and interviews of different lengths, the first platform’s payment practices aligned well with our recommendations. Across surveys and interviews of different lengths, their rates of compensation were between 57% and 152% of our calculator’s recommended incentive.

  • On the other hand, the second platform compensated participants at a much lower level. Across surveys of different lengths, their compensation clocked in somewhere between 8% and 38% of our calculator’s recommendation!

Methodology

The aims of this research, conducted in May 2023, included:

  • Understanding customers’ relative preference for different research incentive types and ways in which incentives are presented. 

  • Testing whether incentive preferences differ for certain groups of participants.

  • Assessing how standard rates of compensation measure up with participants’ valuation of their own time.

Experimental design

We used a Discrete Choice Experiment (DCE). DCEs include a set of realistic, randomly generated options that participants choose between. For example, participants may be asked to select between a series of hypothetical research studies where types and rates of compensation and the lengths and topics of research vary randomly. By analyzing a large number of these choices, we gain insight into preference. 

We used a large online research platform to test 300 people living in the US, with a balanced gender distribution. Each individual made 10 decisions, giving us a total of 3,000 data points.

Discrete Choice Experiment

We tested participant preferences for different types of research incentives, and how perceptions of research incentives varied depending on study characteristics. Participants were presented with information about two hypothetical research studies at a time, and then asked to indicate which one they would rather complete.

Example of choice set:

A graph displaying a discrete choice experiment. One option for an experiment involved questions about consumer breakfast cereal preferences, while one investigated household budgeting and debt management behaviors.

Following the DCE, we gathered demographic data (age, income, employment status) to test preferences between groups. During the experiment, we implemented attention checks to gauge data quality. 

Analysis

Consistent with Hainmueller, 2014, we used a linear regression model to estimate the utilities/value associated with the levels of each attribute in each DCE. Only interactions that were significant at the 95% confidence level were reported. For each stage, we used separate models to test the following:

  • The main effect of each attribute 

  • Interactions between attributes

  • Interactions between attributes and demographic factors

Each model included an estimate associated with changes in the size of the incentive. This estimate can be interpreted as the perceived value of an additional dollar added to the incentive. We used this estimated value of an additional dollar to convert the differences in perceived value between attribute levels into dollar values.

Published October 19, 2023
Updated April 24, 2025

Share this article

Facebook
Twitter
LinkedIn