*This post was previously published at **EA forum** by Peter Hurford*

We all make decisions every day. Some of these decisions are pretty inconsequential, such as what to have for an afternoon snack. Some of these decisions are quite consequential, such as where to live or what to dedicate the next year of your life to. Finding a way to make these decisions better is important.

The folks at Charity Science Health and I have been using the same method to make many of our major decisions for the past few years – everything from where to live to even deciding to create Charity Science Health. The method isn’t particularly novel, but we definitely think it is quite underused. Here it is, as a ten step process:

Come up with a well-defined goal.

Brainstorm many plausible solutions to achieve that goal.

Create criteria through which you will evaluate those solutions.

Create custom weights for the criteria.

Quickly use intuition to prioritize the solutions on the criteria so far (e.g., high, medium, and low)

Come up with research questions that would help you determine how well each solution fits the criteria

Use the research questions to do shallow research into the top ideas (you can review more ideas depending on how long the research takes per idea, how important the decision is, and/or how confident you are in your intuitions)

Use research to rerate and rerank the solutions

Pick the top ideas worth testing and do deeper research or MVP testing, as is applicable

Repeat steps 8 and 9 until sufficiently confident in a decision.

## WHICH CHARITY SHOULD I START?

iThe definitive example for this process was the Charity Entrepreneurship project, where our team decided which charity would be the best possible charity to create.

**Come up with a well-defined goal:** I want to start an effective global poverty charity, where effective is taken to mean a low cost per life saved comparable to current GiveWell top charities.

**Brainstorm many plausible solutions to achieve that goal:** For this, we decided to start by looking at the intervention level. Since there are thousands of potential interventions, we placed a lot of emphasis on plausibly highly effective, and chose to look at GiveWell’s priority programs plus a few that we thought were worthy additions.

**Create criteria through which you will evaluate those solutions / create custom weights for the criteria:** For this decision, we spent a full month of our six month project thinking through the criteria. We weighted criteria based on both importance and the expected variance that would occur between our options. We decided to strongly value cost-effectiveness, flexibility, and scalability. We moderately valued strength of evidence, metric focus, and indirect effects. We weakly valued logistical possibility and other factors.

**Come up with research questions that would help you determine how well each solution fits the criteria:** We came up with the following list of questions and research process.

**Use the research questions to do shallow research into the top ideas, use research to rerate and rerank the solutions:** Since this choice was important and we were pretty uninformed about the different interventions, we did shallow research into all of the choices. We then produced the following spreadsheet:

Afterwards, it was pretty easy to drop 22 out of the 30 possible choices and go with a top eight (the eight that ranked 7 or higher on our scale).
**Pick the top ideas worth testing and do deeper research or MVP testing, as is applicable / Repeat steps 8 and 9 until sufficiently confident in a decision:** We then researched the top eight more deeply, with a keen idea to turn them into concrete charity ideas rather than amorphous interventions. When reranking, we came up with a top five, and wrote up more detailed reports - SMS immunization reminders, tobacco taxation, iron and folic acid fortification, conditional cash transfers, and a poverty research organization. A key aspect to this narrowing was also talking to relevant experts, which we wish we did earlier on in the process as it could quickly eliminate unpromising options.
**Pick the top ideas worth testing and do deeper research or MVP testing, as is applicable:** As we researched further, it became more clear that SMS immunization reminders performed best on the criteria being highly cost-effective, with a high strength of evidence and easy testability. However, the other four finalists are also excellent opportunities and we strongly invite other teams to invest in creating charities in those four areas.

## WHICH CONDO SHOULD I BUY?

**Come up with a well-defined goal:** I want to buy a condo that is (a) a good place to live and (b) a reasonable investment.

**Brainstorm many plausible solutions to achieve that goal:** For this, I searched around on Zillow and found several candidate properties.

**Create criteria through which you will evaluate those solutions:** For this decision, I looked at the purchasing cost of the condo, the HOA fee, whether or not the condo had parking, the property tax, how much I could expect to rent the condo out, whether or not the condo had a balcony, whether or not the condo had a dishwasher, how bright the space was, how open the space was, how large the kitchen was, and Zillow’s projection of future home value.

**Create custom weights for the criteria:** For this decision, I wanted to turn things roughly into a personal dollar value, where I could calculate the benefits minus the costs. The costs were the purchasing cost of the condo turned into a monthly mortgage payment, plus the annual HOA fee, plus the property tax. The benefits were the expected annual rent plus half of Zillow’s expectation for how much the property would increase in value over the next year, to be a touch conservative. I also added some more arbitrary bonuses: +$500 bonus if there was a dishwasher, +$500 bonus if there was a balcony, and up to +$1000 depending on how much I liked the size of the kitchen. I also added +$3600 if there was a parking space, since the space could be rented out to others as I did not have a car. Solutions would be graded on benefits minus costs model.

**Quickly use intuition to prioritize the solutions on the criteria so far:** Ranking the properties was very straightforward. I could skip to plugging in numbers directly from the property data and the photos.

**Come up with research questions that would help you determine how well each solution fits the criteria:** For this, the research was just to go visit the property and confirm the assessments.
**Use the research questions to do shallow research into the top ideas, use research to rerate and rerank the solutions:** Pretty easy, not much changed as I went to actually investigate.
**Pick the top ideas worth testing and do deeper research or MVP testing, as is applicable:** For this, I just ended up purchasing the highest ranking condo, which was a mostly straightforward process. Property A wins!
This is a good example of how easy it is to readapt the process and how you can weight criteria in nonlinear ways.

## HOW SHOULD WE FUNDRAISE?

**Come up with a well-defined goal:** I want to find the fundraising method with the best return on investment.
**Brainstorm many plausible solutions to achieve that goal:** For this, our Charity Science Outreach team conducted a literature review of fundraising methods and asked experts, creating a list of the 25 different fundraising ideas.
**Create criteria through which you will evaluate those solutions / Create custom weights for the criteria:** The criteria we used here was pretty similar to the criteria we later used for picking a charity – we valued ease of testing, the estimated return on investment, the strength of the evidence, and the scalability potential roughly equally.
**Come up with research questions that would help you determine how well each solution fits the criteria:** We created this rubric with questions:

What research says on it (e.g. expected fundraising ratios, success rates, necessary prerequisites)

What are some relevant comparisons to similar fundraising approaches? How well do they work?

What types/sizes of organizations is this type of fundraising best for?

How common is this type of fundraising, in nonprofits generally and in similar nonprofits (global health)?

How would we run a minimum cost experiment in this area?

What is the expected time, cost, and outcome for the experiment?

What is the expected value?

What is the expected time cost to get the best time per $ ratio (e.g., would we have to have 100 staff or a huge budget to make this effective)?

What further research should be done if we were going to run this approach?

**Use the research questions to do shallow research into the top ideas, use research to rerate and rerank the solutions:** After reviewing, we were able to narrow the 25 down to eight finalists: legacy fundraising, online ads, door-to-door, niche marketing, events, networking, peer-to-peer fundraising, and grant writing.
**Pick the top ideas worth testing and do deeper research or MVP testing, as is applicable:** We did MVPs of all eight of the top ideas and eventually decided that three of the ideas were worth pursuing full-time: online ads, peer-to-peer fundraising, and legacy fundraising.

## WHO SHOULD WE HIRE?

**Come up with a well-defined goal:** I want to hire the employee who will contribute the most to our organization.
**Brainstorm many plausible solutions to achieve that goal:** For this, we had the applicants who applied to our job ad.
**Create criteria through which you will evaluate those solutions / Create custom weights for the criteria:** We thought broadly about what good qualities a hire would have, and decided to heavily weight values fit and prior experience with the job, and then roughly equally value autonomy, communication skills, creative problem solving, the ability to break down tasks, and the ability to learn new skills.
**Quickly use intuition to prioritize the solutions on the criteria so far: **We started by ranking hires based on their resumes and written applications. (Note that to protect the anonymity of our applicants, the following information is fictional.)

**Use the research questions to do shallow research into the top ideas, use research to rerate and rerank the solutions:** After our Skype interviews, we reranked all the applicants.

**Pick the top ideas worth testing and do deeper research or MVP testing, as is applicable:** While “MVP testing” may not be polite to extend to people, we do a form of MVP testing by only offering our applicants one month trials before converting to a permanent hire.

## WHICH TELEVISION SHOW SHOULD WE WATCH?

**Come up with a well-defined goal:** Our friend group wants to watch a new TV show together that we’d enjoy the most.
**Brainstorm many plausible solutions to achieve that goal:** We each submitted one TV show, which created our solution pool.
**Create criteria through which you will evaluate those solutions/custom weights for the criteria:** For this decision, the criteria was the enjoyment value of each participant, weighted equally.
**Pick the top ideas worth testing and do deeper research or MVP testing, as is applicable:** We then watched the winning television show, which was Black Mirror. Fun!

## WHICH STATISTICS COURSE SHOULD I TAKE?

**Come up with a well-defined goal:** I want to learn as much statistics as fast as possible, without having the time to invest in taking every course.
**Brainstorm many plausible solutions to achieve that goal:** For this, we searched around on the internet and found ten online classes and three books.
**Create criteria through which you will evaluate those solutions/custom weights for the criteria:** For this decision, we heavily weighted breadth and time cost, weighted depth and monetary cost, and weakly weighted how interesting the course was and whether the course provided a tangible credential that could go on a resume.
**Quickly use intuition to prioritize the solutions on the criteria so far:** By looking at the syllabi, table of contents, and reading around online, we came up with some initial rankings:

**Pick the top ideas worth testing and do deeper research or MVP testing, as is applicable:** Only the first three felt deep enough. Only one of them was free, but we were luckily able to find a way to audit the two expensive classes. After a review of all three, we ended up going with “Master Statistics with R”.

Awesome process! Loved the many clear examples I found it a little hard to fully understand:

Come up with research questions that would help you determine how well each solution fits the criteria:.Maybe adding another simple explanation there can help.