The testers had initial positive impressions of Chef'd's clean design and displayed food photos. Harrison remained engaged throughout the homepage, while Melody lost interest halfway through. Both enjoyed seeing prices below meals. When completing tasks, testers had to infer how to access Spoon University plans, as its logo is not prominently displayed. Selecting proteins caused confusion due to inconsistent interactions. Determining total costs took extra steps of scrolling to find prices.
"Boost Your Digital Presence: Partner with a Leading SEO Agency"
Usabilitytestforchefd zoe green
1. Green1
Zoe Green
Intro to Web Publishing
Usability Test for Chef’d
TEST PREPARATION
With so many meal kit delivery services out in the world, user experience is
important for companies to study to ensure that their visitors have a delightful
experience. Conducting usability tests help to study their websites and point out
problems that users have while completing tasks. All of this information is assessed and
recommendations are given to redesign or change the structure of the site.
I prepared for these usability tests by going through and attempting to complete
each task myself while documenting my problems and overall experience. This initial
test took place on Wednesday at 11:15 a.m. on the third floor of the Northwest Missouri
State library. I was using my Northwest laptop, a HP Elitebook, and Google Chrome as
my browser. There is a silence rule in this area so I had very little distractions, and I was
wearing headphones to block out any possible noise. I was sitting at a window facing
the highrises, and I had both natural and fluorescent lighting so I could see easily. The
biggest distractions I had were from Facebook Messenger on my desktop that would
send notifications, and my cell phone that was sending me texts and other alerts. This
test occurred only two days before an art show that I was assisting in curating, so I had
many people trying to contact me during this usability test.
I typed chefd.com into the search bar, and this sent me to the home page, which
loaded in about four seconds with the last feature being the slideshow. The first thing I
noticed when I scrolled through the landing page was Chef’d is minimalist in style.
2. Green2
There is a lot of white space, and the majority of the text is either black or a dark gray.
The only color on the page comes from occasional orange buttons and photographs,
with little variety until you go down to the partners’ slideshow at the bottom. Nothing felt
overwhelming, and the one of the first things I noticed was a Saint Patrick’s Day
advertisement. It offered a 100-dollar gift card, which is a great incentive for me to like
this company and to look
through the website. The
home page gives a decent
idea of what the company is
about, with three different
sections featuring their food,
such as the first slideshow,
best sellers, and new meals. Chef’d also shows what makes them different from other
meal kit services by featuring their celebrity chefs, notable partners, and well-known
news sources that have written about them. So far, I find it to be a good experience on
the site due to its simple design and that I can easily see what Chef’d does and
promotes.
Task one is to build a Spoon University meal plan and determine the total weekly
cost. In the scenario, you are only working with 65-dollars, and I thought that will
become a problem for the future usability tests. I have studied the content of this
website, and I know that the meal kit service can have high prices, especially if the
users are planning for something weekly. Spoon University is one of their sponsors, so I
clicked on their Partners tab and scroll until I found them. The future users though will
3. Green3
probably not be aware of this way to access Spoon University’s meal plan page, and
their logo does not appear on the homepage until the second slide at the very bottom of
the page. They may not be able to find this specific meal plan right away because there
are eight tabs on the navigation to look through. When I went back to see if there was a
direct way for users to reach Spoon University, I found that it was not until the last tab,
Meal Plans, which listed them in the drop down menu. When I looked through the
Spoon University Meal Plan page, they provided two headers titled, “What It Is” and
“What You’ll Get” with short pieces of information about the service and the type of food
provided. Further down they provided reviews and press specifically about Spoon
University, which help validated and give credibility to Chef’d. Overall, Chef’d offers a
decent amount of information for the visitors, and I believe there will be little confusion
from this page. At the very top, a large, orange button says, “Start Meal Plan” which
when clicked brings you to another page that says “Build Your Box.” When picking my
proteins I encountered two possible problems: the first was that all the protein is
highlighted orange. My first response is to select what protein that I prefer, but that turns
the options gray, meaning to deselect. This goes against the consistency and standards
heuristic by Nielson, and could cause confusion. Second, when I only selected beef and
poultry, I could not continue. A popup said that I had to
select at least six kinds of proteins out of the seven.
This is an error that could have been prevented by
stating that when the user was picking out proteins,
and could hinder the experience of my future tests.
Furthermore, when I selected only three instead of the
4. Green4
requested six, Chef’d still let me continue onto the next part of choosing the number of
meals. When choosing between 0-7 meals in my box, I knew that I would have to stay
at a low number to be able to stay under my budget. I originally clicked on two meals,
and ten different photographs of food displayed. If my users do not scroll through and
read the first page explaining Spoon University’s meal plan, this could cause confusion
as to why there are more than the two meals that originally selected. You cannot see
the price until you scroll to the very bottom, and the meal plan with two entrees is 75-
dollars. Going back up and adjusting the amount of meals to one brought down the
price to 62, and from there I went to checkout to see what the delivery prices were.
However, when I went to checkout, Chef’d alerted me that the Four Cheese Mac n’
Cheese was unavailable. I had to go
back and redo the entire process,
swap the meal and then checkout
again. This went against the error
prevention heuristic of Nielson, and
could cause frustration for my
usability tests. The checkout did not charge delivery, so the correct answer to complete
the task is still one meal for 62-dollars.
Task two is, “You have the budget to cover one extra meal per week, but you want it to
be a spicy meal. Determine your spicy meal options and choose one meal that you are
willing to prepare.” I went through and created my meal plan again, the same way that I
did in the first task. Then I clicked on the provided meal, Chili Mac n’ Cheese, and on
the recipe card was able to see the spiciness level that listed the dish as mild. When I
5. Green5
went to swap the meal, I had to go through each of the 14 options individually and could
not find a spicy dish. Unfortunately, at this point, I
clicked on the Spaghetti Carbonara and the link broke,
and would not allow me to click on any of the other
dishes. Overall, this test failed and ultimately caused a
large problem for the technical part of the test.
Task three was to, “Ensure none of your meals, “Grab ‘n’ Go’s” or snacks is more
than 450 calories per serving. If the calorie count is too high, swap the item for another.”
I still had my meal plan from the last task, and I completed this one in a very similar
way. I clicked on each photo and pulled up the recipe card to find the nutritional
information. One meal I had to swap out, which was a simple process of clicking swap
meal and looking through the other options and their calories. Though this task was
easy, it did take a while, and could frustrate or annoy the users in their experience. The
information is also not clearly available, so it relies on the user to naturally click and
expect a linked photo to the spice level and nutritional information.
The last task was to, “Assume/pretend that it is Monday at 3:30pm, and you plan
to prepare the spicy meal on Saturday night. Will the meal arrive on time?” This was a
difficult and slightly confusing question because Chef’d delivery information cannot be
set at future or past dates. Instead, I went off the time span of five days to conduct this
task. When I went to the checkout during the first task, I had to set the delivery date, so
I went back and looked at the shortest amount of time it would take to receive my meal
kit. The earliest I could receive my food was March 14, a week from today, so this task
6. Green6
was failed. Depending on if my users are accustomed to Amazon Prime with their very
short wait times, this could be upsetting. The task, however, was easy to figure out and
navigate.
Overall, the tasks were easy to complete, or at least figure out that they were not
possible. Though I was already familiar with Chef’d and their content, so I knew it was
going to be a very different experience for the people assisting with the usability test.
The biggest issues that I foresee is the protein selection and how the meal plan is set
up with more food than just the originally selected entrees.
CHOOSING PARTICIPANTS
Tester 1: Harrison Boden
Harrison Boden is a 20-year-old male, who is a sophomore at Northwest Missouri
State University, and is an art major with an emphasis in ceramics. He has two jobs,
which he works about 35 hours total: one as an assistant in the pharmacy of Hy-Vee,
then as a server at Simply Siam. Harrison is familiar with meal kit delivery from one of
his professors using Blue Apron, and seeing ads on social media. He also has been
interested in eating healthier and going out to restaurants less despite not having much
time between school and two jobs. When asked how much time on the internet he
spends each week, he estimated about 90 hours a week due to the fact he has to use
the internet at his pharmacy job. He elaborated and said that about half of that is
recreation. Considering the split between social media and browsing, Harrison said that
he rarely does any browsing, with 90% being social media. Overall, he considered
7. Green7
himself to have high experience with the internet, having used it for his job and
entertainment.
Overall, I believe Harrison to be a reasonable tester because he is interested in
improving his health and preparing more meals at home. He has experience in graphic
design and can give educated assessments over the design and layout of the website.
Environment for Tester 1
Location of test:
I conducted the test for Harrison in the advanced ceramics room of the Fire Arts
Building on Northwest Missouri State’s campus. This is where his private studio is, and
where he spends a large amount of his time using the internet other than work. He is
comfortable here, and right behind his studio is an upright futon where he is usually
relaxing or studying. We took the test sitting on this futon Tuesday, February 6 at 1:15
p.m.
Physical Environment:
The lighting of the environment came from the bright, natural light of the windows
that are floor to ceiling, and also the fluorescent bulbs that hang from across the room
as well. There was no one else in the advanced ceramics room at the time, but noises
such as the kilns, air ducts, and voices coming in from other rooms created a white
noise in the environment. The futon faces the windows so occasionally there was the
distraction of people walking by or into the building.
Technical Environment:
8. Green8
Harrison used a Mac laptop for his test, and connected to the Northwest Missouri
State Wi-Fi to access the internet. He used Safari for this browser, and did not have any
other tabs or programs open while he was completing the usability test.
Tester 2: Melody Monroe
Melody is a 21-year-old female who is a senior at Northwest Missouri State
University, where she majors as an art major with an emphasis in ceramics, and a minor
in business. Melody has two jobs and works about 15 hours a week: one as a kiln
manager at Northwest Missouri State, and the second as a server at Simply Siam. She
is familiar with meal kit delivery services through advertisements on Pandora, and has
looked at the Blue Apron website without purchasing anything. However, Melody is very
health conscious; she consistently works out and carefully watches what she eats while
being a vegetarian. When asked how much she uses the internet per week, she
estimated that she is online about 35 to 45 hours. Melody guessed that about 20 hours
of that is for social media, though she did say that she counted listening to Pandora as
part of her internet usage. Overall, she considers herself highly experienced with the
“regular internet,” as she put it, unlike Harrison who has to do a lot of in-depth research
online for his pharmacy job.
I choose Melody as my second tester because I wanted to be able to test
Chef’d’s usability with someone who is vegetarian. Furthermore, her health conscious
lifestyle creates a strong contrast to Harrison, who is only starting to invest in his health.
Melody also spends about half the amount of time on the internet as Harrison does.
With these comparisons, I hope that their tests will be able to show the user experience
problems that Chef’d has.
9. Green9
Environment for Tester 2:
Location of Test:
Melody also tested in the Fire Arts Building on Northwest Missouri State’s
campus. Though she choose to have the test performed in the open conference room of
the building, where she goes to do a lot of her homework and relax. This room is open
to students during the day to come in and work or spend down time and used only
occasionally for lectures or class time. Melody tested on Wednesday, March 7 at 2:15
p.m.
Physical Environment:
The physical environment has covered windows, and the only light is coming
from the fluorescent lights that are hanging overhead. It is a small room with two long,
black tables pushed together with office chairs surrounding them, and a projector with a
screen on one of the walls. No one was in the room, and most distracting noise was
from the people who were passing by the door. Melody had brought her lunch from
Simply Siam in with her, ate half of it, and then left the remainder behind her laptop. The
strong Thai smell of her food was present the entire time of the test and was distracting.
Overall, it was a calm and quiet environment.
Technical Environment:
Melody performed the usability test using her Northwest Missouri State laptop, a
HP Elitebook, which connected to the campus Wi-Fi. She used Google Chrome as her
browser and did not have any extensions or add-ons. Melody also did not have any
other tabs or programs clearly running while testing.
10. Green10
TEST RESULTS
Initial Site Thoughts
Summary:
Harrison started with his Safari search engine and typed in Chef’d, which pulled
up first in the results. His initial comment was about their name, saying it strange to
pronounce with the “f’d.” The home page loaded quickly, and he immediately noted the
clean, sleek design and the “good pictures.” He liked that the food looked professionally
plated, and commented on how he was interested in the featured meals. Harrison’s art
is clear and minimalist in design so he enjoyed the consistency and small range of color
and text. He made note that Chef’d openly shows their prices for the food, and he
thought that the cost seemed reasonable. Scrolling further down he noticed the
Celebrity Chefs and Proud Partners features, which he commented, “Nice to know who
makes your food…Know where your food is coming from.” Harrison’s attention went to
the app, and mentioned that he would definitely be interested in using it when on
mobile. Overall, he described his first impression of the homepage to be a good and
interesting experience, and from the information, he had a good idea of what Chef’d
provided and how.
Melody’s visit to the homepage also loaded quickly, and she immediately
commented on the simple, light colors. She enjoyed looked at the main feature, the
slideshow, and continued to scroll down the page. There was a popup about 20
seconds into her visit, which definitely irritated her, and she kept accidently going over
the navigation and having her content blocked by the pulldown. She noted that she
11. Green11
became bored by the end of the scrolling through the homepage, though overall she
likes the clean layout. Melody also mentioned that she appreciated Chef’d offering the
prices below the food. Altogether, she felt that she was interested and enjoyed what she
was looking at with no confusion.
When comparing the two testers, they both agreed that the aesthetic of Chef’d
was great and pleasing to look at. They enjoyed the displayed pictures with their prices,
and the basic information that gave them a general idea of what the company was
offering. Both testers agreed that they were interested in the site’s services and wanted
to continue to look around and click on links. Harrison stayed engaged the entire time
while scrolling through the homepage, while Melody had lost interest by about the
celebrity chef feature. In addition, while Harrison did not mind the popup, it clearly
irritated Melody that she was interrupted and had to close it out to continue looking.
Task 1: Build a Spoon University meal plan and determine the total weekly cost.
Summary for Both Testers:
Highlights:
1) Testers had to infer where the Spoon University Meal Plans would be.
On the home page there are eight tabs on the navigation, not including the
search engine icon, and Spoon University isn’t displayed until the very last
Tester
1 Tester2 Average
Average Satisfaction 4 1 2.5
SuccessRate 100% 100% 100%
12. Green12
tab, Meal Plans. The link also is at the very bottom of the page, but is not until
the second page of the Proud Partners slideshow.
2) Testers tried to select protein options instead of deselect.
The first question had the users click
what things they are allergic to, but
then for the proteins they had to
deselect what they did not want. To be
noted: that unlike during my
experience, neither Harrison nor
Melody had the popup about not
including enough protein. During both
of their usability tests, Harrison and
Melody deselected their preferred
protein.
3) Testers were unsure how many meals to select to be able to stay
underneath 65-dollars.
After choosing their protein, Chef’d takes the user to the next question of how
many meals they would like a week, from 0-7. Melody and Harrison knew that
13. Green13
they had to stay under 65-dollars, but they did not
know how much the meal plan costs yet. Harrison
initially chose seven and Melody clicked on three,
and both discovered at the bottom of the next page
under their meals that they are over budget. Both
had to continuously go back and adjust their meals
numbers until finally choosing one meal a week. There was also confusion
when they picked their number of meals, but had six Grab ‘n’ Go’s and two
snacks included as well. While Harrison and Melody scanned the main page
of Spoon University, neither of them read closely enough about what all was
in the box.
Biggest Problem:
The biggest problem that Melody and Harrison had was number two,
selecting their preferred protein. Harrison wanted all the options, so he clicked
on every icon, turning them gray and actually deselecting them. From there
he tried to click the next button, which an error came up that he had no
protein selected. Harrison realized what had happened and was able to
correct his options, though he commented that he did not think the
deselecting made sense when posed with the question, “I like meals that
include.”
Melody’s experience with the protein was even worse, and the reason that
she gave a one for satisfaction. She did not realize that she needed to
14. Green14
deselect the protein, so she accidentally deselected the vegetarian and fish
options. Without Chef’d correcting this error, like Harrison, Melody continued
to the next pages and grew increasingly frustrated that she was given all the
wrong protein options. She started with selecting three meals, and eventually
got down to one meal that was under 65-dollars. At this point she commented
that she was really upset that she would not be able to eat the majority of the
irreplaceable Grab ‘n’ Go’s, and was confused that there were not the right
food options provided. Melody completed the task, but was extremely
unsatisfied. It was not until task 2 when the photo links broke and Melody had
to refresh to start the process all over again that she realized her protein
error. When she discovered what had occurred, Melody was angry and said
she would have left the site a long time ago.
Alignment to Heuristic: Consistency and Standards
This heuristic means that users should not have to wonder whether different
words, situations, or actions mean the same thing. The conventions that were set when
websites were first being designed are what developers and designers should follow so
that there is no confusion. When the user first starts developing their meal plan, they
have to select the things they are allergic to indicate that they want those kinds of meals
excluded. However, the next prompt of, “I like meals that include” it switches to
deselecting the orange highlighted icons. This change is confusing to users, and Chef’d
should have consistency throughout their website. The deselecting of protein also goes
against standards of ecommerce, where users are used to clicking for things that they
want.
15. Green15
Task 2: You have the budget to cover one extra meal per week, but you want it to be a
spicy meal. Determine your spicy meal options and choose one meal that you are
willing to prepare.
Summary for Both Testers:
Tester1 Tester2 Average
Average
Satisfaction 4 3 3.5
SuccessRate 0% 0% 0%
Highlights
1. Testers were unable to complete the task because there was no spicy
meal available.
Harrison and Melody both instinctively clicked on the photo of the meals to be
able to see more information; however, it took Melody longer to be able to see
the spiciness header on the recipe card. The problem simply was that the
desired content was not available, and that created annoyance for both users.
Melody and Harrison both had the expectation that the product would be
available, and it was disappointing to fail the task. Melody was further
frustrated because she had to look through meals that she could not even eat
because she had not discovered her protein error yet.
2. Testers both had links break on them that made them refresh the page.
When Harrison was clicking through all the photos to see the recipe card, he
clicked on the Spaghetti Carbonara, which never loaded. He waited for about
20 seconds until he gave up and tried to continue the task. However, the rest
16. Green16
of the photographs’ links did not work either. Harrison had to refresh the page
and start the entire meal plan over again, and he expressed his annoyance
with the website. The same scenario happened to Melody, though at this
point she at least realized how to select the proper protein for herself.
This problem of broken links and having to refresh is more technical, but it
was large factor for both of them in their satisfaction rates.
3. It look a long time to click through each meal and Grab ‘n’ Go’s to see if
the food was spicy
Including the links breaking, this process look several minutes for Harrison
and Melody. There was no other way to see if the food was spicy other than
to go through and click each photograph. The load time was not long, but the
second each took to load the preview quickly added up.
Biggest Problem
The most prominent issue was highlight three, having to take so much time to go
through all the meals’ recipe cards. The testers tried the provided entrée first, and then
had to click swap meal to go through all the other options’ recipe cards. They both were
feeling impatient, and it contributed to their disappointment to discover that they spent
all that time to realize there was not even an available meal. Melody commented that
she would have left the website by now, and Harrison said that he wanted to just go and
look at the other meals on the site to accomplish the task.
Alignment to Heuristic: Flexibility and Efficiency of Use
17. Green17
This heuristic is about accelerators, such as filters and sorting, that help speed up the
interaction for the expert user while still being easy to use for inexperienced users. It
also allows users to tailor frequent actions. Chef’d offers filters for their All Meals tab,
and one of them is the spiciness, but not for the Spoon University meal plan. This task
required many repetitive actions that having an accelerator may have prevented.
Task 3: Ensure none of your meals, “Grab ‘n’ Go’s” or snacks is more than 450 calories
per serving. If the calorie count is too high, swap the item for another.
Summary for Both Testers:
For a more in-depth success rate, it could also be written as 77% successful.
This is because Harrison and Melody both found seven out of nine of the calorie counts.
Highlights
1) The nutrition facts were difficult to find
Harrison and Melody were able to find the spiciness easily because it one of
the headers
on the
recipe card
when they
clicked on
the photograph link. The nutrition facts, however, are in gray font and users
Tester1 Tester2 Average
Average
Satisfaction 5 2 3.5
SuccessRate 0% 0% 0%
18. Green18
need to click again to be able to see it. Both of the testers did not see the tab
right away and were confused.
2) Testers had to click through each item again to check the nutrition
count
Similar to task two, Harrison and Melody had to take several minutes to once
again go through and check each recipe card. The Four Cheese Mac n’
Cheese link was broken for both of them, and they had to refresh and go
through the meal plan process again.
3) The fruits and healthy snack links did not provide any nutritional
information
Once the testers were at the bottom of the page, they clicked on the images
of fruits and healthy snacks. Instead of the recipe card with information
popping up, Chef’d provided these alerts. Without the calorie information,
Harrison and Melody were unable to finish the task.
Biggest Problem:
19. Green19
The biggest problem was the testers not receiving the nutritional information for the fruit
and healthy snacks. Both testers started with clicking their one entrée, and then working
their way down through their Grab ‘n’ Go’s. The Four Cheese Mac ‘n Cheese link did
not work, which both testers said that they were sure that it would not be under 450
calories. Once they reached the fruit and healthy snack photographs, there was
immediate confusion when those links did not pull up the same things that the other
seven had provided. Harrison commented that he was uncomfortable that he did not get
to know what the healthy snacks were that he planned to buy. Melody was also
frustrated, and stated that she would just have to assume that the snacks were under
the task amount and contained no meat.
Alignment to Heuristic: Consistency and Standards
As mentioned above, the consistency and standards heuristic means that users should
not have to wonder whether different words, situations, or actions mean the same thing.
Follow platform conventions. Chef’d consistently provided the recipe card for the first
seven options on the page, and then for the last two did not. Users should not have to
wonder whether the action of clicking on the photograph link would take them to their
desired information that they are expecting to see. It is a standard that when you
provide a product the user will be able to see more information, and buying healthy
snacks that the user cannot see goes against that.
Task 4: Assume/pretend that it is Monday at 3:30 p.m., and you plan to prepare the
spicy meal on Saturday night. Will the meal arrive in time?
20. Green20
Summary for both testers:
This summary has two contentions: one that the testers are working in the
hypothetical situation that they were able to select a spicy meal. Two, because the
current date of Chef’d cannot be set in the future or past, that the testers are wanting
their food in five days.
Highlights
1) Testers unsure where the delivery dates were located.
Testers did not know where to look for the delivery dates. When they had
done their previous tasks, they did not click “add to cart,” which would bring
them to where they would enter their zip code to see three available dates.
2) Tester had difficulty with the checkout closing out on when clicked off.
When given the timeframe, Harrison wanted to look
at this calendar on his mac while he was looking at
the checkout dates. However, each time he tried to
click back onto the Chef’d website it would close out
his checkout popup.
Tester1 Tester2 Average
Average
Satisfaction 4 4 4
SuccessRate 100% 100% 100%
21. Green21
3) Slow delivery unknown to users until they discovered the checkout
Both Melody and Harrison are used to quick shipping through Amazon Prime,
so they were unhappy to see that they would have to wait at least five days to
get their food. Chef’d advertises that they deliver the kits, but they do not
openly provide the timeframe until the user finally gets to checkout.
On the home page they have advertisements like “need dinner tonight” and
“Eat What You Want When You Want” that give the impression that Chef’d
provides almost instantaneous delivery.
Biggest Problem
The biggest problem was task one, because neither Harrison nor Melody
went immediately to their checkout to discover the delivery dates. Harrison
first went to the navigation and started going back through the tabs to look for
delivery information. After doing that with no success, Harrison noticed his
cart in the right, top corner showed the icon with a 10, representing his
products. He clicked on that and from there clicked on checkout and
continued his task.
Melody first clicked on “How It Works” to try and find the delivery dates, but
did not see any kind of that information. From there she realized that she had
22. Green22
not yet clicked “add to cart” at the bottom of her meal plan. Then the checkout
popup came up on the right side and Melody continued through to the next
page where it offers delivery dates. It is worth mentioning though that they
may have been hesitant to checkout because this was just a usability test.
Harrison and Melody said they assumed that the next step would be credit
card information, which they did not want to provide.
Alignment to Heuristic: Error Prevention
The error prevention heuristic is about careful design that prevents a problem
from occurring in the first place. Either eliminate error-prone conditions or check for
them and prevent users with a confirmation option before they commit to the action. If
part of the meal plan questionnaire was setting the desired delivery date for their kits,
this problem would be avoided. If getting their food within five days was extremely
important for whether they decided to order these products that would have been
information they would have wanted to know right away.
Final Site Thoughts
Harrison and Melody both had great initial impressions of the website’s design
and layout. They also enjoyed the content in the beginning and mentioned they would
consider ordering from a meal kit delivery service in the future. The testers both said
that they first thought it was going to be easy to design their meal kit with how cleanly
designed the website was.
Their feedback changed, however, as they talked about their experiences with
actually going through the Spoon University process. Both agreed that they were
23. Green23
frustrated with the design of certain aspects like the protein deselection, location of the
prices, and how long the tasks took to complete. Melody was the most upset, and again
stated that she would have never ordered from this site after her experience. Harrison
fared better and was overall happy with how the process went for him, despite some
annoyance. In contrast, he stated that he would actually consider ordering from Chef’d
in the future, though likely not through Spoon University.
The testers both voiced a decent amount of frustration with the website’s layout,
and the most positive feedback was with the user interface.
RECOMMENDATIONS TO IMPROVE USER EXPERIENCE
Single problem being fixed:
The single problem to be fixed is from task one of building a Spoon University
Meal Plan for under 65-dollars. It is highlight two, “testers tried to select protein options
instead of deselect.”
Problem Improvement:
The protein selection needs to stay consistent with the rest of the meal plan
survey, and to meet common website standards. First, I recommend rewriting the
24. Green24
header of that section to something more clear. “I like meals that include” is real world
language, but could be more straightforward for error prevention. Instead the header
could read, “Select Proteins” with the icons being in gray just like the allergy icons
above. When the users are interacting with the icons, the protein options should be
something selected instead of deselected.
With these changes, the users know exactly what the question is asking, and
they can follow platform conventions and select the products that they want.
I also recommend preventing any other errors by providing filters for the Spoon
University Meal
Plan, so while the
user is looking
through their meal
options they can
easily filter in case
they accidentally
did make a mistake
while selecting their
25. Green25
protein. This keeps the user from having to go back a few steps, and their meal amount
and choices are not lost. This adjustment could also help with the other tasks that
required users to go through and click each recipe card. I would place this change
below the “Inside Your Box” header and paragraph, right above the meals. With there
being fewer meals under the Spoon University Meal Plan I would not use all nine filters
that Chef’d uses for its All Meals Tab. The college meals are simplistic, quick, and
already in a category, so I would leave out the Category, Cooking Time, and Skill Level
filters.
Overall, the user will clearly know what they are selecting, and if there is a
change, they can easily adjust their preferences. If this had been the design of the
website for my usability tests, Harrison and Melody would have had a more delightful
experience. It gives the user more than one way to avoid or solve problems, and
provides more consistency throughout the website.