Susan drove an average speed of 30 mph for the first 30 minutes of a trip. She did drove 60 mph for the next 30 minutes. If she made no stops, what was the average speed, in miles per hour, for the entire trip?
a. 35
b. 40
c. 45
d. 50
e. 55
Could someone please explain why my logic is flawed?
If the driver drives at an avg rate of x miles per hour and the driver only drives half of 1 hour shouldn't the distance driven be 1/2 of what it would be if a driver drove for one full hour?
first 30 minutes
She drove at avg speed for 30 mph, so if she drove for 1hr then she would driven 30 miles
but she only actually drove for 30 minutes or so would she only have driven 15 miles since she drove have of an hour
Second 30 minutes
She drove at avg speed for 60 mph, so if she drove for 1hr then she would driven 60 miles
but she only actually drove for 30 minutes or so wouldn't she only have driven 30 miles since she drove have of an hour
So Total distance for 1 hours=
15 miles (half of 30 miles since she drove for 30 min)+30 miles (half of 60 miles since she drove for 30 min) = 45 miles
So avg rate is D/Time or 45miles/1 hour
Susan drove an average speed of 30 mph for the first 30 min
This topic has expert replies
- GMATGuruNY
- GMAT Instructor
- Posts: 15539
- Joined: Tue May 25, 2010 12:04 pm
- Location: New York, NY
- Thanked: 13060 times
- Followed by:1906 members
- GMAT Score:790
Please note the words in red above, which reflect the wording of the original problem.jbsocal wrote:Susan drove an average speed of 30 miles per hour for the first 30 miles of a trip and then at an average speed of 60 miles per hour for the remaining 30 miles of the trip. If she made no stops during the trip, what was Susan's average speed, in miles per hour, for the entire trip?
a. 35
b. 40
c. 45
d. 50
e. 55
Time to travel the first 30 miles at a speed of 30mph = d/r = 30/30 = 1 hour.
Time to travel the remaining 30 miles at a speed of 60mph = d/r = 30/60 = .5 hours.
Average speed for the entire trip = (total distance)/(total time) = (30 + 30)/(1 + .5) = 60/1.5 = 40.
The correct answer is B.
Last edited by GMATGuruNY on Tue Mar 19, 2013 9:01 pm, edited 1 time in total.
Private tutor exclusively for the GMAT and GRE, with over 20 years of experience.
Followed here and elsewhere by over 1900 test-takers.
I have worked with students based in the US, Australia, Taiwan, China, Tajikistan, Kuwait, Saudi Arabia -- a long list of countries.
My students have been admitted to HBS, CBS, Tuck, Yale, Stern, Fuqua -- a long list of top programs.
As a tutor, I don't simply teach you how I would approach problems.
I unlock the best way for YOU to solve problems.
For more information, please email me (Mitch Hunt) at [email protected].
Student Review #1
Student Review #2
Student Review #3
Followed here and elsewhere by over 1900 test-takers.
I have worked with students based in the US, Australia, Taiwan, China, Tajikistan, Kuwait, Saudi Arabia -- a long list of countries.
My students have been admitted to HBS, CBS, Tuck, Yale, Stern, Fuqua -- a long list of top programs.
As a tutor, I don't simply teach you how I would approach problems.
I unlock the best way for YOU to solve problems.
For more information, please email me (Mitch Hunt) at [email protected].
Student Review #1
Student Review #2
Student Review #3
GMAT/MBA Expert
- Anju@Gurome
- GMAT Instructor
- Posts: 511
- Joined: Wed Aug 11, 2010 9:47 am
- Location: Delhi, India
- Thanked: 344 times
- Followed by:86 members
As per the question you've posted, your logic is flawless.jbsocal wrote:Could someone please explain why my logic is flawed?
However, I feel the actual question is as follows...
Total distance covered = (30 + 30) miles = 60 milesSusan drove an average speed of 30 mph for the first 30 miles of a trip. She did drove 60 mph for the next 30 miles. If she made no stops, what was the average speed, in miles per hour, for the entire trip?
Total time = (Time taken to cover 30 miles at 30 miles per hour) + (Time taken to cover 30 miles at 60 miles per hour) = 1 hour + 1/2 hour = 3/2 hours
Hence, average speed = (Total distance)/(Total time) = 60/(3/2) mph = 40 mph
The correct answer is B.
Anju Agarwal
Quant Expert, Gurome
Backup Methods : General guide on plugging, estimation etc.
Wavy Curve Method : Solving complex inequalities in a matter of seconds.
§ GMAT with Gurome § Admissions with Gurome § Career Advising with Gurome §
Quant Expert, Gurome
Backup Methods : General guide on plugging, estimation etc.
Wavy Curve Method : Solving complex inequalities in a matter of seconds.
§ GMAT with Gurome § Admissions with Gurome § Career Advising with Gurome §
GMATGuruNY wrote:Please note the words in red above, which reflect the wording of the original problem.jbsocal wrote:Susan drove an average speed of 30 mph for the first 30 miles of a trip. She did drove 60 mph for the next 30 miles. If she made no stops, what was the average speed, in miles per hour, for the entire trip?
a. 35
b. 40
c. 45
d. 50
e. 55
Time to travel the first 30 miles at a speed of 30mph = d/r = 30/30 = 1 hour.
Time to travel the next 30 miles at a speed of 60mph = d/r = 30/60 = .5 hours.
Average speed for the entire trip = (total distance)/(total time) = (30 + 30)/(1 + .5) = 60/1.5 = 40.
The correct answer is B.
my apologies for the typo, this clears up why B is correct.
- hemant_rajput
- Master | Next Rank: 500 Posts
- Posts: 447
- Joined: Sun Apr 22, 2012 7:13 am
- Thanked: 46 times
- Followed by:13 members
- GMAT Score:700
Your method is accurate.jbsocal wrote:Susan drove an average speed of 30 mph for the first 30 minutes of a trip. She did drove 60 mph for the next 30 minutes. If she made no stops, what was the average speed, in miles per hour, for the entire trip?
a. 35
b. 40
c. 45
d. 50
e. 55
Could someone please explain why my logic is flawed?
If the driver drives at an avg rate of x miles per hour and the driver only drives half of 1 hour shouldn't the distance driven be 1/2 of what it would be if a driver drove for one full hour?
first 30 minutes
She drove at avg speed for 30 mph, so if she drove for 1hr then she would driven 30 miles
but she only actually drove for 30 minutes or so would she only have driven 15 miles since she drove have of an hour
Second 30 minutes
She drove at avg speed for 60 mph, so if she drove for 1hr then she would driven 60 miles
but she only actually drove for 30 minutes or so wouldn't she only have driven 30 miles since she drove have of an hour
So Total distance for 1 hours=
15 miles (half of 30 miles since she drove for 30 min)+30 miles (half of 60 miles since she drove for 30 min) = 45 miles
So avg rate is D/Time or 45miles/1 hour
Average speed = Total distance / total time taken.
The flaw is already pointed out by GMATGuruNY.
I'm no expert, just trying to work on my skills. If I've made any mistakes please bear with me.
-
- Junior | Next Rank: 30 Posts
- Posts: 27
- Joined: Sat Feb 16, 2013 2:27 am
- Location: India
- Thanked: 1 times
Here is my take,
If distance is constant
then
Avg speed = 2xy/ (x+y)
x= 30 mph
y= 60 mph
So Avg speed =40 (B)
Thanks,
Neil
If distance is constant
then
Avg speed = 2xy/ (x+y)
x= 30 mph
y= 60 mph
So Avg speed =40 (B)
Thanks,
Neil
GMAT/MBA Expert
- Scott@TargetTestPrep
- GMAT Instructor
- Posts: 7289
- Joined: Sat Apr 25, 2015 10:56 am
- Location: Los Angeles, CA
- Thanked: 43 times
- Followed by:29 members
We are given that Susan drove at an average speed of 30 miles per hour for the first 30 miles of a trip and then at an average speed of 60 miles per hour for the remaining 30 miles of the trip. We must determine her average speed overall. The formula for average speed is:jbsocal wrote:Susan drove an average speed of 30 mph for the first 30 minutes of a trip. She did drove 60 mph for the next 30 minutes. If she made no stops, what was the average speed, in miles per hour, for the entire trip?
a. 35
b. 40
c. 45
d. 50
e. 55
average speed = total distance/total time
For the first half of the trip, we know that Susan's speed was 30 mph and her distance was 30 miles, so her time was 30/30 = 1 hour.
For the second half of the trip, we know that Susan's speed was 60 mph and her distance was 30 miles, so her time was 30/60 = 1/2 hour. Therefore her average speed in miles per hour is:
average speed = (30 + 30)/(1 + ½)
average speed = 60/(3/2)
average speed = 40
Answer: B
Scott Woodbury-Stewart
Founder and CEO
[email protected]
See why Target Test Prep is rated 5 out of 5 stars on BEAT the GMAT. Read our reviews
GMAT/MBA Expert
- Brent@GMATPrepNow
- GMAT Instructor
- Posts: 16207
- Joined: Mon Dec 08, 2008 6:26 pm
- Location: Vancouver, BC
- Thanked: 5254 times
- Followed by:1268 members
- GMAT Score:770
Average speed = (TOTAL distance)/(TOTAL time)jbsocal wrote:Susan drove an average speed of 30 mph for the first 30 minutes of a trip. She did drove 60 mph for the next 30 minutes. If she made no stops, what was the average speed, in miles per hour, for the entire trip?
a. 35
b. 40
c. 45
d. 50
e. 55
TOTAL distance
Susan traveled 30 miles at a speed of 30mph, and 30 miles at a speed of 60mph.
So, TOTAL distance = 60 miles
TOTAL time
TOTAL time = (time spent driving 30 mph) + (time spent driving 60 mph)
time = distance/speed
- Time spent driving 30 mph = 30 miles/30mph = 1 hour
- Time spent driving 60 mph = 30 miles/60mph = 0.5 hours
TOTAL time = (1 hour) + (0.5 hours) = 1.5 hours
----------------------------------
So, Average speed = (TOTAL distance)/(TOTAL time)
= 60 miles/1.5 hours
= 40 mph
= B
Cheers,
Brent