8/14/2019 Ciolli g Arber
1/41
CIOLLI-GARBER 10/15/098:31 PM
132
HOW TO RANK LAW SCHOOL RANKINGS
ANTHONY CIOLLI*
JOSHUA GARBER*
This article discusses the various systems used to rank law schools and methods for
evaluating their validity. The authors conclude that outcome-based rankings are likely
to offer the most utility to prospective students.
I. THE MOST COMMON RANKINGS SYSTEMS ..............................................134A. Reputation Rankings.....................................................................134
1. Single Individual Perception .................................................1352. Faculty Perception..................................................................1353. Attorney/Judge Perception.....................................................1364. Student Perception or Revealed Preference..........................137B. Institutional Characteristics Rankings .........................................137
C. Outcomes Rankings: Employment Placement and Beyond .......138D. Comprehensive Rankings: U.S. News and its Progeny..............139
II. HOW DOES ONE EVALUATE A RANKING SYSTEM? ...............................141A. The Reactionary Response to Law School Rankings .................141B. The Case for Evaluation: Why the Reactionaries Miss the
Point ............................................................................................145C. Developing an Evaluation Paradigm ...........................................146
1. Purpose....................................................................................1462. Audience .................................................................................1473. Methodology...........................................................................148
i. Are measurements used meaningful or arbitrary? ..........148ii. How much error is due to experimenter failings,
rather than inherent error?.............................................148iii. Can others duplicate the findings?.................................151
III. COMMON FAULTS WITH EXISTING LAW SCHOOL RANKINGS..............152A. Disconnect Between Purpose and Methodology ........................152
1. U.S. News and other comprehensive rankings.....................1522. Single Individual Perception Reputation Rankings .............155
* Anthony Ciolli is Appellate Law Clerk to Chief Justice Rhys S. Hodge, Supreme Court of the
Virgin Islands. The opinions in this article are the authors alone and do not reflect the views of
Chief Justice Hodge, the Supreme Court of the Virgin Islands, or the Virgin Islands judiciary.* Joshua R. Garber is a Policy Fellow at the American Civil Liberties Union of NorthernCalifornias Death Penalty Program. The opinions in this article are the authors alone and do not
reflect the views of the American Civil Liberties Union, its members, donors, or supporters.
8/14/2019 Ciolli g Arber
2/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 133
B. Disconnect Between Audience and Purpose ...............................1571. Faculty Perception: Evaluating the Leiter Report ................1572. Why Revealed Preference Reveals Little .............................161
C. Poor Methodology & Avoidable Error ........................................162IV. BEYOND LAW SCHOOL RANKINGS: A PROPOSAL FOR CHANGE .........169Both academic journals and the popular media have published a
multitude of articles related to the impact of higher education rankings. In
fact, the Indiana Law Journalrecently devoted an entire symposium issue
to law school rankings.1 The overwhelming majority of these writings have
painted a very unflattering picture of all higher education rankings,
especially those published by U.S. News & World Report. The Law School
Admission Council has stated that law school rankings have resulted in
many law schools misusing the LSAT in the admissions process. 2 Several
authors, most prominently Daria Roithmayr, argue that this misuse of the
LSAT results in discrimination against minorities, particularly blacks and
Hispanics.3 These criticisms intensified in April 2005, when U.S. News
altered its rankings methodology by averaging a law schools 25 th and 75th
percentile LSAT and GPA figures rather than using self-reported medians.4
Shortly after U.S. News announced this change, the magazine was criticized
for hampering affirmative action efforts,5 and within a year reverted to the
previous system.6 Rankings have also been blamed for a host of other
phenomena, from causing law schools to raise tuition7 to reinforcing an
1 Symposium, The Next Generation of Law School Rankings, 80 IND. L.J. (2006).2 See LAW SCH. ADMISSION COUNCIL, NEW MODELS TO ASSURE DIVERSITY, FAIRNESS,
AND APPROPRIATE TEST USE IN LAW SCH. ADMISSIONS 4, 7, 10 (1999)
3 See Daria Roithmayr, Barriers to Entry: A Market Lock-In Model of Discrimination, 86VA. L. REV. 727, 731-36 (2000) ([L]aw schools have had to adopt the industry standard that
favors whites. schools that want to maintain their national ranking or place graduates in
lucrative positions must admit students based on their [LSAT] scores)4 See U.S. News, Law Methodology (2005), available at
http://www.usnews.com/usnews/edu/grad/rankings/about/06law_meth_brief.php. 5 If only a schools median LSAT is used for rankings purposes, the 25th and 75th percentile
LSAT scores are irrelevant. By definition, half a law schools incoming class will always have an
LSAT score at or below the median; thus, when only the median is used, it does not matter for
rankings purposes if a school with a 170 LSAT median has a 168 or a 162 as its 25 th percentile.
Since blacks and Hispanics as a group have lower LSAT scores than whites and Asians, using
medians as opposed to interquartiles gave law schools the flexibility to admit minorities with
significantly below median LSAT scores without impacting their U.S. News rank. With this
change, some argue that law schools will feel significant pressure to raise their 25th percentile
figures, and thus would have to scale back their affirmative action programs. See
http://www.collegejournal.com/successwork/workplacediversity/20050421-bialik.html?refresh=on.
6 Theodore P. Seto, Understanding the U.S. News Law School Rankings, LOYOLA LAW
SCHOOL LEGAL STUDIES PAPER No. 2007-25, 7 (2007).7 Alvin J.Esau, Competition, Cooperation, or Cartel: A National Law School Accreditation
Process for Canada?, 23 DALHOUSIE L.J. 183, 185 (2000) (Partly as a result of the recent
8/14/2019 Ciolli g Arber
3/41
CIOLLI-GARBER 10/15/09 8:31 PM
134 THE DARTMOUTH LAW JOURNAL Vol. VII:2
educational caste system.8
Unfortunately, many of these articles fail to acknowledge that the
genie is already out of the bottle. Regardless of ones attitude towards U.S.
News and other rankings, there is little doubt that commercial and non-commercial law school rankings systems are here to stay. This Article will
identify the four primary ranking archetypes, advocate for a standardized
method of evaluating a ranking systems validity, and evaluate the most
popular existing law school ranking schemes using this standardized
method. We conclude by proposing outcome-based rankings as a way of
evaluating law schools that will result in the greatest utility for prospective
students and will not suffer from the same deficiencies as U.S. News and its
competitors, provided that the authors have not engaged in avoidable error.
I. THE MOST COMMON RANKINGS SYSTEMS
Law school rankings come in a wide variety of forms, but one caneasily classify most ranking schemes into one of four distinct categories:
reputation rankings, institutional characteristics rankings, outcomes
rankings, and comprehensive rankings. In this section, I will briefly
describe each of these rankings archetypes as well as their common sub-
categories.
A. Reputation Rankings
Reputation rankings are likely the most popular type of rankings
system in terms of sheer numbers. This is not very surprising, since
creating a ranking system based solely on reputation is often relatively
easy. After all, reputation is a proxy for how a law school is perceived by
others any individual can create a reputation ranking scheme simply by
surveying himself or others. Because reputation rankings are so prevalent,
one can further break them down into four sub-categories: single individual
perception, faculty perception, attorney/judge perception, and student
perception.
surveys and rankings of law schools in Canada, there appears to be increased competition
between law schools to stay on top or move up some kind of ladder of prestige. There is increasedcompetition between schools to recruit students, charge them higher fees, get more money from
alumni, institute new programs, chairs, and institutes, and brag about job placements for
graduates.)8 David C. Yamada, Same Old, Same Old: Law School Rankings and the Affirmation of
Hierarchy, 31 SUFFOLK U. L. REV. 249, 261-62 (1997).
8/14/2019 Ciolli g Arber
4/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 135
1. Single Individual Perception
Single individual perception rankings are the most common of the
four reputation ranking schemes, if only because they require less time andeffort to create than any other ranking scheme. As the name implies, single
individual perception rankings represent the viewpoints of one individual.
This individual simply creates an ordinal list of law schools arranged by the
individuals perception of each schools reputation or quality. While the
individual might list some factors that played a role in shaping his rankings,
a precise methodology is never disclosed.
Most single individual perception rankings are not widely
disseminated. Because of their questionable methodologies, these rankings
are often relegated to threads on internet message-boards and blogs.9
Perhaps the most well known single individual perception ranking is the
Gourman Report, first self-published by Jack Gourman in 1967 and then
backed by the Princeton Review in 1997. While early editions of theGourman Report focused on overall university quality, Gourman began to
rank law schools in the 1980 edition.10 Gourman claims that his rankings
are based on multiple measures of quality, such as administration policies
and facilities.11 However, neither Gourman nor the Princeton Review have
been willing to reveal a methodology,12 leading some to speculate that there
is no methodology other than Gourmans personal opinion.13
2. Faculty Perception
Faculty perception rankings are a type of peer-based ranking system in
which an individual attempts to measure the quality or reputation of a
schools faculty by surveying faculty members at various institutions. This
individual then aggregates the survey responses in order to generate a
ranking.
9 See, e.g., http://www.leiterrankings.com/jobs/2008job_biglaw.shtml10 Scott van Alstyne, Ranking the Law Schools: The Reality of Illusion?, AMERICAN BAR
FOUNDATION RESEARCH JOURNAL, VOL. 7, NO. 3 at 656.11 Other qualities Gourman claims to take into account are the relationship between
professors and administrators, support of faculty members, cooperation among professors,
methods of communication between professors and the administration, the openness of the
administration, the use of consultants and committees to solve problems, attitudes about scholarly
research, and the overall cooperation of the administration.
http://chronicle.com/free/v44/i11/11a00101.htm
12 For decades, Gourman has refused to elaborate on his criteria or how he was able toquantify an institutions quality. Id. Evan Schnittman, editor-in-chief of the Princeton Review,
stated that he only knows Gourmans criteria in general terms. Id.13 Alexander W. Astin, director of the Higher Education Research Institute at the University
of California at Los Angeles, stated that An intelligent U.S. senior in high school could sit down
and make up rankings that are similar to Gourman.Id.
8/14/2019 Ciolli g Arber
5/41
CIOLLI-GARBER 10/15/09 8:31 PM
136 THE DARTMOUTH LAW JOURNAL Vol. VII:2
Faculty perception rankings are rather common in other disciplines.
Perhaps the most well known faculty perception ranking is Brian Leiters
The Philosophical Gourmet, a bi-annual ranking of philosophy
departments.14
However, law school faculty perception rankings have beenrelatively rare. The best-known attempt at ranking law schools based on the
perceptions of law school faculty are Leiters educational quality
rankings.15 Modeled after his philosophy department rankings, Leiters law
school educational quality rankings seek to rank law schools based on the
strength of their faculty.16
Unlike Gourman, Leiter discloses the methodology used to generate
his rankings. Leiter asked more than 150 law professors to complete his
survey and evaluated faculty lists on a scale of one to five. 17 After gathering
this data, Leiter aggregated the results and ranked the faculties based on
their mean score.18
3. Attorney/Judge Perception
Rankings by attorneys and judges are similar to faculty perception
rankings in that an individual is not merely stating his or her own personal
opinion of an institutions quality, but aggregating the opinions of many
other individuals in this case, lawyers or judges. Unlike a faculty
perception ranking, an attorney/judge perception ranking is not a peer-
based ranking system.19
Although practitioners frequently serve as adjuncts or legal writing
instructors at many law schools, the overwhelming majority of attorneys
and judges cease to have a formal affiliation with a law school after
obtaining their law degrees. Therefore, unlike faculty perception rankings,
which almost always involve law faculty rating the reputation of other law
faculties, attorney/judge perception rankings can differ in what they
measure. While some attorney/judge perception rankings may measure
academic reputation, others may ask respondents to measure student
employability, or perceptions of student quality, or how graduates perform
on the job.
One set of attorney perception rankings was conducted by Angela
14 Rankings available athttp://www.philosophicalgourmet.com/overall.asp15 Leiters EQR rankings available athttp://www.leiterrankings.com/faculty/
2003faculty_reputation.shtml16 Id.17 Id.18 Id.19 Law firm rankings, such as the annual Vault reputation rankings, are an example of a peer-
based reputation ranking system by attorneys.
8/14/2019 Ciolli g Arber
6/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 137
Cheng and published in the National Law Journalin November 2004.20
The National Law Journalsent a survey to 250 hiring partners, which,
among other things, asked these attorneys to list schools from which they
hired first year associates in the previous year.21
As Leiter did for hisfaculty perception rankings, Cheng aggregated the results and reported
them in her article.
4. Student Perception or Revealed Preference
Student perception rankings also commonly referred to as revealed
preference rankings are a recent but growing trend in higher education
rankings. This trend began in October 2002, when Christopher Avery,
Mark Glickman, Caroline Hoxby, and Andrew Metrick released the first
draft of their paper A Revealed Preference Ranking of U.S. Colleges and
Universities.22 This paper, later revised in September 2004 using a new
methodology, sought to rank undergraduate programs based on studentpreferences or desirability.23 In November 2004, Aaron J. Chalfin released
a similar paper, which ranked law schools based on student desirability.24
While there have been significant methodological differences between
the major revealed preference studies, both have involved surveying
current or prospective students, with the researchers aggregating this
survey data to create an ordinal ranking.
B. Institutional Characteristics Rankings
Institutional characteristics rankings are, for some, an attractive
alternative to reputation rankings. While reputation rankings focus on how
institutions are perceived by others, institutional characteristics rankings
examine the characteristics inherent to the institutions. There are countless
ways individuals could rank law schools based on their characteristics.
Perhaps the most common method involves ranking schools based on a
tangible measure of student quality, such as LSAT medians or interquartile
ranges. Other possibilities include ranking law schools based on
student:faculty ratio, the number of volumes in the law library, the
20 Angela Cheng, Georgetown, Virginia Among Most Mentioned, THE NATIONAL LAW
JOURNAL(2004),available athttp://www.law.georgetown.edu/news/releases/
documents/nlj_000.pdf.
21 Id.22 Avery, Christopher, Glickman, Mark E., Hoxby, Caroline M. and Metrick, Andrew,A
Revealed Preference Ranking of U.S. Colleges and Universities (December 2005). NBER
Working Paper No. W10803. Available at SSRN: http://ssrn.com/abstract=60110523 Id.24 http://www.autoadmit.com/studies/chalfin/chalfin.pdf
8/14/2019 Ciolli g Arber
7/41
8/14/2019 Ciolli g Arber
8/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 139
reputation rankings. However, unlike institutional characteristics, it is
difficult to verify outcomes with any sort of absolute certainty. There is
greater ambiguity involved in outcomes rankings, since one must make
value judgments that are not present when trying to calculate LSATmedians or determine the number of books in a library. For instance, where
does one draw the line when calculating judicial clerkship placement
success? Does one count state family court clerks the same as U.S.
Supreme Court clerks? While two different individuals calculating the
same institutional characteristics will obtain the same numbers, individuals
ranking law schools by student outcomes will likely obtain different results
depending on how they deal with these questions.
While most outcomes rankings focus on employment placement, they
are not limited to that area. Other outcomes rankings can include student
debt after graduation, the percentage of students who go on to obtain
LL.Ms or other graduate degrees after graduation, and a host of other
possibilities.
D. Comprehensive Rankings: U.S. News and its Progeny
One cannot discuss law school rankings without acknowledging the
annual U.S. News & World Report rankings. The U.S. News rankings,
although heavily maligned by some academics,26 are without a doubt
todays most influential law school rankings. In fact, the U.S. News
rankings are so influential that there is little doubt many law schools shape
admissions policies with these rankings in mind.27
How did U.S. News, a biweekly magazine, often considered a tier
below Time andNewsweek, become the dominant force in higher education
rankings? Some have argued that U.S. News provides a service to
prospective law students by providing a host of information about
individual law schools that is not available anywhere else.28 While this is
one of the strongest arguments in support of the U.S. News rankings, it does
not fully explain how the U.S. News rankings were able to rise to
prominence. After all, U.S. News not only did not create the first set of law
school rankings, but it was also not the first magazine to rank law schools.29
26 See, e.g., The Law School Admissions Council, Deans Speak Out(2006), available at,
http://www.lsac.org/Choosing/deans-speak-out-rankings.asp27 Leigh Jones,Law School Deans Feel Heat From Rankings, NATIONAL LAW JOURNAL
(2006), available athttp://www.law.com/jsp/article.jsp?id=1146560723820
28 Russell Korobkin,Harnessing the Positive Power of Rankings: A Response to Posner andSunstein, 81 IND. L.J. 35, 37 (2006).
29 This honor goes to Learning & the Law, a magazine that published the first set of law
school rankings 1975 in an article called Adding up the Law Schools: A Tabulation and Rating
of Their Resources by Charles D. Kelso. See
http://taxprof.typepad.com/taxprof_blog/2006/10/the_first_law_s.html,
8/14/2019 Ciolli g Arber
9/41
CIOLLI-GARBER 10/15/09 8:31 PM
140 THE DARTMOUTH LAW JOURNAL Vol. VII:2
U.S. News established itself as the dominant ranking system through
innovation. While other rankings providers limited themselves to ranking
law schools based on one measure of quality, U.S. News moved beyond
using a single criterion to rank law schools. While the first edition of theU.S. News law school rankings in 1987 did consist only of a faculty
perception reputation ranking, the second edition, released in 1990,
involved a drastic methodology change. These 1990 rankings, and all
subsequent versions, created an overall score by combining a faculty
perception ranking, an attorney/judge perception ranking, a selectivity
ranking, a placement success ranking, a graduation rate ranking, and an
instructional resources ranking.30 While the methodology has changed
several times since then, since 1990 all versions of the U.S. News rankings
have involved some aggregation of multiple reputation and institutional
characteristics rankings.
Thus, U.S. News can receive credit for popularizing comprehensive
rankings. Rather than merely ranking law schools based on reputation orone institutional characteristic, law schools are ranked through a weighted
average of multiple reputation and institutional characteristic rankings.
Comprehensive rankings like U.S. News generally have a key advantage
over other rankings systems when it comes to obtaining the acceptance of
the general public. U.S. News and other comprehensive rankings tend to
convey a sense of scientific rigor. By weighing factor X at 40%, factor Y at
25%, factor Z at 1%, and so on, a comprehensive rankings provider
conveys the impression of statistical certainty, even if the factors were
arbitrarily chosen and the weights lack any defensible empirical or
theoretical basis.31
Such weighting also results in fewer dramatic rises and falls relative to
other rankings. For example, since U.S. News began using a comprehensive
ranking system in 1990, the same fourteen schools Harvard, Yale,
Stanford, Chicago, Columbia, NYU, Penn, Berkeley, Michigan, Duke,
Northwestern, Virginia, Cornell, and Georgetown have held the top
fourteen positions in the rankings.32 In addition, since 1990, each of these
schools, in what is commonly referred to as the Top 14,33 has been
ranked in the U.S. News top ten at least once.34 While there are fluctuations
within the Top 14 from year to year, it is a remarkably stable group. For
instance, the U.S. News top five has always been some combination of
http://www.elsblog.org/the_empirical_legal_studi/2006/10/the_first_ranki.html
30 U.S. News & World Report, Best Graduate Schools (March 9, 1990).31 http://www.washingtonmonthly.com/features/2000/norc.html32 Leiter, Brian. From the Bowels of Cyberspace: The Myth of the the Top 14 (March 19,
2006). Available at: http://leiterlawschool.typepad.com/leiter/2006/03/from_the_bowels.html.33 Id.34 Id.
8/14/2019 Ciolli g Arber
10/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 141
Harvard, Yale, Stanford, Chicago, Columbia, and NYU (commonly called
the HYSCCN group), and 1991, 1992, 2007, and 2008 were the only
years when this group did not make up the entire top six.35 Other ranking
systems have not displayed this type of consistency. Thus, comprehensiverankings like U.S. News are, in general, able to provide year-to-year
fluctuations while at the same time providing enough general stability to
prevent the rankings from appearing that they lack efficacy.
Furthermore, comprehensive rankings inherently split the difference
between other ranking systems. U.S. News, by combining reputation scores
with LSAT medians and employment figures, is able to establish itself as a
consensus ranking system. Even though few individuals are truly
satisfied with the U.S. News methodology or results, so many individuals
provide tepid or lukewarm support for U.S. News that its rankings become
the standard.
Other comprehensive rankings have arisen in the wake ofU.S. News,
but have failed to achieve nearly the same amount of attention oracceptance. Perhaps the most notable of these ranking schemes is Judging
the Law Schools,36 an annual comprehensive ranking by Thomas Brennan,
founder and former President of the Thomas M. Cooley Law School, and
Don LeDec, the current President and Dean of Cooley.37 This ranking
system, which heavily weighs factors such as law school square-footage
and library seating capacity,38 has often been dubbed the Cooley Law
School Rankings,39 and the weighting scheme has come under heavy
criticism for blatantly favoring Cooley40 for instance, the 2008 edition
ranks Cooley higher than Stanford, Penn, Berkeley, Chicago, Cornell and
Duke.41
II. HOW DOES ONE EVALUATE A RANKING SYSTEM?
A. The Reactionary Response to Law School Rankings
While much has been written about how higher education rankings
impact educational institutions by shaping admissions policies and
35 In 1991 and 1992, Michigan displaced NYU for the sixth position. In 2007, Penn tied
Chicago for the sixth position, and in 2008 Berkeley was ranked as the sixth in the country.36 Rankings available athttp://www.cooley.edu/rankings/overall2008.htm37 http://www.cooley.edu/rankings/intro_10th_general.htm
38 Id.39
Leiter,Brian.TheCooleyLawSchoolRankings.(Oct.2005).http://leiterlawschool.typepad.com/leiter/
2005/10/the_cooley_law_.html40 http://taxprof.typepad.com/taxprof_blog/2009/01/size-matters-.html41 Rankings available athttp://www.cooley.edu/rankings/overall2008.htm
8/14/2019 Ciolli g Arber
11/41
CIOLLI-GARBER 10/15/09 8:31 PM
142 THE DARTMOUTH LAW JOURNAL Vol. VII:2
affecting diversity, even more has been written about the utility of law
school rankings. U.S. News and other ranking systems have been heavily
maligned by both academic and non-academic writers. Although some
writers have defended law school rankings,42
the overwhelming majority ofarticles about law school rankings have been negative. Individuals ranging
from university presidents43 to law professors44 to high school seniors45
have extensively criticized rankings for being, among other things,
arbitrary or otherwise useless. Most of these individuals could be best
described as reactionaries, because they not only attack U.S. News and its
methodology, but argue that law school rankings are inherently flawed and
maintain that society would be better off giving these rankings little
credence.
The most publicized reactionary attack against law school rankings
was organized by the American Association of Law Schools in 1998. On
February 18, 1998, a week before the expected release of the 1998 U.S.
News law school rankings, the AALS held a press conference where tendeans and educators, led by John Sexton, then-dean of the New York
University School of Law, urged U.S. News to stop publishing its annual
law school rankings.46 At the press conference and the accompanying press
release, Sexton stated that the rankings were misleading and dangerous,
and argued that a ranking system inherently assumes that every applicant
has identical needs and desires.47 The AALS press conference coincided
with the release of an AALS-commissioned U.S. News validity study48 and
a letter denouncing law school rankings sent to 93,000 law school
applicants by the Law School Admission Council.49 The letter, signed by
164 law school deans, urged applicants to gather their own information
about law schools:
42 See generally Russell Korobkin, In Praise of Law School Rankings: Solutions to
Coordination and Collective Action Problems, 77 TEX. L. REV. 403 (1998).43 http://www.stanford.edu/dept/pres-provost/president/speeches/961206gcfallow.html44 See, e.g., David A. Thomas, The Law School Rankings are Harmful Deceptions: A
Response to Those Who Praise the Rankings and Suggestions for a Better Approach, 40 HOUS.
L.REV. 419, 426-427 (2003) (The inference a reader would draw from the U.S. News rankings
is that the items listed in the rankings are the accepted criteria one should use in determining the
quality of a law school. Not only is that deliberately created impression false and misleading,
but almost none of the U.S. News criteria can possibly be relevant to or probative of any
reasonable or common sense notion of law school quality.).45 See Robert Samuels, Students Call College Rankings into Question, Fox News, August 27,
2003. http://www.foxnews.com/story/0,2933,95772,00.html
46 Press release available at http://www.aals.org/reports/validity.html47 Id.48 See Stephen P. Klein & Laura Hamilton, The Validity of The U.S. News and World Report
Ranking of ABA Law Schools (1998), athttp://www.aals.org/validity.html.49 See The Law School Admission Council, Deans Speak Out(2006), at
http://www.lsac.org/LSAC.asp?url=lsac/deans-speak-out-rankings.asp.
8/14/2019 Ciolli g Arber
12/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 143
The idea that all law schools can be measured by the same yardstick
ignores the qualities that make you and law schools unique, and is
unworthy of being an important influence on the choice you are about to
make. As the deans of law schools that range across the spectrum of severalrating systems, we strongly urge you to minimize the influence of rankings
on your own judgment. In choosing the best school for you, we urge you to
get information about all the schools in which you might have some
interest.50
Although it is difficult to disagree with the letters underlying
principles, the deans argument rings hollow. While the deans urge
prospective students to get information about all the schools in which you
might have some interest, both law schools and the AALS have
historically sought to minimize the dissemination of such information. Paul
Caron and Rafael Gely summarize the AALS impact on 20 th century legal
education quite well:
The AALS. . . removed competitive elements from the legal educationmarket and allowed law schools to behave in a monopolistic manner. Law
schools are different from the typical profit-maximizing firm (which seeks
monopoly power by controlling price and excluding rivals) and the typical
nonprofit firm (which seeks output maximization by distributing its
bounty as widely and as equitably as possible). Instead, law schools seek
to maximize elitist preferences. These preferences include. . . freedom
of law schools to teach the best law students, freedom from faculty
accountability to non-peer groups (such as students), and freedom to
operate in a non-commercial atmosphere.
By eliminating. . . competition among law schools, the AALS
eliminated incentives to measure the success of a law school. To the extent
that all law schools offered the same courses, taught by faculty with similar
credentials, using the same methods, it became more difficult to identify
winners and losers. This elimination of measures of organizational
success was consistent with the elitist preferences of AALS members.
More established schools had little to fear from allowing lesser schools to
exist - as long as they were not too different (i.e., they operated within rigid
AALS standards). Other law schools not only enjoyed the protection
afforded by the cartel-like activities of the AALS, but also were in a
position to claim equality with their more established counterparts.
Standardization permitted every law school to claim to be as good as
whatever conventional wisdom suggested were the leading law schools in
the country.The absence of measures of organizational success also eliminated
incentives to measure individual contributions. The ability of law schools
50 Id.
8/14/2019 Ciolli g Arber
13/41
CIOLLI-GARBER 10/15/09 8:31 PM
144 THE DARTMOUTH LAW JOURNAL Vol. VII:2
to pursue elitist preferences, and hence enjoy freedom from accountability,
allowed them to pursue various objectives that they would have been
unable to seek in a world in which contributions were measured. For
example, law faculty could more easily make hiring decisions based not onteaching effectiveness and scholarly productivity but instead on other
grounds such as race, gender, or educational pedigree.
In short, the AALS presided over a law school world without
objective standards and accountability. By standardizing the industry
product, the AALS eliminated any measures of organizational success and
any incentives to measure individual contributions.51
This AALS stranglehold did not break until U.S. News and other law
school rankings provided institutions with a very strong incentive to cease
operating in a non-competitive environment. U.S. News rankings, by
creating a rat race among law schools that otherwise would not exist,
forced those law schools to reveal the sort of information that they
otherwise would want to hide from applicants.52 Law school rankings, intheir function as information forcing devices, are socially beneficial in that
they result in more efficient matching outcomes between applicants and
schools (and employers).53
David A. Thomas provides a strong counter-argument. Thomas points
out that even those who defend law school rankings implicitly [concede]
that the contents of the rankings have little to do with actually revealing the
quality of a law school, either absolutely or comparatively.54 While some
previously hidden information may become available to the general public,
ordinal rankings that include dubious qualities such as reputation may
result in net social harm rather than social benefit. Thomas argues that if
rankings readers are actually deceived into believing they are learning
about the quality of a law school, such deception may ultimately deter
interested persons from more meaningful investigations.55 Furthermore,
since the American Bar Associations Official Guide to Law Schools is
available for free on the internet56 and contains much of the same statistical
information as U.S. News, one can argue that law school rankings have
already served their purpose as an information forcing device and
eliminating them would not result in any ill effects.
51 Paul L. Caron & Rafael Gely. What Law Schools Can Learn from Billy Beane and the
Oakland Athletics, 82 TEX. L. REV. 1483, 1507-08 (2004).
52 See Scott Baker, Stephen J.Choi, & G. Mitu Gulati. The Rat Race as an InformationForcing Device, 80 IND. L.J. (forthcoming 2005).
53 See id.; see also Korobkin, supra note 53.54 Thomas, supra note 55, at 425.55 Id.56 http://officialguide.lsac.org/
8/14/2019 Ciolli g Arber
14/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 145
B. The Case for Evaluation: Why the Reactionaries Miss the Point
While Thomas makes valid points, his analysis fails because he does
not properly distinguish U.S. News rankings from other law schoolrankings.57 Though his criticisms of both U.S. News and reputation
rankings are justified to an extent, it is inappropriate to attack law school
rankings as a whole due to the perceived failings of one (albeit popular) set
of rankings. Thomas provides little support for this strongly-worded
generalization: Without question, the world would be better off without
those rankings. They benefit no person or entity other than their
publisher.58
Thomass position is akin to no one should rank law schools because
no ranking can possibly be perfect. Because GPA and LSAT are not
perfect predictors of intelligence or law school performance, Thomas
believes they should not be used in a rankings scheme.59
Thomas is correct that LSAT, GPA, and student:faculty ratio are notcompletely free of error. However, this does not make them useless
measures. After all, what in the world is completely free of any error
whatsoever? Let us apply Thomass high standard beyond just law school
rankings. If one were to take Thomass position to its natural conclusion,
one would have to advocate for the elimination of any field of study whose
research does not result in absolutely error-free results, including
economics, political science, psychology, sociology, and the medical and
natural sciences. While error is present in all of these fields, the presence of
error does not automatically mean we should discount all of its scholarship
Researchers can still reach perfectly valid conclusions even in the presence
of reasonable error.
Although it is possible that the world would be better off without one
poorly written or shoddy piece of economics research, arguing the entire
field of economics should be is indefensible and akin to throwing the baby
out with the bathwater. Even if U.S. News rankings are horribly flawed,
57 But see Thomas, supra note 55, at 426 (If the intent of a ranking is merely to display
objective data in numerical order, leaving interpretation and assessment to readers, then the only
challenges one could raise would be to the accuracy or validity of the data.). Although Thomas
does acknowledge this difference between U.S. News and other rankings when critiquing the U.S.
News methodology, he fails to make the same distinction in his later analysis, and paints all law
school rankings in a negative light. See infra note 69.58 Thomas, supra note 55, at 456.59 See Thomas, supra note 55, at 431 (Implicit is the assumption that good students -
those who are a positive influence on their fellow students - can be identified solely or principallyby their GPA or LSAT score. More important, but far less subject to assessment, are whether a
law schools students are emotionally stable, free of substance abuse or other self-defeating
behaviors, honest, kind and compassionate, and balanced in a professional environment that is
often intensely driven and beset by materialism. These factors cannot be measured by reported
statistics or by rankings.).
8/14/2019 Ciolli g Arber
15/41
CIOLLI-GARBER 10/15/09 8:31 PM
146 THE DARTMOUTH LAW JOURNAL Vol. VII:2
arguing that all formal law school rankings result in a net negative effect,
regardless of the rankings methodology or purpose, is an extreme position
that is very difficult, if not impossible, to defend. Not all rankings are
inherently flawed or arbitrary. For example, institutional characteristicrankings are, by definition, based on objective and verifiable facts. One
could very easily generate an ordinal ranking of law schools based on the
number of hours per week the law library is open, and if a prospective
student places an extremely high value on being able to study in the library,
it would be both rational and efficient to use such a ranking to decide
where to interview. Of course, such a ranking might lack utility if
prospective students place significantly higher values on other factors,
which they often do, but students can consider these issues when evaluating
law school rankings.
Keep in mind that in the absence of formal rankings, stakeholders in
legal education will simply rank law schools informally. Prospective law
students, rather than looking at U.S. News or other rankings, will createtheir own personal set of law school rankings based on whatever data they
can informally obtain, potentially containing even greater flaws than U.S.
News.60 If law school rankings did not exist, legal employers would still
have to evaluate law school quality when making hiring decisions, even if
they had little or no knowledge of particular law schools.61
C. Developing an Evaluation Paradigm
Since there are situations where rankings can be highly useful, as well
as situations where they can be useless or even destructive, a blanket
dismissal of all rankings provides little utility to anyone. Rather than saying
all current rankings are flawed, one should develop a standardized method
for evaluating the efficacy of ranking schemes. When evaluating any
ranking, one should pay careful attention to three factors: purpose,
audience, and methodology. I will now discuss each of these factors.
1. Purpose
How does this ranking matter? What does this ranking measure?
60 See Jan Hoffman, Judge Not, Law Schools Demand Of a Magazine That Ranks Them
(1998), available at
http://query.nytimes.com/gst/fullpage.html?res=9806E4DD133FF93AA25751C0A96E958260&sec=&spon=&pagewanted=2 where the Dean of New York University School of Law states that if
reputation rankings asked about Princeton Law School that Princeton, which does not have a law
school, would nevertheless appear in the top twenty.61 Michael Berger, Why U.S.NewsRankings are Useful and Important, 51 J. LEGAL EDUC.
487, 492 (2001).
8/14/2019 Ciolli g Arber
16/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 147
These are the questions one should ask when evaluating the purpose of any
ranking.
Obviously, a law school ranking scheme intends to rank law schools,
but what is the rankings intended purpose? Does the ranker intend tomeasure employment placement, selectivity, or faculty attitudes towards
vanilla ice cream?
One might think this is a very straightforward question, but even after
twenty-two years in existence, there is still no consensus about what U.S.
News is trying to measure. Some authors have assumed that the purpose of
the U.S. News rankings is to rate the quality of legal education.62 Others
state that U.S. Newss primary purpose is to rate schools based on how they
are perceived in the real world by employers. U.S. News itself does not
seem to know the purpose of its own rankings. While on one page U.S.
News says that the purpose of its graduate rankings is to provide an
independent assessment of the academic quality of programs,63 another
page states that one should use the U.S. News rankings to determine howdiplomas from various schools will affect [applicants] earning power and
how successful the schools are at preparing graduates for the bar exam. 64
Unless we are supposed to believe that U.S. News rank perfectly correlates
with academic quality, earning power, and bar exam preparation,65 it
appears that U.S. News provides its readers with contradictory purposes.
2. Audience
Who gains the most utility from these rankings? This is the question
to ask when trying to determine a rankings audience. While most rankings
tend to target prospective students, the audience could be any demographic
group, from faculty to law librarians to the general public.
Audience is as crucial to the evaluation of law school rankings as it is
for evaluating legal writing. A 70 page paper might make a splash as a law
review article but that same paper would make a horrible letter to the editor
of a local newspaper. An ordinal ranking of law schools based by
admissions officer salaries may have a lot of utility for current and
prospective law school admissions officers; however, this same ordinal
ranking would be virtually worthless to prospective students. Like good
writers, effective law school ranking providers will tailor their rankings
towards the appropriate audience.
62 See, e.g., Thomas, supra note 55, at 419.63 http://www.usnews.com/usnews/edu/grad/rankings/about/faq_meth_brief.php#one.64 http://www.usnews.com/usnews/edu/grad/rankings/about/06rank.b_brief.php.65 Even a very quick comparison ofU.S. News rankings and U.S. Newss raw data shows that
this is not the case.
8/14/2019 Ciolli g Arber
17/41
CIOLLI-GARBER 10/15/09 8:31 PM
148 THE DARTMOUTH LAW JOURNAL Vol. VII:2
3. Methodology
An effective methodology will allow a law school ranking scheme to
serve its intended purpose for its intended audience. At a minimum, therankings methodology should conform to standard social science research
methods. It is not possible to provide a concise summary of an entire
research methods textbook in one article.66 However, the most important
research methods concepts can be applied to law school rankings as part of
a simple three part test:
i. Are measurements used meaningful or arbitrary?
It should be self-evident that the tool one uses to measure something
should actually measure what one says it measures. Obviously, one does
not use a ruler to measure ones weight, and one does not use a
thermometer to measure the volume of a partially filled milk carton. Usingthose tools as measuring devices in those contexts would bring about
laughable results.
The same holds true for law school rankings. When ranking law
schools, the measurements used should further the intended purpose. For
example, if ones purpose is to rank law schools based on student:faculty
ratio, ones methodology should not involve distributing a survey asking
faculty for their favorite ice cream flavor. While measuring faculty ice
cream preferences might be appropriate in other contexts,67 the results
would tell us absolutely nothing about school student:faculty ratios,
because the measurement device used is completely arbitrary. For a set of
rankings to have efficacy, the measuring device must have some relation to
what one intends to measure. If one intends to measure student:faculty
ratios, one should collect data on the total number of students and the total
number of faculty at every law school; one should not use data on faculty
ice cream preferences unless one actually wants to measure faculty ice
cream preferences.
ii. How much error is due to experimenter failings, rather than
inherent error?
It is extremely difficult to imagine any type of social science research
that does not result in any error whatsoever. Whether measuring public
66 For a more in-depth explanation of social science research methods, see generally (?)67 For example, Ben & Jerrys may be opening up an ice cream store in the faculty lounge of
every law school in America and would like to know what flavors different faculty prefer so that
they know approximately how much of a certain flavor to keep in stock at each location.
8/14/2019 Ciolli g Arber
18/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 149
opinion towards same-sex marriage or developing an empirical model to
predict a Supreme Court justices vote, error will always exist. However,
all error is not equal. When evaluating any piece of research, one must
distinguish between avoidable and unavoidable error.Unavoidable error is the error that is inherent in any type of empirical
research. As much as an experimenter may want a completely error free
research project, ideal conditions are almost never possible. If a researcher
is examining how peoples weekly church attendance changes over a 20
year period, its a virtual certainty that some individuals will withdraw
from the study before the 20 year period ends, perhaps because of death or
incapacitation, or because they simply do not want to participate anymore.
Such unavoidable error is present in many law school ranking
schemes. Any law school ranking system that uses median LSAT as a
factor must deal with the reality that not every individual attending every
law school has taken the LSAT. Northwestern University, for example, has
a three year joint J.D./M.B.A. degree program that requires the GMAT butnot the LSAT for admission to both the business school and the law
school.68 As much as 10% of Northwesterns incoming first year J.D. class
comes from this program.69 Georgetown University has a sub-
matriculation program where Georgetown undergraduates can apply to
Georgetown University Law Center during their junior year based on their
undergraduate GPA and their SAT/ACT score without submitting an LSAT
score.70 In contrast, most law schools, such as the University of
Pennsylvania Law School, require LSAT scores from every applicant and
will not admit any student to the J.D. program without them, not even
through sub-matriculation or other joint degree programs.71 Thus, one can
say that there is a response rate differential among law schools when it
comes to incoming student LSAT scores. While schools such as the
University of Pennsylvania and Harvard may have 100% reporting rates,
schools like Northwestern may have 90% reporting rates. An individual
constructing a law school ranking system that makes use of LSAT medians
must work with the data available to him or her. Since LSAT scores simply
do not exist for some incoming Northwestern and Georgetown law students
because they were admitted without taking the LSAT, the ranker must deal
with this limitation, perhaps by assuming that the theoretical LSAT scores
of the non-responsive group would not differ significantly from the LSAT
68 http://www.jdmba.northwestern.edu/faq.htm.
69 Northwesterns class of 2011 has 26 J.D./M.B.A. students out of a total class size of 242,meaning about 10.7% of the class consists of J.D./M.B.A. students. See
http://www.jdmba.northwestern.edu/classprofile.htm,
http://www.law.northwestern.edu/admissions/profile/70 http://www.law.georgetown.edu/Admissions/jd_general.html#earlyassurance.71 http://www.law.upenn.edu/prospective/jd/faq.html#jointdegreeapp.
8/14/2019 Ciolli g Arber
19/41
CIOLLI-GARBER 10/15/09 8:31 PM
150 THE DARTMOUTH LAW JOURNAL Vol. VII:2
scores of the rest of the class, or realizing that the number of students
without LSAT scores at these schools is so low that even if their scores
would differ from the rest of the cohort the medians would not be
impacted. Alternatively, the researcher could disclose that this problemexists and explain the biases it introduces, but argue that since LSAT scores
are the best standardized measure of student quality available the benefits
of including LSAT score as a rankings variable outweigh the costs. While
the researcher has several options available to him or her, fixing the error
itself is not one of them.72
Avoidable error, in contrast, involves error that does not have to exist
in an experiment or research project. While unavoidable error is inherent to
the project, avoidable error is completely preventable. Avoidable error is
only present due to the researchers negligence or inexperience.
Avoidable error can manifest itself in many ways, but it seems
especially common in studies that involve survey research. Poor sampling
has the potential to make an otherwise valid study virtually useless. Forexample, if a researcher wanted to examine how the American people feel
about same-sex marriage, he or she would introduce avoidable error if he or
she only included eighteen year old college freshmen at Berkeley in the
sample. Because eighteen year old Berkeley freshmen are not an accurate
reflection of the demographic makeup of the United States, such a survey
would reveal nothing about how Americans in general feel about the issue.
A similar survey about abortion given only to Catholic priests would also
reveal nothing about the general publics attitudes towards the issue. In
both cases, the error is avoidable because the researcher could have given
the survey to a sample that properly reflects the demographic makeup of
the United States.
Law school rankings, too, may have avoidable error due to poor
sampling techniques. U.S. News in particular has been justifiably criticized
for the poor construction of its practitioner reputation survey.73 However,
other types of avoidable error may exist, because a ranker may fail to make
necessary adjustments. For instance, when ranking schools based on
student:faculty ratio, an individual would introduce avoidable error by only
using full-time J.D. enrollment in the calculation. Such a ranking would
72 Of course, if this specific researcher had access to unlimited time and funds, he or she
could obtain the names of every single student enrolled at Northwestern, Georgetown, etc. who
did not take the LSAT and offer to pay each of them a large sum of money to take the LSAT so
he or she could calculate the true median LSAT for each school. While such a scenario is
theoretically possible, in the real world researchers do not have unlimited resources, and variousprivacy laws exist that would make this scenario virtually impossible in practice.
73 See, e.g., Brian Leiter, The U.S. News Law School Rankings: A Guide for the Perplexed
(2003), available athttp://www.utexas.edu/law/faculty/bleiter/rankings/guide.html (criticizing the
U.S. News practitioner survey for having a very low response rate and a geographically biased
sample).
8/14/2019 Ciolli g Arber
20/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 151
ignore part-time and LL.M./S.J.D. students, and thus inflate the ranking of
schools with part-time or graduate programs. Other sources of avoidable
error include basic mistakes, such as inputting incorrect data or making a
mathematical error.Note that not all instances of avoidable error will automatically cause
a set of rankings to become invalid. When evaluating law school rankings
with avoidable error, one must consider the impact that error has on the
rankings themselves. Let us consider one example: a law professor ranks
every law school based on median LSAT score. On April Fools Day, her
research assistant sneaks into her office, opens the professors Excel file
that contains all the rankings, and subtracts five LSAT percentile points
from every law schools median; thus, a school with a 99.4th percentile
median now appears to have a 94.4 th percentile median, a school with a 77th
percentile median appears to have a 72nd percentile median, and so on. The
professor, absent-minded as she is, never notices the change, and uploads
the ordinal rankings to her website. Although there is significant avoidableerror, the ordinal rankings themselves have not been damaged by the error
the 1st school is still the 1st school, the 69th school is still the 69th school,
and the absolute difference in percentile points between any two schools
remains the same. While the underlying data was tampered with, the
damage done was evenly distributed among every law school, and thus the
rankings may still serve their intended purpose. However, if the research
assistant had altered seven random schools and did not make any changes
to the other schools in the dataset, then the avoidable error would certainly
cause the rankings to lose efficacy and become invalid.
iii. Can others duplicate the findings?
Repeatability is one of the hallmarks of social science research. It is
not enough for a researcher to have a valid research design and avoid as
much error as possible. Researchers, as a quality check measure, should
also disclose their methodology so that others may repeat their experiment
to prove its validity. Like social scientists, individuals creating law school
rankings should explain how they generated their rankings. Regardless of
how one feels towards U.S. News rankings, U.S. News has openly
explained its methodology,74 and has allowed others to make attempts to
reverse engineer them.75 While most rankings providers now willingly
share their methodologies, some, most notably the Gourman Report, do
74 See, e.g., Robert Morse, Changing Law School Ranking Formula, U.S. News. June 26,
2008. http://www.usnews.com/blogs/college-rankings-blog/2008/6/26/changing-the-law-school-
ranking-formula.html.75 See, e.g., http://www.deloggio.com/usnews/usnwrpl.htm (membership required).
8/14/2019 Ciolli g Arber
21/41
CIOLLI-GARBER 10/15/09 8:31 PM
152 THE DARTMOUTH LAW JOURNAL Vol. VII:2
not, and thus raise serious questions about their validity.76
III. COMMON FAULTS WITH EXISTING LAW SCHOOL RANKINGS
Having established a paradigm by which one can evaluate the validity
of law school rankings, it is now appropriate to apply this system to current
law school ranking systems in order to illustrate some of the most common
flaws.
A. Disconnect Between Purpose and Methodology
1. U.S. News and other comprehensive rankings
Much has been written about the flaws with comprehensive rankings,
although most criticisms have focused on U.S. News. Such a focus is notsurprising, given the enormous influence of the U.S. News rankings. Other
comprehensive rankings, such as the Cooley rankings, have been virtually
ignored by prospective law students and other legal education
stakeholders.77
There is little doubt that prospective law students are U.S. Newss
intended audience. However, the U.S. News rankings suffer from
significant purpose and methodology problems. As discussed earlier in this
paper,78 U.S. News does a poor job explaining what its rankings are
supposed to measure, claiming that its rankings are an independent
assessment of the academic quality of programs79 while also stating that
its rankings measure how diplomas from various schools will affect
[applicants] earning power and how successful the schools are at
preparing graduates for the bar exam.80
Ultimately, whether U.S. News intends to measure academic quality,
earning power, bar exam preparation, or some combination of all three is a
moot question. The U.S. News rankings do not pass the validity test on
methodological grounds, regardless of which purpose is the true one. While
many articles have been written about the problems with the U.S. News
76 Selengo, supra note 11.77 See Brian Leiter, How to Rank Law Schools, 81 INDIANA LAW JOURNAL 47-52
(2006).78 See Section II, purpose section.
79 Supra note 75; Best Graduate Schools Frequently Asked Questions About TheRankings, http://www.usnews.com/articles/education/best-graduateschools/2008/03/26/frequently
-asked-questionsrankings.html#2.80 Supra note 76; Best Graduate Schools Frequently Asked Questions About The
Rankings How To Use Our Lists Wisely, http://www.usnews.com/articles/education/best-
graduate-schools/2009/04/22/how-to-use-our-lists-wisely-2010.html.
8/14/2019 Ciolli g Arber
22/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 153
methodology,81 Brian Leiter concisely summarizes the fatal flaw of U.S.
News and all other comprehensive rankings: [T]hose elements worth
measuring should be measured separately rather than aggregated on the
basis of unprincipled and unrationalizable schema. One can rank schoolsbased on SSRN downloads, student LSAT scores, faculty reputation,
scholarly impact as measured by citations, job placement, Supreme Court
clerkships, and so on, but there is no way these criteria can be meaningfully
amalgamated.82
Put aside for a moment the numerous flaws with how U.S. News
measures academic reputation,83 practitioner reputation,84 student
selectivity,85 faculty resources,86 and employment placement.87 In fact, let
81 See, e.g., http://www.washingtonmonthly.com/features/2000/norc.html,
http://www.leiterrankings.com/usnews/2000usnews_compare.shtml,
http://www.deloggio.com/data/tiers.html,
http://leiterlawschool.typepad.com/leiter/files/how_to_rank_law_schools.doc.82 Leiter, infra note 90.83 Academic reputation is measured by surveying law school deans, deans of academic
affairs, the chair of faculty appointments, and the most recently tenured faculty members of law
schools. About 70% respond to this survey and this composes a schools Peer Assessment
Score, which accounts for 25% of a schools ranking. Accordingly, it is the most significant
variable in determining a schools rank. However, this score may be heavily influenced by the
fact that law school deans and faculty members may hold allegiances to certain schools
particularly their alma matters and the schools their academic peers attended. Because the
majority of legal academics hail from the same schools and the score itself is highly subjective,
U.S. News may be causing a disservice by basing 25% of its ranking on this survey. See
http://www.usnews.com/articles/education/best-graduate-schools/2008/03/26/law-
methodology.html.84 U.S. News surveys legal professionals, including the hiring partners of law firms, state
attorneys general, and selected federal and state judges to find a schools practitioner reputation
score. While it is good to include this data in the rankings, a schools reputation ranking accounts
for 15% of its overall ranking, and it is questionable that this score should be given the weight.Basing 15% of a law schools overall ranking on a school reputation is especially questionable
because the bar passage rate is given only a weight of 2% of the overall score and employment at
graduation is onlygiven a weight of 4% of the overall score. Both of these factors are, arguably,
more important to an applicant than the practitioner reputation of a school. Additionally,
surveying hiring partners at law firms can make the survey biased to schools that place well in
New York City, where there are more firms than in any other city. Id.See also Brian Leiter, The
U.S. News Law School Rankings: A Guide for the Perplexed (2003), available at
http://www.leiterrankings.com/usnews/guide.shtml.
Measuring student selectivity largely by median undergraduate Grade Point Average (GPA) and
LSAT, as U.S. News does, is problematic for two reasons. First, the median LSAT of a school is
a poor measure of selectivity because 50.1% of a schools population can, for example, have an
LSAT score of 168 but a large portion of the school may nevertheless have an LSAT of 163 or
less. Second, measuring undergraduate GPA as a measure of selectivity ignores the fact that
some of the top students in the country hail from schools known for grade deflation (such as
California Institute of Technology or Swarthmore) or majored in a field of study known for aharsh grading curve. U.S. News claims that a school with a higher median undergraduate GPA or
LSAT should be ranked higher than a school a school with lower scores, but unless the quality of
the undergraduate institution attended/major studied and the LSAT scores from the rest of the
class are considered, the magazine does a poor job of measuring student selectivity. See
http://www.usatoday.com/news/education/2007-03-27-princeton-grades_N.htm, discussing grade
8/14/2019 Ciolli g Arber
23/41
CIOLLI-GARBER 10/15/09 8:31 PM
154 THE DARTMOUTH LAW JOURNAL Vol. VII:2
us imagine that U.S. News found a way to correct for all of these flaws, but
otherwise kept its methodology the same. What would these new and
improved U.S. News rankings tell us? Would they tell us about academic
quality? Not really if the purpose of the U.S. News rankings is tomeasure academic quality, why does their methodology include factors
such as bar passage rates, percent of the class employed nine months after
graduation, and financial aid expenditures per student? Perhaps the
rankings are meant to measure the earning power of graduates. But if this is
U.S. Newss true purpose, why include factors such as the number of titles
in the law schools library, or acceptance rate? Such extraneous factors do
nothing but create distortions in the rankings. There is absolutely no
empirical evidence to suggest that purchasing 10,000 library books will
have any impact on graduate earning power, and no reason to believe that
decreasing financial aid expenditures per student will decrease faculty
quality.
These additional factors create nothing but noise even if evidenceshows that there is a relationship between the two variables. For example,
academic reputation is known to strongly correlate with national
employment placement.88 However, why should a rankings scheme whose
purpose is to predict employment prospects or earning power include
academic reputation as an input if information on employment placement is
already available? While there is a correlation between the two variables,
the correlation is not perfect, and even if the correlation is very strong,
actual employment placement data will be by definition better than a
deflation at Princeton University, a top school that produces graduates with lower GPAs than
many schools below its caliber.86 The main problem with how U.S. News calculates faculty resource ranking is that at no
point does the magazine consider the strength of faculty in both publishing and teaching ability
(instead, it focuses on expenditures per student, student:faculty ratio, and library resources). No
points are assigned to faculties that are particularly scholarly, such as Chicagos. Also, there is
no place for the students to rank the quality of faculty/instruction at their school. Thus, although
a school may have a high faculty resource score, this in no way means that the faculty at the
school is better than the faculty at another school. See
http://www.usnews.com/articles/education/best-graduate-schools/2008/03/26/law-
methodology.html87 U.S. News values 4% of its overall ranking on employment at graduation and 14% of its
overall ranking on employment 9 months after graduation (and the magazine states that For the
nine-month employment rate, 25 percent of those whose status is unknown are counted as
employed. Students, in picking a law school, are hoping to be employed when they graduate.
At the very least, students hope to employed 6 months after graduation when they are required to
start making loan repayments. As a result, U.S. News is acting arbitrarily in placing 14% of itsoverall ranking on employment 9 months after graduation.
http://www.usnews.com/articles/education/best-graduate-schools/2008/03/26/law-
methodology.html88 See Anthony Ciolli, The Legal Employment Market: Determinants of Elite Firm Placement
and How Law Schools Stack Up, 45 JURIMETRICS J. 413 (2005).
8/14/2019 Ciolli g Arber
24/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 155
variable that closely correlates with employment placement.
The only way the current U.S. News methodology is valid is if the
purpose of the U.S. News rankings is also the methodology. That is, the
purpose ofU.S. News is to rank law schools by placing a 25% weight onacademic reputation and selectivity respectively, 20% weight on placement
success, and 15% weight on practitioner reputation as well as faculty
resources. If that is the actual purpose of the rankings, then the U.S. News
methodology makes perfect sense. However, this creates a new validity
problem: what utility do prospective law students, or any other audience,
gain from such a ranking? U.S. News has provided no justification as to
why these weights are necessary.89 There is also no evidence that even a
significant minority (let alone a majority) of prospective law students feel
these are appropriate weights. Given these facts, it is difficult to explain
how the U.S. News rankings can have any real utility for prospective law
students.
2. Single Individual Perception Reputation Rankings
U.S. News and other comprehensive rankings are often invalid due to
disconnect between purpose and methodology. Reputation rankings
generally suffer from similar problems, although different subsets of
reputation rankings also experience other validity issues.
The validity problems with single individual perception reputation
rankings closely parallel the problems with U.S. News. Depending on the
rankings stated purpose, the validity problem may be either methodology or
purpose. A single individual perception ranking will rarely experience a
methodology problem if the purpose of the ranking is to determine how a
specific individual views the law schools.90 But once again, even if such a
ranking is not methodological flawed, it lacks a useful purpose. After all,
why should a prospective law student care about how Jay Brody ranks the
law schools?91 Are there any tangible benefits to attending Brodys 4 th
ranked law school over his 17th ranked school?92
Not surprisingly, most single individual perception reputation
rankings claim that their rankings strongly correlate with some other factor.
89 Stephen P. Klein and Laura Hamilton, The Validity of the U.S. News and World Report
Ranking of ABA Law Schools (1998), available athttp://www.aals.org/reports/validity.html90 Perhaps the only way such a ranking could experience a methodology problem is if a third
party were asked to provide the rankings on behalf of the other person. For instance, if Justice
Scalia were asked to list Justice Breyers top ten law schools, he might provide a different listfrom the one Justice Breyer would provide.
91
http://web.archive.org/web/20030727215004/http://brody.com/law/resources/brody_law_school_r
ankings.php.92 Id.
8/14/2019 Ciolli g Arber
25/41
CIOLLI-GARBER 10/15/09 8:31 PM
156 THE DARTMOUTH LAW JOURNAL Vol. VII:2
Jay Brody does not claim that the purpose of his rankings are merely to tell
the world about his own perceptions about law school quality instead,
Brody says that prospective law students should pay attention to his
rankings because they represent how law schools are perceived in theemployment market.93 In other words, one is led to believe that the purpose
of the Brody rankings is to measure career placement.
Most single individual reputation rankings, therefore, run into a
similar methodology problem as with U.S. News and other comprehensive
rankings: if the purpose of the Brody rankings is to rank law schools based
on career placement, should not the Brody rankings actually try to
meaningfully measure career placement? Brody has made no visible
attempt to include actual career placement variables in his rankings
scheme, nor does he provide evidence that his individual perceptions about
law school quality correlate with a law schools ability to place its
graduates. When there is such an obvious disconnect between purpose and
methodology, the ranking scheme does not pass the validity test.Notably, there is a limited situation where a single individual
perception ranking can be both methodologically sound as well as useful to
the intended audience. This can occur when the individual ranking the law
schools is also in a position to make hiring decisions for specific positions
that the audience finds desirable. For example, let us assume that a
Supreme Court justice decided to rank the law schools, in order to let
prospective law students know how he perceives the quality of various
schools. Unlike Brody, there are many individuals who put great weight in
a Supreme Court justices view of law school quality. This is not just due to
the prestige of the justices position, but also because the justice has direct
control over a commodity that is highly desirable to many law students:
which individuals get a clerkship with that justice. Thus, a Supreme Court
justices single individual perception reputation rankings could have great
utility, for a prospective law student would know that attending a law
school ranked highly by the justice may enhance their chances of obtaining
a Supreme Court clerkship. Similar situations may occur if the hiring
partner of an elite law firm that many prospective law students aspire to,
such as Wachtell or Cravath, decided to rank the law schools. However, it
seems no such individuals or firms have decided to rank the law schools
based solely on their own perceptions.94
93 Id.94 Although Judge Richard Posner has written an article about law school rankings that
includes his own ranking scheme, this scheme is based on amalgamating other rankings, and does
not necessarily reflect how Judge Posner himself perceives law schools. See Posner, Richard,
"Law School Rankings," 81Indiana Law Journal13 (2006).
8/14/2019 Ciolli g Arber
26/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 157
B. Disconnect Between Audience and Purpose
1. Faculty Perception: Evaluating the Leiter Report
Brian Leiters Educational Quality Survey, the most well known law
school faculty perception rankings, does not suffer from the same
confusion of purpose as U.S. News Professor Leiter explicitly states that
the purpose of his survey is to measure the scholarly quality of law school
faculty,95 and he does not try to mislead readers into believing that his
rankings are meant to correlate with student employment prospects or any
other unrelated variable.96 Similarly, Professor Leiters methodology, while
certainly open to criticism,97 is still sound enough to serve the rankings
intended purpose; that is, Leiters survey actually attempts to measure how
faculty perceive other faculty.
The Leiter rankings, however, do suffer from a flaw, albeit a
somewhat less serious one. While the purpose and methodology may besound, Leiter considers prospective J.D. students his audience.98 Certainly,
some prospective J.D. students may have an interest in Leiters findings.
However, whether one likes it or not, faculty perception of the scholarly
quality of faculty at other law schools has little to no direct impact on
prospective law students. This is not to say that faculty quality is not
important to students however, faculty perception of the quality of other
faculties bears little utility for prospective students. Faculty perceptions of
the scholarly quality of other faculties do not correlate well with faculty
teaching quality. In fact, Leiters own teaching quality rankings confirm
that there is little to no relation between student evaluations of teaching
quality and faculty perception of scholarly quality.99
Leiter argues that his rankings may be the decisive factor for those
interested in pursuing careers in law teaching.100 While it is conceivable
that some prospective law students might already have an interest in legal
academia prior to enrolling in law school and might therefore take faculty
perception of scholarly quality very seriously, basing an enrollment
decision on the Leiter rankings may not be any wiser than basing an
95 http://www.leiterrankings.com/faculty/1999faculty_reputation.shtml.96 Id.97 For example, Richard G. Heck, Jr. argues that Leiters rankings of graduate schools should
not be based on a single factor, the quality of the facultys research, because the correlation
between this and the quality of a students graduate education is arguably small. Leiter may be
making the same mistake by ranking law schools by faculty research.http://frege.brown.edu/heck/philosophy/aboutpgr.php.
98 http://www.leiterrankings.com/faculty/1999faculty_reputation.shtml.99 http://www.leiterrankings.com/faculty/2003faculty_best.shtml.
100 http://www.utexas.edu/law/faculty/bleiter/rankings/students.html (last visited: January
2009).
8/14/2019 Ciolli g Arber
27/41
CIOLLI-GARBER 10/15/09 8:31 PM
158 THE DARTMOUTH LAW JOURNAL Vol. VII:2
enrollment decision on U.S. News rankings. In fact, deciding law schools
based on the Leiter faculty perception rankings may result in even bigger
enrollment mistakes.
In addition to the overall faculty perception rankings, Leiter alsoincludes faculty perception rankings for various specialty areas ranging
from administrative law to legal history.101 Leiter emphasizes the
importance of these specialty rankings to prospective law students on his
rankings webpage: If you have particular intellectual interests or particular
professional goals, make sure the schools youre considering will meet
them. In some cases, the rankings by faculty quality in the specialty areas
will provide useful information on this score. A school with a strong
faculty in, e.g., international and comparative law, will undoubtedly have
good offerings in those areas.102 While Leiter cautions that differences in
0.1 or less in the specialty rankings should not be given much weight,103 the
implication by omission is that differences higher than 0.1 should be
treated as significant differences.These faculty perception specialty rankings have little use even to
students who are seriously considering legal academia. For example, in the
faculty quality in law and philosophy specialty ranking, Cardozo Law
School is ranked 5th with a 4.0 score and the University of San Diego is
ranked 10th with a 3.6 score, while Harvard University is ranked 12 th with a
3.4 score, the University of Chicago is ranked 13th with a 3.3 score, and
Stanford University ranks 18th with a 3.1 score.104 For law and economics,
Georgetown University is ranked 19th with a 3.1 score while George Mason
University is ranked 9th with a 4.0 score.105 While these faculty perception
rankings may very well be accurate, they provide very little utility to
prospective law students, including prospective law students who know
with an absolute certainty that they intend to become legal academics in
those specialty fields. Once again, Leiter himself provides the data showing
the limitations of his faculty perception rankings. In his survey of where
tenure-track faculty hired since 1996 went to law school, Leiter finds that
108 Harvard, 104 Yale, and 48 Stanford graduates who had not earned an
LL.M. or other graduate degree secured tenure-track jobs.106 In contrast,
only one Cardozo and one George Mason graduate who had not earned an
additional graduate degree obtained tenure-track jobs, and the Cardozo
graduate works at her own alma mater.107 The University of San Diego
101 http://www.leiterrankings.com/faculty/index.shtml
102 http://www.leiterrankings.com/students/index.shtml103 http://www.leiterrankings.com/faculty/index.shtml104 http://www.utexas.edu/law/faculty/bleiter/rankings/philo.html105 http://www.utexas.edu/law/faculty/bleiter/rankings/economics.html106 http://www.leiterrankings.com/jobs/2006job_teaching.shtml107 Id.
8/14/2019 Ciolli g Arber
28/41
CIOLLI-GARBER 10/15/09 8:31 PM
Summer 2009 LAW SCHOOL RANKINGS 159
produced no graduates who obtained tenure-track jobs who did not earn an
additional graduate degree.108 While faculty members perceive some
second tier schools like Cardozo as having faculty in certain specialty areas
that are significantly better than faculty at top schools such as Harvard andYale, the data plainly shows that an individual seeking to enter legal
academia is far better off attending inferior J.D. programs at Harvard and
Yale.
Furthermore, even if the faculty perception rankings did not have such
flaws, the nature of the J.D. as a three year program pursued by future
lawyers and future academics alike makes prospective J.D. students an
inappropriate audience for this ranking. First, many individuals who apply
to J.D. programs thinking they intend to become law professors end up
foregoing the academy without even trying to obtain a tenure-track job.
The reasons for this can range from failings as students (finishing in the
bottom half of the class, not making law review, and so on) to realizing that
academia is simply not a good fit (they discover that they hate writingacademic articles, would prefer to earn more money in private practice, and
so forth). This sort of invisible attrition should not come as a surprise given
that, unlike Ph.D. programs, virtually all law students apply to law school
without ever setting foot in a law school class or knowing anything about
law as an academic discipline. Given how only a slim number of
individuals from any law school class will actually seek to become legal
academics, even among those who entered law school with that ambition, it
is irresponsible for J.D. students to make enrollment decisions based on
how their J.D. schools faculty is perceived by faculty at other law schools.
Second, and perhaps more importantly, faculty perception rankings
are extremely fluid. Even something as minor as one faculty member
switching schools, retiring, or dying can result in a significant shift in how
an entire schools faculty is perceived by others. Because faculty departures
and additions occur relatively frequently, faculty perception rankings can
change even more significantly than U.S. News rankings between the time
the prospective students makes the enrollment decision, graduates, and
enters the teaching market. While faculty departures will naturally have
little negative impact on schools such as Yale that are already perceived as
highly prestigious by non-academics,109 such departures, even if only
temporary, could have more significant effects on schools that are
traditionally viewed as having less prestige. Leiter himself illustrates this
by quoting a student on his blog: If you are going to Tufts to work with
Dennett and he is gone or not teaching in one or both of your firstsemesters, then it is not clear that Tufts distinguished faculty will be of
108 Id.109 http://leiterlawschool.typepad.com/leiter/2005/08/how_to_rank_law.html.
8/14/2019 Ciolli g Arber
29/41
CIOLLI-GARBER 10/15/09 8:31 PM
160 THE DARTMOUTH LAW JOURNAL Vol. VII:2
much help to you.110 While in reference to two year philosophy M.A.
programs, the same principle holds true for J.D. programs: the temporary or
permanent departure of one or two important faculty members can
eliminate any hypothetical advantage to attending a school that traditionallycarries less prestige but is perceived well by law school faculty.
Prospective J.D. students, then, have little use for a faculty perception
ranking, even those interested in becoming law professors. Professor
Leiters faculty perception rankings, however, could have considerable
utility for students who are contemplating enrollment in LL.M. or J.S.D.
programs with the intention of entering law teaching afterwards. An
audience of LL.M./J.S.D. applicants would gain more utility from such
rankings than prospective J.D. students for several reasons. First, full time
LL.M. and J.S.D. programs usually do not last longer than one academic
year, except in cases where an individual is pursuing both an LL.M. and a
J.S.D.111 Thus, except in cases of unexpected deaths; informed LL.M. and
J.S.D. applicants will know in advance whether specific faculty memberswill be available during their stay in the program. Second, since these
LL.M. and J.S.D. students have already experienced law school classes
during their J.D. programs and are pursuing their advanced degrees to
enhance their chances of obtaining academic jobs, how faculty perceive
their graduate schools truly becomes the decisive factor in the enrollment
decision, since the perceptions of other stakeholders, such as elite law
firms, are no longer relevant since future hiring decisions will be by faculty
members. Third, LL.M. and J.S.D. students will likely enter the market for
law teaching immediately after earning their graduate degrees. Thus, unlike
J.D. students, they do not need to worry about the prestige of their graduate
institution dropping in the eyes of faculty on hiring committees between
enrolling in the program and going on the market. Fourth, since many
LL.M. and all J.S.D. programs often involve significantly more one-on-one
interactions with faculty than J.D. programs, one could argue that the
prestige of LL.M. and especially J.S.D. students are more closely tied to the
prestige of their faculty advisors than J.D. students. Given these realities,
Leiter might consider increasing the validity of his rankings by targeting
prospective graduate law students rather than prospective J.D. students.
110 http://leiterreports.typepad.com/blog/2004/10/ma_programs_in_.html.111 For example, it would take at least 3 years to obtain a LL.M./J.S.D. in International Human
Rights from Saint Thomas University. http://www.stu.edu/Academics/Programs/
LLMJSDinInterculturalHumanRights/ProgramInformation/tabid/1227/JSDinInterculturalHuman
Rights/tabid/1237/Default.as