National Assessment of Educational Progress

Last updated
NAEP Logo NatlAssessmentOfEduProgress.jpg
NAEP Logo
The Nation's Report Card Logo TheNationsReportCard.png
The Nation's Report Card Logo

The National Assessment of Educational Progress (NAEP) is the largest continuing and nationally representative assessment of what U.S. students know and can do in various subjects. NAEP is a congressionally mandated project administered by the National Center for Education Statistics (NCES), within the Institute of Education Sciences (IES) of the United States Department of Education. The first national administration of NAEP occurred in 1969. [1] The National Assessment Governing Board (NAGB) is an independent, bipartisan board that sets policy for NAEP and is responsible for developing the framework and test specifications.The National Assessment Governing Board, whose members are appointed by the U.S. Secretary of Education, includes governors, state legislators, local and state school officials, educators, business representatives, and members of the general public. Congress created the 26-member Governing Board in 1988.

Contents

NAEP results are designed to provide group-level data on student achievement in various subjects, and are released as The Nation's Report Card. [2] There are no results for individual students, classrooms, or schools. NAEP reports results for different demographic groups, including gender, socioeconomic status, and race/ethnicity. Assessments are given most frequently in mathematics, reading, science and writing. Other subjects such as the arts, civics, economics, geography, technology and engineering literacy (TEL) and U.S. history are assessed periodically.

In addition to assessing student achievement in various subjects, NAEP also surveys students, teachers, and school administrators to help provide contextual information. Questions asking about participants' race or ethnicity, school attendance, and academic expectations help policy makers, researchers, and the general public better understand the assessment results.

Teachers, principals, parents, policymakers, and researchers all use NAEP results to assess student progress across the country and develop ways to improve education in the United States. NAEP has been providing data on student performance since 1969. [3] [4]

NAEP uses a sampling procedure that allows the assessment to be representative of the geographical, racial, ethnic, and socioeconomic diversity of the schools and students in the United States.[ citation needed ] Data is also provided on students with disabilities and English language learners. NAEP assessments are administered to participating students using the same test booklets and procedures, except accommodations for students with disabilities, [5] [6] so NAEP results are used for comparison of states and urban districts that participate in the assessment.

There are two NAEP websites: the NCES NAEP website and The Nation's Report Card website. The first site details the NAEP program holistically, while the second focuses primarily on the individual releases of data.

History

NAEP began in 1964, with a grant from the Carnegie Corporation to set up the Exploratory Committee for the Assessment of Progress in Education (ESCAPE). The first national assessments were held in 1969. Voluntary assessments for the states began in 1990 on a trial basis and in 1996 were made a permanent feature of NAEP to be administered every two years. In 2002, selected urban districts participated in the state-level assessments on a trial basis and continue as the Trial Urban District Assessment (TUDA).[need citation]

The development of a successful NAEP program has involved many, including researchers, state education officials, contractors, policymakers, students, and teachers. [7]

Assessments

There are two types of NAEP assessments, main NAEP and long-term trend NAEP. This separation makes it possible to meet two objectives:

  1. As educational priorities change, develop new assessment instruments that reflect current educational content and assessment methodology.
  2. Measure student progress over time.

Main

Main NAEP assessments are conducted in a range of subjects with fourth-, eighth- and twelfth-graders across the country. Assessments are given most frequently in mathematics, reading, science, and writing. Other subjects such as the arts, civics, economics, geography, technology and engineering literacy (TEL), and U.S. history are assessed periodically.

These assessments follow subject-area frameworks that are developed by the NAGB and use the latest advances in assessment methodology. [8] Under main NAEP, results are reported at the national level, and in some cases, the state and district levels.

National

National NAEP reports statistical information about student performance and factors related to educational performance for the nation and for specific demographic groups in the population (e.g., race/ethnicity, gender). It includes students from both public and nonpublic (private) schools and depending on the subject reports results for grades 4, 8, and 12.

State

State NAEP results are available in some subjects for grades 4 and 8. This allows participating states to monitor their own progress over time in mathematics, reading, science, and writing. They can then compare the knowledge and skills of their students with students in other states and with the nation.

The assessments given in the states are exactly the same as those given nationally. Traditionally, state NAEP was assessed only at grades 4 and 8. However, a 2009 [9] pilot program allowed 11 states (Arkansas, Connecticut, Florida, Idaho, Illinois, Iowa, Massachusetts, New Hampshire, New Jersey, South Dakota, and West Virginia) to receive scores at the twelfth-grade level.

Through 1988, NAEP reported only on the academic achievement of the nation as a whole and for demographic groups within the population. Congress passed legislation in 1988 authorizing a voluntary Trial State Assessment. Separate representative samples of students were selected from each state or jurisdiction that agreed to participate in state NAEP. Trial state assessments were conducted in 1990, 1992, and 1994. Beginning with the 1996 assessment, the authorizing statute no longer considered the state component a "trial.”

A significant change to state NAEP occurred in 2001 with the reauthorization of the Elementary and Secondary Education Act, also referred to as "No Child Left Behind" legislation. This legislation requires that states which receive Title I funding must participate in state NAEP assessments in mathematics and reading at grades 4 and 8 every two years. State participation in other subjects assessed by state NAEP (science and writing) remains voluntary.

Like all NAEP assessments, state NAEP does not provide individual scores for the students or schools assessed.

Trial Urban District Assessment

The Trial Urban District Assessment (TUDA) is a project developed to determine the feasibility of using NAEP to report on the performance of public school students at the district level. As authorized by congress, NAEP has administered the mathematics, reading, science, and writing assessments to samples of students in selected urban districts.

TUDA began with six urban districts in 2002, and has since expanded to 27 districts for the 2017 assessment cycle.

District200220032005200720092011201320152017
Albuquerque Public Schools xxxx
Atlanta Public Schools xxxxxxxxx
Austin Independent School District xxxxxxx
Baltimore City Public Schools xxxxx
Boston Public Schools xxxxxxxx
Charlotte-Mecklenburg Schools xxxxxxxx
Chicago Public Schools xxxxxxxxx
Clark County (NV) School District x
Cleveland Metropolitan School District xxxxxxxx
Dallas Independent School District xxxx
Denver Public Schools x
Detroit Public Schools xxxxx
District of Columbia Public Schools xxxxxxxxx
Duval County (FL) Public Schools xx
Fort Worth (TX) Independent School District x
Fresno Unified School District xxxxx
Guilford County (NC) Schools x
Hillsborough County (FL) Public Schools xxxx
Houston Independent School District xxxxxxxxx
Jefferson County (KY) Public Schools xxxxx
Los Angeles Unified School District xxxxxxxxx
Miami-Dade County Public Schools xxxxx
Milwaukee Public Schools xxxx
New York City Department of Education xxxxxxxxx
School District of Philadelphia xxxxx
San Diego Unified School District xxxxxxxx
Shelby County (TN) Schools x

Long-term trend

Long-term trend NAEP is administered to 9-, 13-, and 17-year-olds periodically at the national level. Long-term trend assessments measure student performance in mathematics and reading and allow the performance of today's students to be compared with students since the early 1970s.

Although long-term trend and main NAEP both assess mathematics and reading, there are several differences between them. In particular, the assessments differ in the content assessed, how often the assessment is administered, and how the results are reported. These and other differences mean that results from long-term trend and main NAEP cannot be compared directly. [10]

Although NAEP has been administered since the 1970's, in 2021 US DOE officials have decided to postpone the assessment in math and reading due to the COVID-19 pandemic. The reasons for postponing include the possibility of skewed student samples as well as results due to differing distance learning options and because of safety concerns for proctors and students. [11]

Assessment schedule

NAGB sets the calendar for NAEP assessments. Please refer to the entire assessment schedule for all NAEP assessments since 1968 and those planned through 2017.

Main NAEP assessments are typically administered over approximately six weeks between the end of January and the beginning of March of every year. Long-term trend assessments are typically administered every four years by age group between October and May. All of the assessments are administered by NAEP-contracted field staff across the country.

NAEP State Coordinators (NSC)

NAEP is conducted in partnership with states. The NAEP program provides funding for a full-time NSC in each state. He or she serves as the liaison between NAEP, the state's education agency, and the schools selected to participate.

NSCs provide many important services for the NAEP program and are responsible for:

New digitally-based assessments (DBA)

While most NAEP assessments are administered in a paper-and-pencil based format, NAEP is evolving to address the changing educational landscape through its transition to digitally-based assessments. NAEP is using the latest technology available to deliver assessments to students, and as technology evolves, so will the nature of delivery of the DBAs. The goal is for all NAEP assessments to be paperless by the end of the decade. The 2011 writing assessment was the first to be fully computer-based.

Interactive Computer Tasks (ICTs)

In 2009, ICTs were administered as part of the paper-and-pencil science assessment. The computer delivery affords measurement of science knowledge, processes, and skills not able to be assessed in other modes. Tasks included performance of investigations that include observations of phenomena that would otherwise take a long time, modeling of phenomena on a very large scale or invisible to the naked eye, and research of extensive resource documents.

Mathematics Computer-Based Study

This special study in multi-stage testing, implemented in 2011, investigated the use of adaptive testing principles in the NAEP context. A sample of students were given an online mathematics assessment which adapts to their ability level. All of the items in the study are existing NAEP items.

Technology and Engineering Literacy (TEL) Assessment

The TEL assessment framework describes technology and engineering literacy as the capacity to use, understand, and evaluate technology as well as to understand technological principles and strategies needed to develop solutions and achieve goals. The three areas of the assessment are:

Eighth-grade students throughout the nation took the assessment in winter of 2014. Results from this assessment were released in May 2016.

Writing Computer-Based Assessment

In 2011, NAEP transitioned its writing assessment (at grades 8 and 12) from paper and pencil to a computer-based administration in order to measure students' ability to write using a computer. The assessment takes advantage of many features of current digital technology and the tasks are delivered in multimedia formats, such as short videos and audio. Additionally, in an effort to include as many students as possible, the writing computer-based assessment system has embedded within it several universal design features such as text-to-speech, adjustable font size, and electronic spell check. In 2012, NAEP piloted the computer-based assessment for students at grade 4.

Studies using NAEP data

In addition to the assessments, NAEP coordinates a number of related special studies that often involve special data collection processes, secondary analyses of NAEP results, and evaluations of technical procedures.

Achievement gaps

Achievement gaps occur when one group of students outperforms another group and the difference in average scores for the two groups is statistically significant (that is, larger than the margin of error). In initial report releases NAEP highlights achievement gaps across student groups. However, NAEP has also releases a number of reports and data summaries that highlight achievement gap. – Some examples include the School Composition and the Black-White Achievement Gap and the Hispanic-White and the Black-White Achievement Gap Performance. [12] These publications use NAEP scores in mathematics and/or reading for these groups to either provide data summaries or illuminate patterns and changes in these gaps over time. Research reports, like the School Composition and Black-White Achievement Gap, also include caveats and cautions to interpreting the data.

High School Transcript Study (HSTS)

The HSTS explores the relationship between grade 12 NAEP achievement and high school academic careers by surveying the curricula being followed in our nation's high schools and the course-taking patterns of high school students through a collection of transcripts. Recent studies have placed an emphasis on STEM education and how it correlates to student achievement on the NAEP mathematics and science assessments.

NAEP-TIMSS Linking Study

The Trends in International Mathematics and Science Study (TIMSS) is an international assessment by the International Association for the Evaluation of Educational Achievement (IEA) that measures student learning in mathematics and science. NCES initiated the NAEP-TIMSS linking study so that states and selected districts can compare their own students' performance against international benchmarks. The linking study was conducted in 2011 at grade 8 in mathematics and science. NCES will "project", state and district-level scores on TIMSS in both subjects using data from NAEP.

National Indian Education Study (NIES)

The NIES is a two-part study designed to describe the condition of education for American Indian/Alaska Native students in the United States. The first part of the study consists of assessment results in mathematics and reading at grades 4 and 8. The second part presents the results of a survey given to American Indian/Alaska Native students, their teachers and their school administrators. The surveys focus on the students' cultural experiences in and out of school.

Mapping State Proficiency Standards

Under the 2001 reauthorization of the Elementary and Secondary Education Act (ESEA) of 1965, states develop their own assessments and set their own proficiency standards to measure student achievement. Each state controls its own assessment programs, including developing its own standards, resulting in great variation among the states in statewide student assessment practices. This variation creates a challenge in understanding the achievement levels of students across the United States. Since 2003, NCES has supported research that compares the proficiency standards of NAEP with those of individual states. State assessments are placed onto a common scale defined by NAEP scores, which allows states' proficiency standards to be compared not only to NAEP, but also to each other. NCES has released the Mapping State Proficiency Standards report using state data for mathematics and reading in 2003, 2005, 2007, 2009, and most recently 2013. [13]

Past studies

Over the years, NCES has conducted a number of other studies related to different aspects of the NAEP program. A few studies from the recent past are listed below:

Criticism

NAEP's heavy use of statistical hypothesis testing has drawn some criticism related to interpretation of results. For example, the Nation's Report Card reported "Males Outperform Females at all Three Grades in 2005" as a result of science test scores of 100,000 students in each grade. [14] Hyde and Linn criticized this claim, because the mean difference was only 4 out of 300 points, implying a small effect size and heavily overlapped distributions. They argue that "small differences in performance in the NAEP and other studies receive extensive publicity, reinforcing subtle, persistent, biases." [15]

NAEP's choice of which answers to mark right or wrong has also been criticized, a problem which happens in other countries too. [16] For example, a history question asked about the 1954 Brown v. Board of Education ruling, and explicitly referred to the 1954 decision which identified the problem, not the 1955 decision which ordered desegregation. NAEP asked students to "describe the conditions that this 1954 decision was designed to correct." They marked students wrong who mentioned segregation without mentioning desegregation. In fact the question asked only about existing conditions, not remedies, and in any case the 1954 decision did not order desegregation. [17] [18] The country waited until the 1955 Brown II decision to hear about "all deliberate speed." Another history question marked students wrong who knew the US fought Russians as well as Chinese and North Koreans in the Korean War. Other released questions on math and writing have had similar criticism. Math answers have penalized students who understand negative square roots, interest on loans, and errors in extrapolating a graph beyond the data. [19] [20]

NAEP's claim to measure critical thinking has also been criticized. UCLA researchers found that students could choose the correct answers without critical thinking. [21]

NAEP scores each test by a statistical method, sets cutoffs for "basic" and "proficient" standards, and gives examples of what students at each level accomplished on the test. The process to design the tests and standards has been criticized by Western Michigan University (1991), the National Academy of Education (1993), the Government Accountability Office (1993), the National Academy of Sciences (1999), [22] [23] the American Institutes for Research and RTI International (2007), [24] Brookings Institution (2007 [25] and 2016 [24] ), the Buros Center for Testing (2009), [22] and the National Academies of Sciences, Engineering, and Medicine (2016). [24]

Interpretation of NAEP results has been difficult: NAEP's category of "proficient" on a reading test given to fourth graders reflects students who do well on the test and are at seventh grade level. [24] NAEP's category of "proficient" on a math test given to eighth graders reflects students who do well on the test and are at twelfth grade level. [25] The fact that few eighth graders are proficient by this standard and achieve at twelfth grade level has been misinterpreted to allege that few eighth graders achieve even at eighth grade level. [26] NAEP says, "Students who may be proficient in a subject, given the common usage of the term, might not satisfy the requirements for performance at the NAEP achievement level" [24] James Harvey, principal author of A Nation at Risk, says, "It's hard to avoid concluding that the word was consciously chosen to confuse policymakers and the public." [24]

See also

Related Research Articles

<span class="mw-page-title-main">Standardized test</span> Test administered and scored in a predetermined, standard manner

A standardized test is a test that is administered and scored in a consistent, or "standard", manner. Standardized tests are designed in such a way that the questions and interpretations are consistent and are administered and scored in a predetermined, standard manner.

<span class="mw-page-title-main">No Child Left Behind Act</span> 2002 United States education reform law; repealed 2015

The No Child Left Behind Act of 2001 (NCLB) was a U.S. Act of Congress that reauthorized the Elementary and Secondary Education Act; it included Title I provisions applying to disadvantaged students. It supported standards-based education reform based on the premise that setting high standards and establishing measurable goals could improve individual outcomes in education. The Act required states to develop assessments in basic skills. To receive federal school funding, states had to give these assessments to all students at select grade levels.

Achievement gaps in the United States are observed, persistent disparities in measures of educational performance among subgroups of U.S. students, especially groups defined by socioeconomic status (SES), race/ethnicity and gender. The achievement gap can be observed through a variety of measures, including standardized test scores, grade point average, dropout rates, college enrollment, and college completion rates. The gap in achievement between lower income students and higher income students exists in all nations and it has been studied extensively in the U.S. and other countries, including the U.K. Various other gaps between groups exist around the globe as well.

<span class="mw-page-title-main">Programme for International Student Assessment</span> Scholastic performance study by the OECD

The Programme for International Student Assessment (PISA) is a worldwide study by the Organisation for Economic Co-operation and Development (OECD) in member and non-member nations intended to evaluate educational systems by measuring 15-year-old school pupils' scholastic performance on mathematics, science, and reading. It was first performed in 2000 and then repeated every three years. Its aim is to provide comparable data with a view to enabling countries to improve their education policies and outcomes. It measures problem solving and cognition.

The New England Common Assessment Program was a series of reading, writing, mathematics and science achievement tests, administered annually, which were developed in response to the Federal No Child Left Behind Act. Starting in 2005, school students in New Hampshire, Rhode Island, and Vermont participated in NECAP, and Maine joined the assessment program in 2009. It was a collaborative project of the New Hampshire, Rhode Island and Vermont departments of education, with assistance from the National Center for the Improvement of Educational Assessments. Measured Progress, an assessment contractor from Dover, New Hampshire, coordinates production, administration, scoring and reporting.

The Standards of Learning (SOL) is a public school standardized testing program in the Commonwealth of Virginia. It sets forth learning and achievement expectations for core subjects for grades K-12 in Virginia's Public Schools. The standards represent what many teachers, school administrators, parents, and business and community leaders believe schools should teach and students should learn. The Virginia Department of Education, schools, and school systems routinely receive essential feedback on the effectiveness of implementation and address effective instructional strategies and best practices. The Standards of Learning is supportive of and a direct response to the No Child Left Behind Act, which was signed into law by then-President George W. Bush on January 8, 2002. They address student achievement in four critical areas: (1) English, (2) mathematics, (3) science, and (4) history/social studies. Students are assessed in English and mathematics in grades 3-8 and upon completion of certain high school level courses. Science and history SOL are administered in grades 4, 5, and 8 and at the end of completing high school courses in these respective subjects.

Adequate Yearly Progress (AYP) was a measurement defined by the United States federal No Child Left Behind Act that allowed the U.S. Department of Education to determine how every public school and school district in the country was performing academically according to results on standardized tests. As defined by National Council on Measurement in Education (NCME), AYP was "the amount of annual achievement growth to be expected by students in a particular school, district, or state in the U.S. federal accountability system, No Child Left Behind (NCLB)." AYP has been identified as one of the sources of controversy surrounding George W. Bush administration's Elementary and Secondary Education Act. Private schools were not required to make AYP.

<span class="mw-page-title-main">Progress in International Reading Literacy Study</span> International study of fourth graders literacy

The IEA's Progress in International Reading Literacy Study (PIRLS) is an international study of reading (comprehension) achievement in 9-10 year olds. It has been conducted every five years since 2001 by the International Association for the Evaluation of Educational Achievement (IEA). It is designed to measure children's reading literacy achievement, to provide a baseline for future studies of trends in achievement, and to gather information about children's home and school experiences in learning to read.

<span class="mw-page-title-main">STAR (software)</span>

STAR Reading, STAR Early Literacy and STAR Math are standardized, computer-adaptive assessments created by Renaissance Learning, Inc., for use in K–12 education. Each is a "Tier 2" assessment of a skill that can be used any number of times due to item-bank technology. These assessments fall somewhere between progress monitoring tools and high-stakes tests.

The Colorado Student Assessment Program (CSAP) was an assessment required by the No Child Left Behind Act administered by the Unit of Student Assessment in the Colorado Department of Education (CDE). The CSAP was designed to measure how well students are learning material from the Colorado Model Content Standards, the established content standards that all Colorado public school students should learn. The CSAP only tested four of the thirteen subject areas in the Colorado Model Content Standards.

<span class="mw-page-title-main">Trends in International Mathematics and Science Study</span> Study of international math and science skills

The IEA's Trends in International Mathematics and Science Study (TIMSS) is a series of international assessments of the mathematics and science knowledge of students around the world. The participating students come from a diverse set of educational systems in terms of economic development, geographical location, and population size. In each of the participating educational systems, a minimum of 4,000 to 5,000 students is evaluated. Contextual data about the conditions in which participating students learn mathematics and science are collected from the students and their teachers, their principals, and their parents via questionnaires.

In an educational setting, standards-based assessment is assessment that relies on the evaluation of student understanding with respect to agreed-upon standards, also known as "outcomes". The standards set the criteria for the successful demonstration of the understanding of a concept or skill.

Sandra Stotsky is Professor emerita in the Department of Education Reform at the University of Arkansas, and held the 21st Century Chair in Teacher Quality. Her research ranges from teacher licensure tests, e.g., (1), coherence in the literature and reading curriculum, e.g., (2), and academic achievement in single-sex classrooms, e.g., (3) to critiques of Common Core’s standards in English language arts, e.g., (4) mathematics.(5), and US History and civic education (6), and other aspects of the Common Core project, e.g., (7), and to reviews of books in education, e.g., (8) She is an advocate of standards-based reform and strong academic standards and assessments for students and teachers.

The Interactive Mathematics Program (IMP) is a four-year, problem-based mathematics curriculum for high schools. It was one of several curricula funded by the National Science Foundation and designed around the 1989 National Council of Teachers of Mathematics (NCTM) standards. The IMP books were authored by Dan Fendel and Diane Resek, professors of mathematics at San Francisco State University, and by Lynne Alper and Sherry Fraser. IMP was published by Key Curriculum Press in 1997 and sold in 2012 to It's About Time.

Education in Kentucky includes elementary school, middle school, high school, and post-secondary institutions. Most Kentucky schools and colleges are accredited through the Southern Association of Colleges and Schools (SACS).

The Pennsylvania System of School Assessment (PSSA) is a standardized test administered in public schools in the state of Pennsylvania. Students in grades 3-8 are assessed in English language arts skills and mathematics. Students in grades 4 and 8 are also assessed in skills relating to natural science, including the field of data interpretation and analysis. Since 2013, high school students have taken the Keystone Exam in place of the PSSA for their standardized testing. The PSSA's were made by a company in New Jersey. The PSSA is written, owned and administered by Pearson Education. There are reporting categories for each subject which list eligible content to be tested in each grade. Assessment Anchors specify what is considered eligible content for each grade level tested. A Proficient or Advanced level is needed to be able to qualify as passing the PSSA.

<span class="mw-page-title-main">Literacy in the United States</span> Overview of literacy in the United States

Literacy in the United States was categorized by the National Center for Education Statistics into different literacy levels, with 92% of American adults having at least "Level 1" literacy in 2014. Nationally, over 20% of adult Americans have a literacy proficiency at or below Level 1. Adults in this range have difficulty using or understanding print materials. Those on the higher end of this category can perform simple tasks based on the information they read, but adults below Level 1 may only understand very basic vocabulary or be functionally illiterate. According to a 2020 report by the U.S. Department of Education, 54% of adults in the United States have English prose literacy below the 6th-grade level.

The education system in Mississippi consists of elementary, middle, and high schools as well as boarding schools, public and private schools, and colleges and universities.

The racial achievement gap in the United States refers to disparities in educational achievement between differing ethnic/racial groups. It manifests itself in a variety of ways: African-American and Hispanic students are more likely to earn lower grades, score lower on standardized tests, drop out of high school, and they are less likely to enter and complete college than whites, while whites score lower than Asian Americans.

<span class="mw-page-title-main">Elementary schools in the United States</span> Primary education in the United States

In the United States, elementary schools are the main point of delivery of primary education, for children between the ages of 4–11 and coming between pre-kindergarten and secondary education.

References

  1. "History and Innovation - What is the Nation's Report Card | NAEP". nces.ed.gov. Retrieved 2022-04-05.
  2. "Nation's report card".
  3. Mullis, Ina V. S. (2019). White Paper on 50 Years of NAEP Use: Where NAEP Has Been and Where It Should Go Next (PDF) (Report). American Institutes for Research. Retrieved 2021-03-12.
  4. Jones, Lyle V. (1996). "A history of the National Assessment of Educational Progress and some questions about its future" . Educational Researcher. 25 (7): 15–22. doi:10.3102/0013189X025007015. JSTOR   1176519. S2CID   145442224 . Retrieved 2021-03-12 via JSTOR.
  5. "Revised, November 2018 G ALLOWABLE ACCOMMODATIONS FOR STUDENTS WITH DISABILITIES" (PDF). gadoe.org. 2018-11-05.
  6. Freedman, Miriam (2009-03-04). "NAEP and Testing Students with Disabilities and English Language Learners" (PDF). NAGB.
  7. "Measuring Student Progress Since 1964". National Center for Education Statistics. Retrieved 2011-09-29.
  8. "Frameworks and Specifications". National Center for Education Statistics. Retrieved 2011-09-29.
  9. "Results for public school students in 11 states available for the first time". National Center for Education Statistics. Retrieved 2011-09-29.
  10. "What are the main differences between Long-Term Trend NAEP and Main NAEP?". National Center for Education Statistics. Retrieved 2011-09-29.
  11. "Commissioner's Remarks-Due to COVID Pandemic, NCES to delay National Assessment of Education Progress (NAEP) assessment-November 25, 2020". nces.ed.gov. Retrieved 2021-01-28.
  12. Achievement Gaps NAEP (homesite), retrieved 13 April 2013
  13. Mapping State Proficiency Standards National Center for Education Statistics, retrieved 13 April 2013
  14. "Male and Female Students Make Gains Since 2000 at Grade 4; Males Outperform Females at all Three Grades in 2005". The Nation's Report Card. U.S. Department of Education. Retrieved 16 September 2012.
  15. Hyde, Janet Shibley; Marcia C. Linn (27 October 2006). "Gender similarities in mathematics and science". Science. 314 (5799): 599–600. doi:10.1126/science.1132154. PMID   17068246. S2CID   34045261 . Retrieved 16 September 2012.
  16. Cassidy, Caitlin (2023-11-18). "An urgent overhaul of VCE exams is needed after multiple errors, experts say. But how did this happen?". The Guardian. ISSN   0261-3077 . Retrieved 2023-11-20.
  17. Liberman, Mark (2011-06-22). "Language Log » A reading comprehension test". U of Pennsylvania. Retrieved 2020-09-07.
  18. Wineburg, Sam (2011-10-24). "Testing students' knowledge of the civil rights movement". Los Angeles Times. Retrieved 2020-09-07.
  19. Burke, Paul. "Wrong "Correct" Answers: The Scourge of the NAEP". History News Network - George Washington University. Retrieved 2020-09-07.
  20. Burke, Paul (1990-08-28). "U.S. STUDENTS THE MYTH OF MASSIVE FAILURE". Washington Post. ISSN   0190-8286 . Retrieved 2020-09-07.
  21. Winebeurg, Sam, Mark Smith and Joel Breakstone (2017-09-19). "The 'nation's report card' says it assesses critical thinking in history — but NAEP gets an F on that score". Washington Post. ISSN   0190-8286 . Retrieved 2020-09-07.{{cite news}}: CS1 maint: multiple names: authors list (link)
  22. 1 2 Harvey, James (2011-11-04). "NAEP: A flawed benchmark producing the same old story". Washington Post. Retrieved 2020-09-07.
  23. "NAEP Levels Found To Be Flawed". www.fairtest.org. Retrieved 2020-09-07.
  24. 1 2 3 4 5 6 Harvey, James (February 2018). "The Problem with "Proficient"". Educational Leadership.
  25. 1 2 Loveless, Tom (2016-06-13). "The NAEP proficiency myth". Brookings. Retrieved 2020-09-07.
  26. Strauss, Valerie (2016-05-23). "Why a social media fight between Campbell Brown and her critics matters". Washington Post. ISSN   0190-8286 . Retrieved 2020-09-07.

Further reading