This guide provides an overview for how to use all subscriptions to the Lexile Analyzer® including the Professional Lexile Analyzer® and the Lexile Analyzer® Editor Assistant™. The Lexile Analyzer® is used to evaluate the text complexity of books, articles, passages, and other texts to report its reading demand as a Lexile® text measure. The Lexile Analyzer® is a web-based tool that determines the Lexile® measure of professionally edited, complete, conventional prose text. The accuracy of the Lexile® measures produced from the Lexile Analyzer® depends on your following the text-preparation procedures and formatting conventions detailed in this guide.
Certain categories of text should not be measured as a Lexile measure. Because The Lexile® Framework for Reading was built upon the measurement of professionally edited, complete, conventional prose text, the Lexile Analyzer will return an inaccurate Lexile measure for other kinds of text. Follow these guidelines as you choose texts to measure:
Please observe the following text-preparation guidelines before you submit your sample file to the Lexile Analyzer.
Here are some guidelines for removing non-prose text:
Additional considerations should be made when editing a text for measurement. Historical notes, introductions, “About the author" pieces, and previews of the next book in a series should typically be removed. Such text is often written separately from the main text and thus contains unique textual characteristics that can influence the Lexile measure. However, such decisions should be carefully considered while preparing your text for analysis. Some frontmatter and backmatter may be a legitimate part of the larger text and should be included. As a general guideline, if text appears to be written by the same author for the same audience, then it should be included in the Lexile analysis.
In the layout of children’s picture books, single sentences are sometimes distributed across multiple pages of a book. In the activity of reading, these page breaks function as sentence endings, so a pipe character (|) should be inserted at each page break in your file. The Lexile Analyzer interprets pipes as sentence-ending punctuation. In the example below from Ludwig Bemelmans’ Madeline (Puffin Books), a pipe would be placed in the plain text file after the words “crank" to emulate the effect that page breaks have on reading, and the comma would be removed.
When using resources downloaded from websites, be sure to remove the non-prose and web page-specific elements, as indicated in the example below from a CNN.com article:
Only the main body of the article (G) and the complete-sentence figure caption (D) should be measured. The article title (A), date line (B), and image (C), as well as website-specific elements such as social sharing links (E) and margin advertisement text (F), should not be measured.
Also be careful to eliminate all HTML code and URLs from your sample file when measuring web resources.
If you typed your text in the text editing box or if the words and sentences in your file are recognizable, proceed to step 2.
If the text you are trying to measure is in an image file or part of printed material, you can scan the pages, save them as a PDF file, and load them into an optical character recognition (OCR) program. There are several different OCR programs you can use to convert an image to text, including ones offered by Acrobat and ABBYY FineReader.
Occasionally OCR results can be inconsistent, particularly involving punctuation marks such as periods not being recognized at all (see below). These inconsistencies will impact the accuracy of the Lexile measure. Unfortunately, the repair of a poorly OCRed file can take as long as typing the text.
If a text is converted from hard copy to electronic format using an OCR application, some problems may occur in the conversion process. These tend to relate to the specific software used, and special care should be taken to ensure the accuracy of the electronic facsimile.
The Lexile Analyzer is designed to measure professionally-edited, complete, conventional prose text. It should not be used on non-prose, unpunctuated, or unconventional text. The Lexile Analyzer determines sentence length through recognition of sentence endings, so sentences must be conventionally punctuated to be recognized (refer to the FAQ How does the Lexile Analyzer work? for sentence ender information). Likewise, the Lexile Analyzer determines word frequency by recognizing correctly spelled, well-formed words. Otherwise, the Lexile Analyzer will not return a useful estimated Lexile measure.
You should keep in mind that the usefulness of an estimated Lexile measure depends on the proper preparation of a text for analysis. Seemingly minor errors can result in significant variation in Lexile measures.
The Lexile Analyzer requires a UTF-8 plain text document (*.txt file) for proper processing and Lexile measurement. A plain text file is one which uses only the basic UTF-8 character set and contains no special formatting. If you submit files of an incorrect format to the Lexile Analyzer, an incorrect Lexile measure will be returned.
Note: The Lexile Analyzer cannot measure Microsoft Word, PDF, HTML or scanned image files such as JPGs.
If the source text to be measured is in an electronic document format, such as a word processing document or a rich text document, the file usually can be converted into the plain text format using the settings in the application’s Save As… menu.
It is advisable to open your plain text document to check it for conversion errors. Pay attention to special characters such as quotation marks, apostrophes, ellipses, or accented characters. Also, ensure that em-dashes and en-dashes have appropriate spacing on either side. They will be converted to hyphens, which can make the Lexile Analyzer interpret the two words on either side of an em-dash as a single hyphenated word unless there are spaces on either side of the hyphen.
Select the “Analyze" button and your Lexile Analyzer results appear on the screen. The type of subscription to the Lexile Analyzer that you have will determine the data and features you have access to. All subscriptions provide access to the Lexile measure, Mean Sentence Length, Mean Log Word Frequency, and Word Count. For the descriptions of additional data and features displayed for the Professional Lexile Analyzer and Lexile Analyzer Editor Assistant, see the informational pop-up boxes on the results page.
The Professional Analyzer maintains a log history and the Lexile Analyzer Editor Assistant provides a document management system with features like category tagging and editing history. To learn more about the type of subscriptions and what features they offer, see the FAQ What type of subscriptions are available for the Lexile Analyzer?
MetaMetrics (developers of the Lexile Framework) measures a book at a publisher’s request. Books are always measured in their entirety. Publishers pay for this service, as well as the right to use the Lexile measure in their marketing materials.
In order to ensure the most accurate Lexile measure, MetaMetrics’ text measurement process includes the following steps, with quality checks at each stage:
Several publishers and content developers use the Professional Lexile Analyzer and Lexile Analyzer Editor Assistant to help write text at certain reading levels and measure their texts developmental level before submitting for a certified Lexile measure. Upon submission to MetaMetrics, these files have been prepared using our text-preparation guidelines. These files are reviewed by our resource measurement coordinators to assure that the editing guidelines have been met. They are then submitted for Lexile code review and Lexile measures are returned to the publisher. It is only after review by MetaMetrics’ resource measurement team that these measures are deemed “certified" and then available for distribution via marketing materials, websites, and searchable in “Find a Book."
The Lexile Analyzer works in steps: first by calculating a number of text complexity variables for a given text such as the length of sentences. Text complexity variables are then combined to calculate the Lexile measure of the text. The Lexile measure is an indication of the degree of challenge a particular text will pose for a particular reader. Lexile measures are units of measure just like degrees Fahrenheit.
Like the measurement of temperature, the measurement of text complexity requires different measurement instruments at different locations on the scale. For example, you would use a different instrument to measure a person’s body temperature than you would to measure the temperature inside your oven, but each instrument still provides measures on the same scale.
Similarly, for texts used in the early grades (generally texts intended for grades K thru 2), different text complexity variables are required to provide the most accurate measurement possible. For texts intended for readers above grade 2, a Lexile measure has a semantic and syntactic component represented by MLF (Mean Log Frequency) and LMSL (Log Mean Sentence Length) respectively. Research has shown that word frequency can be used as a proxy for vocabulary difficulty, and sentence length can be used as a proxy for sentence complexity. Word frequency is calculated using the MetaMetrics corpus of over 1.4 billion words intended for K–12 readers. For texts intended for readers in grades 2 and below, nine variables are used to measure four aspects of text important for early-reading: structure, syntax, semantics, and decoding. Structure refers to the systematic patterning and repetition often found in early-grades texts, but not typically found in texts for older readers. Decoding refers to the complexity of the orthographic patterns in the text and is related to the important task of learning to decode printed text into spoken or silently read words.
When a text is analyzed, all of its text complexity variables are calculated and a determination is made automatically as to whether the early-reading text complexity variables are required for analysis. For texts 650L and below, four additional early-reading indicators, relevant for early-grades texts, are reported: structure, syntax, semantics, and decoding.
Consequently, when using the Lexile Analyzer to measure text, you should keep in mind two keys to getting an accurate Lexile measure:
Proper file preparation, as detailed in the earlier section “Step 2: Prepare your text for the Lexile Analyzer" is the crucial step for ensuring Lexile measurement accuracy. File preparation errors or oversights, such as missing or incorrect punctuation or sections of unconventional prose or non-prose, may compromise your Lexile Analyzer results and return an estimated Lexile measure too far from the actual Lexile measure to be of use to you.
The measurement impact of editing errors and oversights is more severe the shorter the length of the input file. For this reason, special attention is encouraged when preparing a short passage, article, or children’s text for analysis.
There are two Lexile Analyzer products offered on a paid subscription basis. These products are intended for commercial education companies and content developers. The products include:
If you are interested in using the Lexile Analyzer for commercial purposes, visit Licensing the Professional Lexile Analyzer.
Once you are logged into your account, click on My Account on the top menu bar.
The next screen will show you what the products to which you have access.
You can also see your subscription type on the upper left hand corner once you are logged into your account. For example, if you have access to the Lexile Analyzer Editor Assistant, you will see this: