You cannot go straight from raw text to fitting a machine learning or deep learning model.
You must clean your text first, which means splitting it into words and handling punctuation and case.
In fact, there is a whole suite of text preparation methods that you may need to use, and the choice of methods really depends on your natural language processing task.
In this tutorial, you will discover how you can clean and prepare your text ready for modeling with machine learning.
After completing this tutorial, you will know:
- How to get started by developing your own very simple text cleaning tools.
- How to take a step up and use the more sophisticated methods in the NLTK library.
- How to prepare text when using modern text representation methods like word embeddings.Let’s get started.
Metamorphosis by Franz Kafka
Let’s start off by selecting a dataset.
In this tutorial, we will use the text from the book Metamorphosis by Franz Kafka. No specific reason, other than it’s short, I like it, and you may like it too. I expect it’s one of those classics that most students have to read in school.
The full text for Metamorphosis is available for free from Project Gutenberg.
You can download the ASCII text version of the text here:
- Metamorphosis by Franz Kafka Plain Text UTF-8 (may need to load the page twice).
Download the file and place it in your current working directory with the file name “metamorphosis.txt“.
The file contains header and footer information that we are not interested in, specifically copyright and license information. Open the file and delete the header and footer information and save the file as “metamorphosis_clean.txt“.
The start of the clean file should look like:
One morning, when Gregor Samsa woke from troubled dreams, he found himself transformed in his bed into a horrible vermin.
The file should end with:
And, as if in confirmation of their new dreams and good intentions, as soon as they reached their destination Grete was the first to get up and stretch out her young body.
Poor Gregor…
Text Cleaning Is Task Specific
After actually getting a hold of your text data, the first step in cleaning up text data is to have a strong idea about what you’re trying to achieve, and in that context review your text to see what exactly might help.
Take a moment to look at the text. What do you notice?
Here’s what I see:
- It’s plain text so there is no markup to parse (yay!).
- The translation of the original German uses UK English (e.g. “travelling“).
- The lines are artificially wrapped with new lines at about 70 characters (meh).
- There are no obvious typos or spelling mistakes.
- There’s punctuation like commas, apostrophes, quotes, question marks, and more.
- There’s hyphenated descriptions like “armour-like”.
- There’s a lot of use of the em dash (“-“) to continue sentences (maybe replace with commas?).
- There are names (e.g. “Mr. Samsa“)
- There does not appear to be numbers that require handling (e.g. 1999)
- There are section markers (e.g. “II” and “III”), and we have removed the first “I”.
I’m sure there is a lot more going on to the trained eye.
We are going to look at general text cleaning steps in this tutorial.
Nevertheless, consider some possible objectives we may have when working with this text document.
For example:
- If we were interested in developing a Kafkaesque language model, we may want to keep all of the case, quotes, and other punctuation in place.
- If we were interested in classifying documents as “Kafka” and “Not Kafka,” maybe we would want to strip case, punctuation, and even trim words back to their stem.
Use your task as the lens by which to choose how to ready your text data.
Manual Tokenization
Text cleaning is hard, but the text we have chosen to work with is pretty clean already.
We could just write some Python code to clean it up manually, and this is a good exercise for those simple problems that you encounter. Tools like regular expressions and splitting strings can get you a long way.
1. Load Data
Let’s load the text data so that we can work with it.
The text is small and will load quickly and easily fit into memory. This will not always be the case and you may need to write code to memory map the file. Tools like NLTK (covered in the next section) will make working with large files much easier.
We can load the entire “metamorphosis_clean.txt” into memory as follows:
Running the example loads the whole file into memory ready to work with.
2. Split by Whitespace
Clean text often means a list of words or tokens that we can work with in our machine learning models.
This means converting the raw text into a list of words and saving it again.
A very simple way to do this would be to split the document by white space, including ” “, new lines, tabs and more. We can do this in Python with the split() function on the loaded string.
Running the example splits the document into a long list of words and prints the first 100 for us to review.
We can see that punctuation is preserved (e.g. “wasn’t” and “armour-like“), which is nice. We can also see that end of sentence punctuation is kept with the last word (e.g. “thought.”), which is not great.
3. Select Words
Another approach might be to use the regex model (re) and split the document into words by selecting for strings of alphanumeric characters (a-z, A-Z, 0-9 and ‘_’).
For example:
Again, running the example we can see that we get our list of words. This time, we can see that “armour-like” is now two words “armour” and “like” (fine) but contractions like “What’s” is also two words “What” and “s” (not great).
3. Split by Whitespace and Remove Punctuation
Note: This example was written for Python 3.
We may want the words, but without the punctuation like commas and quotes. We also want to keep contractions together.
One way would be to split the document into words by white space (as in “2. Split by Whitespace“), then use string translation to replace all punctuation with nothing (e.g. remove it).
Python provides a constant called string.punctuation that provides a great list of punctuation characters. For example:
Results in:
Python offers a function called translate() that will map one set of characters to another.
We can use the function maketrans() to create a mapping table. We can create an empty mapping table, but the third argument of this function allows us to list all of the characters to remove during the translation process. For example:
We can put all of this together, load the text file, split it into words by white space, then translate each word to remove the punctuation.
We can see that this has had the desired effect, mostly.
Contractions like “What’s” have become “Whats” but “armour-like” has become “armourlike“.
If you know anything about regex, then you know things can get complex from here.
4. Normalizing Case
It is common to convert all words to one case.
This means that the vocabulary will shrink in size, but some distinctions are lost (e.g. “Apple” the company vs “apple” the fruit is a commonly used example).
We can convert all words to lowercase by calling the lower() function on each word.
For example:
Running the example, we can see that all words are now lowercase.
Note
Cleaning text is really hard, problem specific, and full of tradeoffs.
Remember, simple is better.
Simpler text data, simpler models, smaller vocabularies. You can always make things more complex later to see if it results in better model skill.
Next, we’ll look at some of the tools in the NLTK library that offer more than simple string splitting.
Tokenization and Cleaning with NLTK
The Natural Language Toolkit, or NLTK for short, is a Python library written for working and modeling text.
It provides good tools for loading and cleaning text that we can use to get our data ready for working with machine learning and deep learning algorithms.
1. Install NLTK
You can install NLTK using your favorite package manager, such as pip:
After installation, you will need to install the data used with the library, including a great set of documents that you can use later for testing other tools in NLTK.
There are few ways to do this, such as from within a script:
Or from the command line:
For more help installing and setting up NLTK, see:
2. Split into Sentences
A good useful first step is to split the text into sentences.
Some modeling tasks prefer input to be in the form of paragraphs or sentences, such as word2vec. You could first split your text into sentences, split each sentence into words, then save each sentence to file, one per line.
NLTK provides the sent_tokenize() function to split text into sentences.
The example below loads the “metamorphosis_clean.txt” file into memory, splits it into sentences, and prints the first sentence.
Running the example, we can see that although the document is split into sentences, that each sentence still preserves the new line from the artificial wrap of the lines in the original document.
One morning, when Gregor Samsa woke from troubled dreams, he found
himself transformed in his bed into a horrible vermin.3. Split into Words
NLTK provides a function called word_tokenize() for splitting strings into tokens (nominally words).
It splits tokens based on white space and punctuation. For example, commas and periods are taken as separate tokens. Contractions are split apart (e.g. “What’s” becomes “What” “‘s“). Quotes are kept, and so on.
For example:
Running the code, we can see that punctuation are now tokens that we could then decide to specifically filter out.
4. Filter Out Punctuation
We can filter out all tokens that we are not interested in, such as all standalone punctuation.
This can be done by iterating over all tokens and only keeping those tokens that are all alphabetic. Python has the function isalpha() that can be used. For example:
Running the example, you can see that not only punctuation tokens, but examples like “armour-like” and “‘s” were also filtered out.
5. Filter out Stop Words (and Pipeline)
Stop words are those words that do not contribute to the deeper meaning of the phrase.
They are the most common words such as: “the“, “a“, and “is“.
For some applications like documentation classification, it may make sense to remove stop words.
NLTK provides a list of commonly agreed upon stop words for a variety of languages, such as English. They can be loaded as follows:
You can see the full list as follows:
You can see that they are all lower case and have punctuation removed.
You could compare your tokens to the stop words and filter them out, but you must ensure that your text is prepared the same way.
Let’s demonstrate this with a small pipeline of text preparation including:
- Load the raw text.
- Split into tokens.
- Convert to lowercase.
- Remove punctuation from each token.
- Filter out remaining tokens that are not alphabetic.
- Filter out tokens that are stop words.
Running this example, we can see that in addition to all of the other transforms, stop words like “a” and “to” have been removed.
I note that we are still left with tokens like “nt“. The rabbit hole is deep; there’s always more we can do.
6. Stem Words
Stemming refers to the process of reducing each word to its root or base.
For example “fishing,” “fished,” “fisher” all reduce to the stem “fish.”
Some applications, like document classification, may benefit from stemming in order to both reduce the vocabulary and to focus on the sense or sentiment of a document rather than deeper meaning.
There are many stemming algorithms, although a popular and long-standing method is the Porter Stemming algorithm. This method is available in NLTK via the PorterStemmer class.
For example:
Running the example, you can see that words have been reduced to their stems, such as “trouble” has become “troubl“. You can also see that the stemming implementation has also reduced the tokens to lowercase, likely for internal look-ups in word tables.
You can also see that the stemming implementation has also reduced the tokens to lowercase, likely for internal look-ups in word tables.
There is a nice suite of stemming and lemmatization algorithms to choose from in NLTK, if reducing words to their root is something you need for your project.
Additional Text Cleaning Considerations
We are only getting started.
Because the source text for this tutorial was reasonably clean to begin with, we skipped many concerns of text cleaning that you may need to deal with in your own project.
Here is a short list of additional considerations when cleaning text:
- Handling large documents and large collections of text documents that do not fit into memory.
- Extracting text from markup like HTML, PDF, or other structured document formats.
- Transliteration of characters from other languages into English.
- Decoding Unicode characters into a normalized form, such as UTF8.
- Handling of domain specific words, phrases, and acronyms.
- Handling or removing numbers, such as dates and amounts.
- Locating and correcting common typos and misspellings.
- …
The list could go on.
Hopefully, you can see that getting truly clean text is impossible, that we are really doing the best we can based on the time, resources, and knowledge we have.
The idea of “clean” is really defined by the specific task or concern of your project.
A pro tip is to continually review your tokens after every transform. I have tried to show that in this tutorial and I hope you take that to heart.
Ideally, you would save a new file after each transform so that you can spend time with all of the data in the new form. Things always jump out at you when to take the time to review your data.
Have you done some text cleaning before? What are you preferred pipeline of transforms?
Let me know in the comments below.Tips for Cleaning Text for Word Embedding
Recently, the field of natural language processing has been moving away from bag-of-word models and word encoding toward word embeddings.
The benefit of word embeddings is that they encode each word into a dense vector that captures something about its relative meaning within the training text.
This means that variations of words like case, spelling, punctuation, and so on will automatically be learned to be similar in the embedding space. In turn, this can mean that the amount of cleaning required from your text may be less and perhaps quite different to classical text cleaning.
For example, it may no-longer make sense to stem words or remove punctuation for contractions.
Tomas Mikolov is one of the developers of word2vec, a popular word embedding method. He suggests only very minimal text cleaning is required when learning a word embedding model.
Below is his response when pressed with the question about how to best prepare text data for word2vec.
There is no universal answer. It all depends on what you plan to use the vectors for. In my experience, it is usually good to disconnect (or remove) punctuation from words, and sometimes also convert all characters to lowercase. One can also replace all numbers (possibly greater than some constant) with some single token such as .
All these pre-processing steps aim to reduce the vocabulary size without removing any important content (which in some cases may not be true when you lowercase certain words, ie. ‘Bush’ is different than ‘bush’, while ‘Another’ has usually the same sense as ‘another’). The smaller the vocabulary is, the lower is the memory complexity, and the more robustly are the parameters for the words estimated. You also have to pre-process the test data in the same way.
…
In short, you will understand all this much better if you will run experiments.
Read the full thread on Google Groups.
Further Reading
This section provides more resources on the topic if you are looking go deeper.
- Metamorphosis by Franz Kafka on Project Gutenberg
- nltk.tokenize package API
- nltk.stem package API
- Chapter 3: Processing Raw Text, Natural Language Processing with Python
Summary
In this tutorial, you discovered how to clean text or machine learning in Python.
Specifically, you learned:
- How to get started by developing your own very simple text cleaning tools.
- How to take a step up and use the more sophisticated methods in the NLTK library.
- How to prepare text when using modern text representation methods like word embeddings.
Do you have any questions?
Ask your questions in the comments below and I will do my best to answer.Do you have experience with cleaning text?
Please share your experiences in the comments below.
No comments:
Post a Comment