From 11ec8368dea759bbb074ede158a762586996d610 Mon Sep 17 00:00:00 2001 From: Jaques Grobler Date: Wed, 15 Jan 2014 12:23:05 +0100 Subject: [PATCH] small fixes --- .../text_analytics/working_with_text_data.rst | 14 +++++++------- 1 file changed, 7 insertions(+), 7 deletions(-) diff --git a/doc/tutorial/text_analytics/working_with_text_data.rst b/doc/tutorial/text_analytics/working_with_text_data.rst index d4b42bf2e8a3e..29eceb85ec378 100644 --- a/doc/tutorial/text_analytics/working_with_text_data.rst +++ b/doc/tutorial/text_analytics/working_with_text_data.rst @@ -537,31 +537,31 @@ Bonus point if the utility is able to give a confidence level for its predictions. -Where to go from Here +Where to from here ------------------ Here are a few suggestions to help further your scikit-learn intuition upon the completion of this tutorial: -- Try playing around with the `analyzer` and `token normalisation` under +* Try playing around with the ``analyzer`` and ``token normalisation`` under :class:`CountVectorizer` -- If you don't have labels, try using +* If you don't have labels, try using :ref:`Clustering ` on your problem. -- If you have multiple labels per document, e.g categories, have a look +* If you have multiple labels per document, e.g categories, have a look at the :ref:`Multiclass and multilabel section ` -- Try using :ref:`Truncated SVD ` for +* Try using :ref:`Truncated SVD ` for `latent semantic analysis `_. -- Have a look at using +* Have a look at using :ref:`Out-of-core Classification ` to learn from data that would not fit into the computer main memory. -- Have a look at the :ref:`Hashing Vectorizer ` +* Have a look at the :ref:`Hashing Vectorizer ` as a memory efficient alternative to :class:`CountVectorizer`.