From d8a935a88f087a9c3161421411f6c2a0dfc11017 Mon Sep 17 00:00:00 2001 From: Spencer Davies Date: Thu, 18 Jun 2020 19:30:38 -0600 Subject: [PATCH] two simple typos fixed use `a` in front of unicode instead of `an` and when talking about multiple parameters it needs to be plural with the `s`. Thanks! --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index ee2ed8aa..88a2a9e0 100644 --- a/README.md +++ b/README.md @@ -521,8 +521,8 @@ Main Functions 1. Cut -------- * The `jieba.cut` function accepts three input parameters: the first parameter is the string to be cut; the second parameter is `cut_all`, controlling the cut mode; the third parameter is to control whether to use the Hidden Markov Model. -* `jieba.cut_for_search` accepts two parameter: the string to be cut; whether to use the Hidden Markov Model. This will cut the sentence into short words suitable for search engines. -* The input string can be an unicode/str object, or a str/bytes object which is encoded in UTF-8 or GBK. Note that using GBK encoding is not recommended because it may be unexpectly decoded as UTF-8. +* `jieba.cut_for_search` accepts two parameters: the string to be cut; whether to use the Hidden Markov Model. This will cut the sentence into short words suitable for search engines. +* The input string can be a unicode/str object, or a str/bytes object which is encoded in UTF-8 or GBK. Note that using GBK encoding is not recommended because it may be unexpectly decoded as UTF-8. * `jieba.cut` and `jieba.cut_for_search` returns an generator, from which you can use a `for` loop to get the segmentation result (in unicode). * `jieba.lcut` and `jieba.lcut_for_search` returns a list. * `jieba.Tokenizer(dictionary=DEFAULT_DICT)` creates a new customized Tokenizer, which enables you to use different dictionaries at the same time. `jieba.dt` is the default Tokenizer, to which almost all global functions are mapped.