From cfc1e965d51ed2fc2c6462e979094e47e47ba2f4 Mon Sep 17 00:00:00 2001 From: Vsevolod Stakhov Date: Thu, 2 Apr 2015 14:10:09 +0100 Subject: [PATCH] Fix tokenization of the last token in a message. --- src/libstat/tokenizers/tokenizers.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/libstat/tokenizers/tokenizers.c b/src/libstat/tokenizers/tokenizers.c index eebc57c22d..4b8f0cfb39 100644 --- a/src/libstat/tokenizers/tokenizers.c +++ b/src/libstat/tokenizers/tokenizers.c @@ -295,7 +295,7 @@ rspamd_tokenize_text (gchar *text, gsize len, gboolean is_utf, res = g_array_sized_new (FALSE, FALSE, sizeof (rspamd_fstring_t), 128); while (func (&buf, &pos, &token, &cur, is_utf, &l)) { - if (min_len > 0 && l < min_len) { + if (l == 0 || (min_len > 0 && l < min_len)) { token.begin = pos; continue; } -- 2.47.3