util.utf8: improve CJK character detection

Previously the CJK character detection defined only characters in the
range U+4000..U+AFFF as "CJK characters". This excludes an incredibly
large number of CJK characters within the BMP, let alone the whole two
planes dedicated to rarer CJK characters (the SIP and TIP). As a result,
a very large number of Chinese, Japanese, and Korean characters were not
detected as being CJK characters.

While slightly less elegant-looking, it is far more accurate to compute
the codepoint from the utf8 character and then see if it falls within
one of the defined CJK blocks. This is not future-proof against future
CJK ideograph extensions in future Unicode versions, but there is no
real way to accurately predict such changes so this is the best we can
do without accidentally treating characters explicitily defined as being
non-CJK in Unicode as CJK.

While we're at it, copy Lua 5.3's utf8.charpattern constant definition
so that we can more easily write utf8 iterators with string.gmatch (at
least in the interim until there is a rework of utf8 handling in
KOReader and everything is rebuilt on top of utf8proc).

Some unit tests are added for Korean and Japanese text, and the existing
unit tests needed a minor adjustment to handle the fact that
isSplittable now correctly detects CJK punctuation as a character to
compare against the forbidden split rules.

Signed-off-by: Aleksa Sarai <cyphar@cyphar.com>
This commit is contained in:
Aleksa Sarai
2021-10-23 21:12:38 +11:00
committed by Frans de Jonge
parent 8e017de898
commit 6f1b70e5eb
2 changed files with 132 additions and 11 deletions

View File

@@ -88,6 +88,23 @@ describe("util module", function()
"","","","","","","","","","","","","","","",
}, words)
end)
it("should split Japanese words", function()
local words = util.splitToWords("色は匂へど散りぬるを我が世誰ぞ常ならむ")
assert.are_same({
"","","","","","","","","","",
"","","","","","","","","",
}, words)
end)
it("should split Korean words", function()
-- Technically splitting on spaces is correct but we treat Korean
-- as if it were any other CJK text.
local words = util.splitToWords("대한민국의 국기는 대한민국 국기법에 따라 태극기")
assert.are_same({
"","","","",""," ","","",""," ",
"","","",""," ","","","",""," ",
"",""," ","","","",
}, words)
end)
it("should split words of multilingual text", function()
local words = util.splitToWords("BBC纪录片")
assert.are_same({"BBC", "", "", ""}, words)
@@ -108,7 +125,7 @@ describe("util module", function()
table.insert(table_of_words, word)
word = ""
end
if i == #table_chars then table.insert(table_of_words, word) end
if i == #table_chars and word ~= "" then table.insert(table_of_words, word) end
end
assert.are_same({
"Pójdźże, ",
@@ -121,7 +138,7 @@ describe("util module", function()
"gavilán",
}, table_of_words)
end)
it("should split text to line - CJK", function()
it("should split text to line - CJK Chinese", function()
local text = "彩虹是通过太阳光的折射引起的。"
local word = ""
local table_of_words = {}
@@ -134,12 +151,76 @@ describe("util module", function()
table.insert(table_of_words, word)
word = ""
end
if i == #table_chars then table.insert(table_of_words, word) end
if i == #table_chars and word ~= "" then table.insert(table_of_words, word) end
end
assert.are_same({
"","","","","","","","","","","","","","","",
}, table_of_words)
end)
it("should split text to line - CJK Japanese", function()
local text = "色は匂へど散りぬるを我が世誰ぞ常ならむ"
local word = ""
local table_of_words = {}
local c
local table_chars = util.splitToChars(text)
for i = 1, #table_chars do
c = table_chars[i]
word = word .. c
if util.isSplittable(c) then
table.insert(table_of_words, word)
word = ""
end
if i == #table_chars and word ~= "" then table.insert(table_of_words, word) end
end
assert.are_same({
"","","","","","","","","","",
"","","","","","","","","",
}, table_of_words)
end)
it("should split text to line - CJK Korean", function()
local text = "대한민국의 국기는 대한민국 국기법에 따라 태극기"
local word = ""
local table_of_words = {}
local c
local table_chars = util.splitToChars(text)
for i = 1, #table_chars do
c = table_chars[i]
word = word .. c
if util.isSplittable(c) then
table.insert(table_of_words, word)
word = ""
end
if i == #table_chars and word ~= "" then table.insert(table_of_words, word) end
end
assert.are_same({
"","","","",""," ","","",""," ",
"","","",""," ","","","",""," ",
"",""," ","","","",
}, table_of_words)
end)
it("should split text to line - mixed CJK and latin", function()
local text = "This is Russian: русский язык, Chinese: 汉语, Japanese: 日本語、 Korean: 한국어。"
local word = ""
local table_of_words = {}
local c
local table_chars = util.splitToChars(text)
for i = 1, #table_chars do
c = table_chars[i]
word = word .. c
if util.isSplittable(c) then
table.insert(table_of_words, word)
word = ""
end
if i == #table_chars and word ~= "" then table.insert(table_of_words, word) end
end
assert.are_same({
"This ", "is ",
"Russian: ", "русский ", "язык, ",
"Chinese: ", "","",", ",
"Japanese: ", "","","","", " ",
"Korean: ", "","","","",
}, table_of_words)
end)
it("should split text to line with next_c - unicode", function()
local text = "Ce test : 1) est très simple ; 2 ) simple comme ( 2/2 ) > 50 % ? ok."
local word = ""
@@ -154,7 +235,7 @@ describe("util module", function()
table.insert(table_of_words, word)
word = ""
end
if i == #table_chars then table.insert(table_of_words, word) end
if i == #table_chars and word ~= "" then table.insert(table_of_words, word) end
end
assert.are_same({
"Ce ",
@@ -187,7 +268,7 @@ describe("util module", function()
table.insert(table_of_words, word)
word = ""
end
if i == #table_chars then table.insert(table_of_words, word) end
if i == #table_chars and word ~= "" then table.insert(table_of_words, word) end
end
assert.are_same({
"Ce ",