Skip to content

Commit

Permalink
Adjusted tokenisation tests to account for removal of 的 from suffixes
Browse files Browse the repository at this point in the history
  • Loading branch information
andreihar committed Jun 17, 2024
1 parent bfadf5a commit 1b8b82a
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion tests/test_tokenisation.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@ def test_best_solution_tokenisation():

def test_suffix():
t = Tokeniser()
assert ['咱', '的', '食飯', '是', '誠', '好食'] == t.tokenise("咱的食飯是誠好食")
assert ['咱的', '食飯', '是', '誠', '好食'] == t.tokenise("咱的食飯是誠好食")
assert ['卯死', '矣'] == t.tokenise("卯死矣")

def test_simplified():
Expand Down

0 comments on commit 1b8b82a

Please sign in to comment.