What happens to Wordle and Chinese idiom Knowledge Graph?

Could I create something between the human brain and the game-cheater/ruiner to make it more of fun? With Knowledge Graph?

You may have seen tweets like this in past months, where the color dots in emoji was shared in SNS randomly.

Feel free to Google Wordle first if you don’t know its meaning yet.

wordle_SNS_share

For all magics being used to solve wordle, I am impressed by Grant Sanderson, who explained us the information theory when solving wordle, in an elegent and delightful way.

I am not going create yet another wordle-solver today, instead, it’s more about an intresting variant of wordle.

To truly enjoy the fun of wordle, mostly we should be a native speaker, and it is not surprising that there is a Spanish wordle out there, and still tweets on wordle(es) are being shared literially every second now.

While for non alphabetic languages like Chineses, do we have the luck to have fun with wordling? The answer is yes, while it’s a bit different.

For the reason Chinese charactors, also called Hanzi or Han chactors are in from of Logogram, each charactor in Chinese is made up of radicals in quite different forms, each Chinese word can be 2/3/4 charactors. Most of the crossword games in Chinese are idiom, proverb or ancient poetry based: the slot to be filled are one Chinese.

Chinese Idiom Puzzle

ref: Chinese idiom puzzle

Thus, a wordle but in Chinese idiom will be quite strange, or even hilarious as Chee and JinGen discussed in this tweet thread, where you could see the candidate characters are counted in thousands: there is no chance to guess the idiom in 10 attempts!

wordle-but-in-Chinese-idioms

As one of the fun on playing wordle could be the feedback loop of guess --> result in limited attempts, while the scale of Chinese Charactor had pushed the Chinese wordle variant creators leveraging other aspects of the charactor: pronunciation.

Each charactor in Chinease is monosyllabic without exceptions, and when it comes to its pronunciation, they are all made up from two parts(initial and final, and they could be written in roman letters), which comes in tens level of possbilities.

There are bounch of Chinese wordle varients asking player to guess idiom leveraging pinyin:

While, to me, a native Chinese speaker, it’s either too hard to play with condtions of pronunciation parts(pinyin) or too easy to guess on given around 20 Chinese charactors. Then, the varient stands out here is the “handle/ๆฑ‰ๅ…œ"(Hanzi-Wordle) created by Antfu.

“Handle” introduced the tones with genius to add an extra dimension of all charactors per each guess attempt, which helped player to have more information on filtering the knowledge in the brain.

Note, for each Chinese charactor, there will be a tone in 1 of 4 tones in its pronunciation.

Let’s see what it’s like to play the “Handle”:

  • There will be 4 Chinese Charactors to be filled in 10 times of guess
  • Not only the charactor self will be colored in result:
    • For example in first line, the green “้—จ” in position 2 is correct whereas in second line, the orange “ไป“” is corret while the possition should be all but not the first slot.
  • There will be extra hints on:
    • Pinyin parts for both part1(initial) and part2(final)
      • In third line of the boxes, the green “qiao” refers to the first charactor is ponouced in “qiao” with initial:“q” and final:“iao”, although we filled the wrong charactor in the writing dimension.
      • In third line, the orange “uo” refers to there is one chacarctor in other poisition with the final part of the pinyin as “uo”.
    • Tones of the charactor:
      • In third line, the green “-” stands for the third charactor is in tone-1.

/en/resolve-wordle/handle-intro.webp

As a non-native English speaker, the way I was playing wordle is to cheating relying on helpers:

After knowing on the third letter to be “O”, I googled and got this: 5-letter-words-with-o-in-the-middle and do the searching ๐Ÿ˜.

wordle-en-play

The way to play with helpers works for me to have fun yet not ruin it by an automated cheat resolver(it’s only simulating my brain as a native-speaker!), so that I could somehow experience the same as Millions of people out there without cheating.

While for Chinese “Handle” players, from my perspective, it’s still a bit harder(to find answers in 10 guesses), and the way my wife and I were playing “Handle” when lining up at the restaurant door ended up googling: idiom list with word ‘foo’, yet still having a lot of fun.

Could I create something between the human brain and the game-cheater/ruiner to make it more of fun?

The answer is yes: a game extension as a secondary brain. For this helper/secondary brain, the solution for “handle” differenciates from the English wordle, unlike the auto-solver, similar algorithms could help on both cases:

  • In wordle(English), player searches in their brain or from a helper like the web page: 5-letter-words-with-o-in-the-middle.

  • In handle(Chinese), it’s harder to be searching based on hints like tones/initial parts of pinyin in fulltext webpage searching anymore, the reason hehind is that the multidimensional filter condtions are not indexed by normal webpages. As I mentioned, the key of the helper to be leveraged to (not ruining the game) is to be the extension of the brain, then the question is: how does our brain work on handling the knowledge of “handle”(yes, I was preparing for this pun for so long!)?

Thus, why not do it in a graph/neural network way? And here we go, let’s create a knowledge graph of Chinese idiom and see how it goes with the “handle” game.

It’s indeed entertaining to me, and I could write Graph Queries[*] by hand or via Visualization tools[**] to help explore things in this graph, because I can we’re doing the “thinking” process the similar way in our own brain, but not so well-informed.

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
# There is one Character not in 1st position, with tone-4, final part as "ai", but is not "็ˆฑ"
MATCH (char0:`character`)<-[with_char_0:with_character]-(x:idiom)-[with_pinyin_0:with_pinyin]->(pinyin_0:character_pinyin)-[:with_pinyin_part]->(final_part_0:pinyin_part{part_type: "final"})
WHERE id(final_part_0) == "ai" AND pinyin_0.character_pinyin.tone == 4 AND with_pinyin_0.position != 0 AND with_char_0.position != 0 AND id(char0) != "็ˆฑ"

# There is one Character in tone-1 not in 2nd position
MATCH (x:idiom) -[with_pinyin_1:with_pinyin]->(pinyin_1:character_pinyin)
WHERE pinyin_1.character_pinyin.tone == 1 AND with_pinyin_1.position != 1

# There is one Character with final part as "ing", not in 4th position
MATCH (x:idiom) -[with_pinyin_2:with_pinyin]->(:character_pinyin)-[:with_pinyin_part]->(final_part_2:pinyin_part{part_type: "final"})
WHERE id(final_part_2) == "ing" AND with_pinyin_2.position != 3

# The 4th Character is tone-2
MATCH (x:idiom) -[with_pinyin_3:with_pinyin]->(pinyin_3:character_pinyin)
WHERE pinyin_3.character_pinyin.tone == 2 AND with_pinyin_3.position == 3

RETURN x, count(x) as c ORDER BY c DESC
  • **

studio_en

If you are intrested in how you could do it from your side, here are some details. Before getting started, if you would like to hands-on do it, please refer to appendix chapter on getting your own Nebula Graph Database cluster ready, which are a couple of lines commands to be run.

So, if we visit the “handle” game ๐Ÿ‘‰๐Ÿป https://handle.antfu.me/. We could start with one guess i.e. “็ˆฑๆ†Žๅˆ†ๆ˜Ž”.

We will get hint of the first guess as:

/en/resolve-wordle/initial_word.webp

Not bad, we have three of the charactor with informative hints!

  • There is one Character not in 1st position, with tone-4, final part as “ai”, but is not “็ˆฑ”
  • There is one Character in tone-1 not in 2nd position
  • There is one Character with final part as “ing”, not in 4th position
  • The 4th Character is tone-2

Then we just query it from Nebula Graph:

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
# There is one Character not in 1st position, with tone-4, final part as "ai", but is not "็ˆฑ"
MATCH (char0:`character`)<-[with_char_0:with_character]-(x:idiom)-[with_pinyin_0:with_pinyin]->(pinyin_0:character_pinyin)-[:with_pinyin_part]->(final_part_0:pinyin_part{part_type: "final"})
WHERE id(final_part_0) == "ai" AND pinyin_0.character_pinyin.tone == 4 AND with_pinyin_0.position != 0 AND with_char_0.position != 0 AND id(char0) != "็ˆฑ"

# There is one Character in tone-1 not in 2nd position
MATCH (x:idiom) -[with_pinyin_1:with_pinyin]->(pinyin_1:character_pinyin)
WHERE pinyin_1.character_pinyin.tone == 1 AND with_pinyin_1.position != 1

# There is one Character with final part as "ing", not in 4th position
MATCH (x:idiom) -[with_pinyin_2:with_pinyin]->(:character_pinyin)-[:with_pinyin_part]->(final_part_2:pinyin_part{part_type: "final"})
WHERE id(final_part_2) == "ing" AND with_pinyin_2.position != 3

# The 4th Character is tone-2
MATCH (x:idiom) -[with_pinyin_3:with_pinyin]->(pinyin_3:character_pinyin)
WHERE pinyin_3.character_pinyin.tone == 2 AND with_pinyin_3.position == 3

RETURN x, count(x) as c ORDER BY c DESC

And there are 7 results left in this single guess for us!

1
2
3
4
5
6
7
("ๆƒŠๆ„š้ช‡ไฟ—" :idiom{pinyin: "['jing1', 'yu2', 'hai4', 'su2']"})
("ๆƒŠไธ–้ช‡ไฟ—" :idiom{pinyin: "['jing1', 'shi4', 'hai4', 'su2']"})
("ๆƒŠ่ง้ช‡้—ป" :idiom{pinyin: "['jing1', 'jian4', 'hai4', 'wen2']"})
("ๆฒฝๅๅ–็›ด" :idiom{pinyin: "['gu1', 'ming2', 'mai4', 'zhi2']"})
("ๆƒŠๅฟƒ้ช‡็ฅž" :idiom{pinyin: "['jing1', 'xin1', 'hai4', 'shen2']"})
("่†ๆฃ˜่ฝฝ้€”" :idiom{pinyin: "['jing1', 'ji2', 'zai4', 'tu2']"})
("ๅ‡บๅ–็ต้ญ‚" :idiom{pinyin: "['chu1', 'mai4', 'ling2', 'hun2']"})

And as a Chinese native speaker, the common one would be “ๆƒŠไธ–้ช‡ไฟ—”, and let’s give a try!

ๆฑ‰ๅ…œ D59

We could modify our query to make every part of the subquery returned, thus, they could also be renderred visually:

 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
13
14
15
16
17
# There is one Character not in 1st position, with tone-4, final part as "ai", but is not "็ˆฑ"
MATCH p0=(char0:`character`)<-[with_char_0:with_character]-(x:idiom)-[with_pinyin_0:with_pinyin]->(pinyin_0:character_pinyin)-[:with_pinyin_part]->(final_part_0:pinyin_part{part_type: "final"})
WHERE id(final_part_0) == "ai" AND pinyin_0.character_pinyin.tone == 4 AND with_pinyin_0.position != 0 AND with_char_0.position != 0 AND id(char0) != "็ˆฑ"

# There is one Character in tone-1 not in 2nd position
MATCH p1=(x:idiom) -[with_pinyin_1:with_pinyin]->(pinyin_1:character_pinyin)
WHERE pinyin_1.character_pinyin.tone == 1 AND with_pinyin_1.position != 1

# There is one Character with final part as "ing", not in 4th position
MATCH p2=(x:idiom) -[with_pinyin_2:with_pinyin]->(:character_pinyin)-[:with_pinyin_part]->(final_part_2:pinyin_part{part_type: "final"})
WHERE id(final_part_2) == "ing" AND with_pinyin_2.position != 3

# The 4th Character is tone-2
MATCH p3=(x:idiom) -[with_pinyin_3:with_pinyin]->(pinyin_3:character_pinyin)
WHERE pinyin_3.character_pinyin.tone == 2 AND with_pinyin_3.position == 3

RETURN p0,p1,p2,p3

And then we query it in Nebula-Studio-Console, and click “View Subgraphs” to see how the searching was done like it was in our own brains:

studio_en

If you happened to get hands dirty(or being intrested to) on Graph Database, you could checkout the Nebula Graph project now and its Docs to have more fun of it! Also, here are some only playgrounds if you prefer to try it without deployment on your own envrioment.

If you are intrested in the MATCH query syntax and would like to actually do exersices with each daily handle challenge, check below Documents:

Happy Graphing!

I put the code and process here: https://github.com/wey-gu/chinese-graph, feel free to check that out.

The process would be:

  • Modeling the Knowledge
  • Preprocessing the data
  • ETL data to a Graph Database: Nebula Graph
  • Have fun on Nebula Graph

For Graph Modeling, it’s actually quite straight forward, the mind model for me was to put the knowledge I cares as vertcies and connect them with their relationships first.

You will come back to iterate or optimize the modeling when you are actually playing with the data afterwards, thus, if you could imagine how the graph will be queried in the first place, the graph modeling could be adopted accordingly. Otherwise, don’t over design it, just do it the intuitive way.

Here, I put the vertices with properties as:

  • idiom
  • character
  • pinyin
    • tone
  • pinyin_part
    • type

The edges with properteis as:

  • with_character
  • with_pinyin
  • with_pinyin_part

schema_v1_en

With Nebula-UP, it’s an onliner call

1
curl -fsSL nebula-up.siwei.io/install.sh | bash -s -- v3.0.0
 1
 2
 3
 4
 5
 6
 7
 8
 9
10
11
12
# clone the code for data genration and data loading
git clone https://github.com/wey-gu/chinese-graph.git && cd chinese-graph

python3 graph_data_generator.py # generate data

# load data with Nebula-Importer
docker run --rm -ti \
    --network=nebula-docker-compose_nebula-net \
    -v ${PWD}/importer_conf.yaml:/root/importer_conf.yaml \
    -v ${PWD}/output:/root \
    vesoft/nebula-importer:v3.0.0 \
    --config /root/importer_conf.yaml

Related Content