Have a personal or library account? Click to login
From Character to Poem: Nested Contexts and Scalar Limits of Parallelism Detection in Classical Chinese Poetry Cover

From Character to Poem: Nested Contexts and Scalar Limits of Parallelism Detection in Classical Chinese Poetry

Open Access
|Feb 2026

Figures & Tables

johd-12-488-g1.png
Figure 1

Character Model architecture.

johd-12-488-g2.png
Figure 2

Couplet Model architecture.

johd-12-488-g3.png
Figure 3

Poem-4 (left) and Poem-1 (right) Model architectures.

Table 1

Models and inference pipelines, with performance metrics (mean ± standard deviation) acquired from 100 trials.

MODEL/INFERENCEACCURACYPRECISIONRECALLF1
Character0.932 ± 0.0340.930 ± 0.0410.936 ± 0.0370.932 ± 0.032
Couplet0.947 ± 0.0190.923 ± 0.0400.979 ± 0.0210.949 ± 0.017
Char → Couplet0.881 ± 0.0260.897 ± 0.0380.864 ± 0.0460.879 ± 0.026
Poem-10.886 ± 0.0290.860 ± 0.0590.932 ± 0.0420.892 ± 0.022
Poem-1 (2 epochs)0.905 ± 0.0220.881 ± 0.0450.941 ± 0.0370.908 ± 0.019
Couplet → Poem-10.839 ± 0.0510.777 ± 0.0730.969 ± 0.0310.860 ± 0.035
Char → Poem-10.808 ± 0.0270.847 ± 0.0420.760 ± 0.0760.797 ± 0.036
Poem-40.696 ± 0.0170.742 ± 0.0550.601 ± 0.1020.655 ± 0.048
Poem-4 (2 epochs)0.735 ± 0.0220.759 ± 0.0520.686 ± 0.0770.715 ± 0.031
Poem-4 → Poem-10.659 ± 0.0270.673 ± 0.0660.674 ± 0.1440.657 ± 0.061
johd-12-488-g4.png
Figure 4

F1 Score Distribution by Target. Metrics computed against silver labels generated by the teacher model.

johd-12-488-g5.png
Figure 5

Attention distribution in a regulated poem, top layer of the fine-tuned SikuBERT classifier (Poem-1). Each heatmap is a head, each row is a couplet, each cell is a Chinese character or punctuation. Darker color indicates higher attention score from the [CLS] token. The [CLS] and [SEP] tokens have been removed for better visibility. Notice the isomorphic attention distribution in the inner couplets: in Head 1 (top left), for example, the third (parallel) couplet elicits higher attention at positions 1, 2, and 5 in both lines. Punctuation marks often serve as “attention sinks,” providing a stable anchor for information flow across layers.

DOI: https://doi.org/10.5334/johd.488 | Journal eISSN: 2059-481X
Language: English
Submitted on: Nov 24, 2025
|
Accepted on: Jan 21, 2026
|
Published on: Feb 25, 2026
Published by: Ubiquity Press
In partnership with: Paradigm Publishing Services
Publication frequency: 1 issue per year

© 2026 Maciej Kurzynski, published by Ubiquity Press
This work is licensed under the Creative Commons Attribution 4.0 License.