Have a personal or library account? Click to login
Is more always better? Measuring the quality of ranking data through information entropy Cover

Is more always better? Measuring the quality of ranking data through information entropy

By: Yishan Liu,  Yu Xiao,  Xin Long and  Jun Wu  
Open Access
|Nov 2025

Figures & Tables

Figure 1.

Comparison of rank aggregation under different distributions of input ranking information. In (a), the input information is concentrated among a few objects, whereas in (b), the input information is evenly distributed.
Comparison of rank aggregation under different distributions of input ranking information. In (a), the input information is concentrated among a few objects, whereas in (b), the input information is evenly distributed.

Figure 2.

Network representation of input information with varying distributions. (a) illustrates the network corresponding to a concentrated input distribution, whereas (b) illustrates the network for an evenly distributed input.
Network representation of input information with varying distributions. (a) illustrates the network corresponding to a concentrated input distribution, whereas (b) illustrates the network for an evenly distributed input.

Figure 3.

A measurement of input ranking information quality based on degree entropy (Hd): the more evenly the information is distributed, the higher Hd; the more concentrated the information, the lower Hd.
A measurement of input ranking information quality based on degree entropy (Hd): the more evenly the information is distributed, the higher Hd; the more concentrated the information, the lower Hd.

Figure 4.

A measurement of input ranking information quality based on edge-weighted entropy (Hw): the more evenly the information is distributed, the higher Hw; the more concentrated the information, the lower Hw.
A measurement of input ranking information quality based on edge-weighted entropy (Hw): the more evenly the information is distributed, the higher Hw; the more concentrated the information, the lower Hw.

Figure 5.

Degree entropy (Hd) and edge-weighted entropy (Hw) under different distribution parameters (α) with L0=2,3 and 4 respectively, where n=20, m=1,000. The results are averaged over 100 independent trials.
Degree entropy (Hd) and edge-weighted entropy (Hw) under different distribution parameters (α) with L0=2,3 and 4 respectively, where n=20, m=1,000. The results are averaged over 100 independent trials.

Figure 6.

The relative change in Kendall’s tau-b (∆τb) under different distribution parameters (α), where n=50, mb = 1,000, mc = 2,000 and L0 = 10. Different colors represent different data generation models, e.g. green color denotes the MM. The results are averaged over 100 independent trials.
The relative change in Kendall’s tau-b (∆τb) under different distribution parameters (α), where n=50, mb = 1,000, mc = 2,000 and L0 = 10. Different colors represent different data generation models, e.g. green color denotes the MM. The results are averaged over 100 independent trials.

Figure 7.

The Kendall’s tau-b (τb) under different distribution parameters (α), where n = 50, m = 1,000, and L0 = 10. The results are averaged over 100 independent trials.
The Kendall’s tau-b (τb) under different distribution parameters (α), where n = 50, m = 1,000, and L0 = 10. The results are averaged over 100 independent trials.

Figure 8.

Degree entropy (Hd) under different distribution parameters (α) and numbers of rankings (m), where n = 50 and L0 = 10. The results are averaged over 100 independent trials.
Degree entropy (Hd) under different distribution parameters (α) and numbers of rankings (m), where n = 50 and L0 = 10. The results are averaged over 100 independent trials.

Figure 9.

Edge-weighted entropy (Hw) under different distribution parameters (α) and numbers of rankings (m), where n=50 and L0=10. The results are averaged over 100 independent trials.
Edge-weighted entropy (Hw) under different distribution parameters (α) and numbers of rankings (m), where n=50 and L0=10. The results are averaged over 100 independent trials.

Figure 10.

Computational efficiency of degree entropy (Hd) under different numbers of objects (n) and rankings (m). The results are averaged over 100 independent trials.
Computational efficiency of degree entropy (Hd) under different numbers of objects (n) and rankings (m). The results are averaged over 100 independent trials.

Figure 11.

Computational efficiency of edge-weighted entropy (Hw) under different numbers of objects (n) and rankings (m). The results are averaged over 100 independent trials.
Computational efficiency of edge-weighted entropy (Hw) under different numbers of objects (n) and rankings (m). The results are averaged over 100 independent trials.

Figure 12.

The collaborative impact of the length of input rankings (L0) and distribution parameters (α). Each cell represents Kendall’s tau-b under different combinations of L0 and α. The gradual change in color from blue to red indicates a gradual increase in Kendall’s tau-b. The results are averaged over 100 independent trials.
The collaborative impact of the length of input rankings (L0) and distribution parameters (α). Each cell represents Kendall’s tau-b under different combinations of L0 and α. The gradual change in color from blue to red indicates a gradual increase in Kendall’s tau-b. The results are averaged over 100 independent trials.

Degree entropy (Hd) and edge-weighted (HW) of the course evaluation dataset under different distribution characteristics, with n=9 and m=146_ The results are averaged over 100 independent trials_

Methodmn
153045607590105120
Hd2.1952.1912.1892.1812.1772.1732.1662.158
Hw3.5773.5663.5583.5393.5273.5173.4963.473

Spearman’s rho of the data generated by the three ranking data generation models under different values of α_ The results are averaged over 100 independent trials_

Modelα
0.001.002.003.004.005.006.007.008.00
MM0.5190.5100.5220.5320.5340.4620.5070.5350.496
P-L0.6590.6610.6240.6850.6690.6380.6770.6600.692
IA0.8070.8170.8520.9010.8950.9190.8900.9080.872

Degree entropy of the data generated by the three ranking data generation models under different values of α_ The results are averaged over 100 independent trials_

Modelα
0.001.002.003.004.005.006.007.008.00
MM3.9123.9123.8953.8123.6883.5113.3513.2113.102
P-L3.9123.9123.8953.8143.6883.5153.3543.2133.113
IA3.9123.9123.8953.8123.6853.5103.3543.2113.106

Degree entropy (Hd) and edge-weighted (Hw) of the election dataset under different distribution characteristics, with n=9 and m=28,245_ The results are averaged over 100 independent trials_

Methodmn
1,0002,0003,0004,0005,0006,0007,0008,000
Hd2.1912.1882.1832.1782.1712.1562.1182.082
Hw3.5683.5613.5483.5363.5163.4783.3833.292

Kendall’s tau of the data generated by the three ranking data generation models under different values of α_ The results are averaged over 100 independent trials_

Modelα
0.001.002.003.004.005.006.007.008.00
MM0.3930.3940.3990.4390.4170.3830.3960.3470.373
P-L0.5170.5400.4970.5640.5050.5190.5700.5510.592
IA0.7820.7730.7810.7850.7820.7750.7840.8030.795

Edge-weighted entropy of the data generated by the three ranking data generation models under different values of α_ The results are averaged over 100 independent trials_

Modelα
0.001.002.003.004.005.006.007.008.00
MM7.0986.4005.5134.9644.6434.4444.3194.2324.168
P-L7.0976.3995.5114.9654.6424.4464.3184.2324.168
IA7.0986.3995.5124.9654.6404.4454.3184.2304.168

Kendall’s tau-b between the aggregated rankings obtained by the five methods from three sets of baseline data and the ground truth rankings, where mb = 1,000, n = 50 and L0 = 10_ The results are averaged over 100 independent trials_

ModelBMDMVBMVRCG
MM0.890.890.940.930.96
P-L0.920.910.960.940.97
IA0.930.920.950.960.98
DOI: https://doi.org/10.2478/jdis-2025-0055 | Journal eISSN: 2543-683X | Journal ISSN: 2096-157X
Language: English
Page range: 105 - 131
Submitted on: Jul 1, 2025
|
Accepted on: Oct 13, 2025
|
Published on: Nov 7, 2025
In partnership with: Paradigm Publishing Services

© 2025 Yishan Liu, Yu Xiao, Xin Long, Jun Wu, published by Chinese Academy of Sciences, National Science Library
This work is licensed under the Creative Commons Attribution 4.0 License.