Datasets:
sample_id stringlengths 14 14 | doc_id stringlengths 10 10 | mode stringclasses 2
values | image_width int64 112 6.8k | image_height int64 95 9.07k | bbox listlengths 4 4 | answer stringlengths 1 434 | level int64 1 4 | target_index int64 0 44 | num_context_images int64 1 55 | context_img_paths listlengths 1 55 | language stringclasses 22
values | mask_ratio float64 0 0.14 | char_length int64 1 434 | file_name stringlengths 41 43 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
sample_0000001 | doc_000956 | multi | 1,224 | 1,584 | [
0.59,
0.461,
0.885,
0.475
] | (d) Comparison between NOMA, SDMA, and OMA. | 3 | 3 | 5 | [
"images/doc_000956/doc_000956_page0001.png",
"images/doc_000956/doc_000956_page0002.png",
"images/doc_000956/doc_000956_page0003.png",
"images/doc_000956/doc_000956_page0004.png",
"images/doc_000956/doc_000956_page0005.png"
] | English | 0.004192 | 43 | images/doc_000956/doc_000956_page0004.png |
sample_0000002 | doc_001009 | multi | 1,224 | 1,584 | [
0.292,
0.369,
0.444,
0.386
] | Quarter fixed effect | 2 | 15 | 16 | [
"images/doc_001009/doc_001009_page0001.png",
"images/doc_001009/doc_001009_page0010.png",
"images/doc_001009/doc_001009_page0011.png",
"images/doc_001009/doc_001009_page0012.png",
"images/doc_001009/doc_001009_page0013.png",
"images/doc_001009/doc_001009_page0014.png",
"images/doc_001009/doc_001009_page... | English | 0.002522 | 20 | images/doc_001009/doc_001009_page0009.png |
sample_0000003 | doc_001010 | multi | 1,224 | 1,584 | [
0.254,
0.073,
0.426,
0.095
] | Attribute Classifier Training | 2 | 1 | 7 | [
"images/doc_001010/doc_001010_page0001.png",
"images/doc_001010/doc_001010_page0002.png",
"images/doc_001010/doc_001010_page0003.png",
"images/doc_001010/doc_001010_page0004.png",
"images/doc_001010/doc_001010_page0005.png",
"images/doc_001010/doc_001010_page0006.png",
"images/doc_001010/doc_001010_page... | English | 0.003814 | 29 | images/doc_001010/doc_001010_page0002.png |
sample_0000004 | doc_000913 | multi | 1,152 | 1,566 | [
0.096,
0.25,
0.246,
0.267
] | Input: paint path segments | 2 | 10 | 17 | [
"images/doc_000913/doc_000913_page0001.png",
"images/doc_000913/doc_000913_page0010.png",
"images/doc_000913/doc_000913_page0011.png",
"images/doc_000913/doc_000913_page0012.png",
"images/doc_000913/doc_000913_page0013.png",
"images/doc_000913/doc_000913_page0014.png",
"images/doc_000913/doc_000913_page... | English | 0.002602 | 26 | images/doc_000913/doc_000913_page0003.png |
sample_0000005 | doc_000941 | multi | 1,201 | 1,600 | [
0.127,
0.613,
0.377,
0.634
] | Fully connected layer with 2 classes. | 3 | 12 | 14 | [
"images/doc_000941/doc_000941_page0001.png",
"images/doc_000941/doc_000941_page0010.png",
"images/doc_000941/doc_000941_page0011.png",
"images/doc_000941/doc_000941_page0012.png",
"images/doc_000941/doc_000941_page0013.png",
"images/doc_000941/doc_000941_page0014.png",
"images/doc_000941/doc_000941_page... | English | 0.005464 | 37 | images/doc_000941/doc_000941_page0008.png |
sample_0000006 | doc_000999 | multi | 1,224 | 1,584 | [
0.074,
0.536,
0.491,
0.569
] | We now discuss how our dynamic KDE approach contributes to improving large-scale scientific data exploration in VR. | 3 | 10 | 14 | [
"images/doc_000999/doc_000999_page0001.png",
"images/doc_000999/doc_000999_page0010.png",
"images/doc_000999/doc_000999_page0011.png",
"images/doc_000999/doc_000999_page0012.png",
"images/doc_000999/doc_000999_page0013.png",
"images/doc_000999/doc_000999_page0014.png",
"images/doc_000999/doc_000999_page... | English | 0.013481 | 115 | images/doc_000999/doc_000999_page0006.png |
sample_0000007 | doc_000888 | multi | 1,191 | 1,684 | [
0.148,
0.881,
0.501,
0.902
] | Bayesian Inference of Astrophysical Parameters | 3 | 31 | 32 | [
"images/doc_000888/doc_000888_page0001.png",
"images/doc_000888/doc_000888_page0010.png",
"images/doc_000888/doc_000888_page0011.png",
"images/doc_000888/doc_000888_page0012.png",
"images/doc_000888/doc_000888_page0013.png",
"images/doc_000888/doc_000888_page0014.png",
"images/doc_000888/doc_000888_page... | English | 0.007458 | 46 | images/doc_000888/doc_000888_page0009.png |
sample_0000008 | doc_000957 | multi | 1,224 | 1,584 | [
0.525,
0.922,
0.902,
0.951
] | Fig. 2: Distribution matching for the unicycle model. | 3 | 6 | 8 | [
"images/doc_000957/doc_000957_page0001.png",
"images/doc_000957/doc_000957_page0002.png",
"images/doc_000957/doc_000957_page0003.png",
"images/doc_000957/doc_000957_page0004.png",
"images/doc_000957/doc_000957_page0005.png",
"images/doc_000957/doc_000957_page0006.png",
"images/doc_000957/doc_000957_page... | English | 0.011033 | 53 | images/doc_000957/doc_000957_page0007.png |
sample_0000009 | doc_000911 | multi | 1,224 | 1,584 | [
0.347,
0.665,
0.482,
0.688
] | High-Altitude Platforms
RU+DU | 3 | 2 | 9 | [
"images/doc_000911/doc_000911_page0001.png",
"images/doc_000911/doc_000911_page0002.png",
"images/doc_000911/doc_000911_page0003.png",
"images/doc_000911/doc_000911_page0004.png",
"images/doc_000911/doc_000911_page0005.png",
"images/doc_000911/doc_000911_page0006.png",
"images/doc_000911/doc_000911_page... | English | 0.003085 | 29 | images/doc_000911/doc_000911_page0003.png |
sample_0000010 | doc_000879 | multi | 1,224 | 1,584 | [
0.437,
0.893,
0.748,
0.919
] | (I) Left compatibility conditions | 3 | 11 | 19 | [
"images/doc_000879/doc_000879_page0001.png",
"images/doc_000879/doc_000879_page0010.png",
"images/doc_000879/doc_000879_page0011.png",
"images/doc_000879/doc_000879_page0012.png",
"images/doc_000879/doc_000879_page0013.png",
"images/doc_000879/doc_000879_page0014.png",
"images/doc_000879/doc_000879_page... | English | 0.008111 | 33 | images/doc_000879/doc_000879_page0002.png |
sample_0000011 | doc_001001 | multi | 1,224 | 1,584 | [
0.115,
0.703,
0.34,
0.725
] | Opening the left sink cabinet door | 3 | 16 | 18 | [
"images/doc_001001/doc_001001_page0001.png",
"images/doc_001001/doc_001001_page0010.png",
"images/doc_001001/doc_001001_page0011.png",
"images/doc_001001/doc_001001_page0012.png",
"images/doc_001001/doc_001001_page0013.png",
"images/doc_001001/doc_001001_page0014.png",
"images/doc_001001/doc_001001_page... | English | 0.004964 | 34 | images/doc_001001/doc_001001_page0008.png |
sample_0000012 | doc_000996 | multi | 1,224 | 1,584 | [
0.068,
0.062,
0.5,
0.082
] | For notional convenience, we define the effective channel | 3 | 2 | 5 | [
"images/doc_000996/doc_000996_page0001.png",
"images/doc_000996/doc_000996_page0002.png",
"images/doc_000996/doc_000996_page0003.png",
"images/doc_000996/doc_000996_page0004.png",
"images/doc_000996/doc_000996_page0005.png"
] | English | 0.008522 | 57 | images/doc_000996/doc_000996_page0003.png |
sample_0000013 | doc_000894 | multi | 1,191 | 1,684 | [
0.186,
0.553,
0.635,
0.572
] | Below, the Euler diagram and bar plot show that | 3 | 3 | 13 | [
"images/doc_000894/doc_000894_page0001.png",
"images/doc_000894/doc_000894_page0010.png",
"images/doc_000894/doc_000894_page0011.png",
"images/doc_000894/doc_000894_page0012.png",
"images/doc_000894/doc_000894_page0013.png",
"images/doc_000894/doc_000894_page0002.png",
"images/doc_000894/doc_000894_page... | English | 0.008411 | 47 | images/doc_000894/doc_000894_page0012.png |
sample_0000014 | doc_000925 | multi | 1,224 | 1,584 | [
0.432,
0.082,
0.56,
0.11
] | Feature extraction and classification | 3 | 1 | 7 | [
"images/doc_000925/doc_000925_page0001.png",
"images/doc_000925/doc_000925_page0002.png",
"images/doc_000925/doc_000925_page0003.png",
"images/doc_000925/doc_000925_page0004.png",
"images/doc_000925/doc_000925_page0005.png",
"images/doc_000925/doc_000925_page0006.png",
"images/doc_000925/doc_000925_page... | English | 0.003644 | 37 | images/doc_000925/doc_000925_page0002.png |
sample_0000015 | doc_001026 | multi | 1,191 | 1,684 | [
0.394,
0.726,
0.567,
0.744
] | Original Ecosystems | 2 | 12 | 16 | [
"images/doc_001026/doc_001026_page0001.png",
"images/doc_001026/doc_001026_page0010.png",
"images/doc_001026/doc_001026_page0011.png",
"images/doc_001026/doc_001026_page0012.png",
"images/doc_001026/doc_001026_page0013.png",
"images/doc_001026/doc_001026_page0014.png",
"images/doc_001026/doc_001026_page... | English | 0.003117 | 19 | images/doc_001026/doc_001026_page0006.png |
sample_0000016 | doc_000915 | multi | 1,224 | 1,584 | [
0.097,
0.386,
0.345,
0.403
] | Key nodes Set of Knowledge Graph | 3 | 13 | 18 | [
"images/doc_000915/doc_000915_page0001.png",
"images/doc_000915/doc_000915_page0010.png",
"images/doc_000915/doc_000915_page0011.png",
"images/doc_000915/doc_000915_page0012.png",
"images/doc_000915/doc_000915_page0013.png",
"images/doc_000915/doc_000915_page0014.png",
"images/doc_000915/doc_000915_page... | English | 0.004308 | 32 | images/doc_000915/doc_000915_page0005.png |
sample_0000017 | doc_000979 | multi | 1,224 | 1,584 | [
0.788,
0.262,
0.9,
0.295
] | : Upsampling Block
(Patch Expanding) | 3 | 3 | 8 | [
"images/doc_000979/doc_000979_page0001.png",
"images/doc_000979/doc_000979_page0002.png",
"images/doc_000979/doc_000979_page0003.png",
"images/doc_000979/doc_000979_page0004.png",
"images/doc_000979/doc_000979_page0005.png",
"images/doc_000979/doc_000979_page0006.png",
"images/doc_000979/doc_000979_page... | English | 0.003723 | 36 | images/doc_000979/doc_000979_page0004.png |
sample_0000018 | doc_000950 | multi | 1,224 | 1,584 | [
0.208,
0.307,
0.806,
0.345
] | Figure 1: The basic blocks of the DNN architecture. The procedure is repeated over a very large set of input data. | 4 | 11 | 17 | [
"images/doc_000950/doc_000950_page0001.png",
"images/doc_000950/doc_000950_page0010.png",
"images/doc_000950/doc_000950_page0011.png",
"images/doc_000950/doc_000950_page0012.png",
"images/doc_000950/doc_000950_page0013.png",
"images/doc_000950/doc_000950_page0014.png",
"images/doc_000950/doc_000950_page... | English | 0.023141 | 114 | images/doc_000950/doc_000950_page0004.png |
sample_0000019 | doc_001025 | multi | 1,224 | 1,584 | [
0.103,
0.869,
0.891,
0.917
] | Figure 8. Examples of temporal angle computation capturing the opening angle of the main vascular arcades. The computed value is show on the top left corner of each sample. | 4 | 3 | 17 | [
"images/doc_001025/doc_001025_page0001.png",
"images/doc_001025/doc_001025_page0010.png",
"images/doc_001025/doc_001025_page0011.png",
"images/doc_001025/doc_001025_page0012.png",
"images/doc_001025/doc_001025_page0013.png",
"images/doc_001025/doc_001025_page0014.png",
"images/doc_001025/doc_001025_page... | English | 0.037952 | 172 | images/doc_001025/doc_001025_page0012.png |
sample_0000020 | doc_001006 | multi | 1,191 | 1,684 | [
0.106,
0.763,
0.491,
0.832
] | In this section, we seek to determine the characteristics of fine-tuning with ES that cause catastrophic forgetting. We do this by analyzing two features: the update norm and sparsity. | 4 | 5 | 11 | [
"images/doc_001006/doc_001006_page0001.png",
"images/doc_001006/doc_001006_page0010.png",
"images/doc_001006/doc_001006_page0011.png",
"images/doc_001006/doc_001006_page0002.png",
"images/doc_001006/doc_001006_page0003.png",
"images/doc_001006/doc_001006_page0004.png",
"images/doc_001006/doc_001006_page... | English | 0.026812 | 184 | images/doc_001006/doc_001006_page0004.png |
sample_0000021 | doc_000914 | multi | 1,224 | 1,584 | [
0.169,
0.56,
0.371,
0.576
] | Distributed-system layer (middleware) | 3 | 2 | 10 | [
"images/doc_000914/doc_000914_page0001.png",
"images/doc_000914/doc_000914_page0010.png",
"images/doc_000914/doc_000914_page0002.png",
"images/doc_000914/doc_000914_page0003.png",
"images/doc_000914/doc_000914_page0004.png",
"images/doc_000914/doc_000914_page0005.png",
"images/doc_000914/doc_000914_page... | English | 0.003191 | 37 | images/doc_000914/doc_000914_page0002.png |
sample_0000022 | doc_000954 | multi | 1,191 | 1,684 | [
0.775,
0.236,
0.835,
0.268
] | Wireless Channels | 2 | 7 | 13 | [
"images/doc_000954/doc_000954_page0001.png",
"images/doc_000954/doc_000954_page0010.png",
"images/doc_000954/doc_000954_page0011.png",
"images/doc_000954/doc_000954_page0012.png",
"images/doc_000954/doc_000954_page0013.png",
"images/doc_000954/doc_000954_page0002.png",
"images/doc_000954/doc_000954_page... | English | 0.001911 | 17 | images/doc_000954/doc_000954_page0004.png |
sample_0000023 | doc_000934 | multi | 1,224 | 1,584 | [
0.439,
0.514,
0.747,
0.54
] | $ D(\text{Pred}(Enc(\text{Sentence1})), \text{Enc}(\text{Sentence2})) $ | 3 | 1 | 16 | [
"images/doc_000934/doc_000934_page0001.png",
"images/doc_000934/doc_000934_page0010.png",
"images/doc_000934/doc_000934_page0011.png",
"images/doc_000934/doc_000934_page0012.png",
"images/doc_000934/doc_000934_page0013.png",
"images/doc_000934/doc_000934_page0014.png",
"images/doc_000934/doc_000934_page... | English | 0.008032 | 79 | images/doc_000934/doc_000934_page0010.png |
sample_0000024 | doc_001028 | multi | 1,224 | 1,584 | [
0.169,
0.837,
0.714,
0.862
] | Table 7 provides the complete list of all 394 evaluations organized by platform. | 3 | 2 | 17 | [
"images/doc_001028/doc_001028_page0001.png",
"images/doc_001028/doc_001028_page0010.png",
"images/doc_001028/doc_001028_page0011.png",
"images/doc_001028/doc_001028_page0012.png",
"images/doc_001028/doc_001028_page0013.png",
"images/doc_001028/doc_001028_page0014.png",
"images/doc_001028/doc_001028_page... | English | 0.013771 | 80 | images/doc_001028/doc_001028_page0011.png |
sample_0000025 | doc_000923 | multi | 1,191 | 1,684 | [
0.497,
0.325,
0.952,
0.359
] | Table 2. Names used for the on-off spectra taken along the long-slit shown in Fig. 2 and their extensions. | 4 | 8 | 14 | [
"images/doc_000923/doc_000923_page0001.png",
"images/doc_000923/doc_000923_page0010.png",
"images/doc_000923/doc_000923_page0011.png",
"images/doc_000923/doc_000923_page0012.png",
"images/doc_000923/doc_000923_page0013.png",
"images/doc_000923/doc_000923_page0014.png",
"images/doc_000923/doc_000923_page... | English | 0.015293 | 106 | images/doc_000923/doc_000923_page0004.png |
sample_0000026 | doc_000962 | multi | 1,191 | 1,684 | [
0.151,
0.347,
0.337,
0.361
] | Simulated Interaction / Communication | 2 | 1 | 6 | [
"images/doc_000962/doc_000962_page0001.png",
"images/doc_000962/doc_000962_page0002.png",
"images/doc_000962/doc_000962_page0003.png",
"images/doc_000962/doc_000962_page0004.png",
"images/doc_000962/doc_000962_page0005.png",
"images/doc_000962/doc_000962_page0006.png"
] | English | 0.002645 | 37 | images/doc_000962/doc_000962_page0002.png |
sample_0000027 | doc_000951 | multi | 1,224 | 1,584 | [
0.734,
0.195,
0.832,
0.206
] | Transformation Layer | 2 | 11 | 17 | [
"images/doc_000951/doc_000951_page0001.png",
"images/doc_000951/doc_000951_page0010.png",
"images/doc_000951/doc_000951_page0011.png",
"images/doc_000951/doc_000951_page0012.png",
"images/doc_000951/doc_000951_page0013.png",
"images/doc_000951/doc_000951_page0014.png",
"images/doc_000951/doc_000951_page... | English | 0.001052 | 20 | images/doc_000951/doc_000951_page0004.png |
sample_0000028 | doc_000998 | multi | 1,191 | 1,684 | [
0.101,
0.626,
0.493,
0.679
] | These results statistically reinforce our finding that PP's utility is task and model-dependent, and often detrimental to rationale quality. | 4 | 17 | 19 | [
"images/doc_000998/doc_000998_page0001.png",
"images/doc_000998/doc_000998_page0010.png",
"images/doc_000998/doc_000998_page0011.png",
"images/doc_000998/doc_000998_page0012.png",
"images/doc_000998/doc_000998_page0013.png",
"images/doc_000998/doc_000998_page0014.png",
"images/doc_000998/doc_000998_page... | English | 0.020838 | 140 | images/doc_000998/doc_000998_page0008.png |
sample_0000029 | doc_000936 | multi | 1,224 | 1,584 | [
0.267,
0.25,
0.754,
0.274
] | Figure 1: Representative meshes from the curated artifact dataset. | 3 | 5 | 10 | [
"images/doc_000936/doc_000936_page0001.png",
"images/doc_000936/doc_000936_page0010.png",
"images/doc_000936/doc_000936_page0002.png",
"images/doc_000936/doc_000936_page0003.png",
"images/doc_000936/doc_000936_page0004.png",
"images/doc_000936/doc_000936_page0005.png",
"images/doc_000936/doc_000936_page... | English | 0.011917 | 66 | images/doc_000936/doc_000936_page0005.png |
sample_0000030 | doc_000939 | multi | 1,224 | 1,584 | [
0.501,
0.463,
0.932,
0.484
] | Fig. 2. Average semantic efficiency versus transmit power at satellites. | 3 | 3 | 5 | [
"images/doc_000939/doc_000939_page0001.png",
"images/doc_000939/doc_000939_page0002.png",
"images/doc_000939/doc_000939_page0003.png",
"images/doc_000939/doc_000939_page0004.png",
"images/doc_000939/doc_000939_page0005.png"
] | English | 0.009182 | 72 | images/doc_000939/doc_000939_page0004.png |
sample_0000031 | doc_001018 | multi | 1,684 | 2,382 | [
0.101,
0.668,
0.899,
0.765
] | We report a case of BFP secondary to trauma, which has not been previously documented. We emphasize that when encountering bilateral facial nerve paralysis following trauma, clinicians should avoid prematurely attributing the condition solely to the injury. Instead, a comprehensive history inquiry and physical examinat... | 4 | 3 | 5 | [
"images/doc_001018/doc_001018_page0001.png",
"images/doc_001018/doc_001018_page0002.png",
"images/doc_001018/doc_001018_page0003.png",
"images/doc_001018/doc_001018_page0004.png",
"images/doc_001018/doc_001018_page0005.png"
] | English | 0.077091 | 374 | images/doc_001018/doc_001018_page0004.png |
sample_0000032 | doc_001008 | multi | 1,190 | 1,684 | [
0.095,
0.555,
0.624,
0.576
] | Using the Cobb-Douglas production function in its intensive form, | 3 | 8 | 15 | [
"images/doc_001008/doc_001008_page0001.png",
"images/doc_001008/doc_001008_page0010.png",
"images/doc_001008/doc_001008_page0011.png",
"images/doc_001008/doc_001008_page0012.png",
"images/doc_001008/doc_001008_page0013.png",
"images/doc_001008/doc_001008_page0014.png",
"images/doc_001008/doc_001008_page... | English | 0.01115 | 65 | images/doc_001008/doc_001008_page0003.png |
sample_0000033 | doc_000997 | multi | 1,224 | 1,584 | [
0.494,
0.704,
0.891,
0.798
] | We propose a light-weight LLM inference profiling tool based on eBPF for mobile systems. By extracting the core functions of the software stack of llama.cpp, combined with hardware observabilities through performance monitoring counters, we derive fine-grained profiling results with three types of views to identify the... | 4 | 1 | 12 | [
"images/doc_000997/doc_000997_page0001.png",
"images/doc_000997/doc_000997_page0010.png",
"images/doc_000997/doc_000997_page0011.png",
"images/doc_000997/doc_000997_page0012.png",
"images/doc_000997/doc_000997_page0002.png",
"images/doc_000997/doc_000997_page0003.png",
"images/doc_000997/doc_000997_page... | English | 0.036993 | 345 | images/doc_000997/doc_000997_page0010.png |
sample_0000034 | doc_000970 | multi | 1,224 | 1,584 | [
0.038,
0.187,
0.822,
0.211
] | Fig. 9. Global equivalent circuit model of hybrid AC/DC/DS MG with FTS dynamic concatenator and frequency/voltage restoration control. | 4 | 13 | 14 | [
"images/doc_000970/doc_000970_page0001.png",
"images/doc_000970/doc_000970_page0010.png",
"images/doc_000970/doc_000970_page0011.png",
"images/doc_000970/doc_000970_page0012.png",
"images/doc_000970/doc_000970_page0013.png",
"images/doc_000970/doc_000970_page0014.png",
"images/doc_000970/doc_000970_page... | English | 0.019162 | 134 | images/doc_000970/doc_000970_page0009.png |
sample_0000035 | doc_000953 | multi | 1,224 | 1,584 | [
0.505,
0.575,
0.946,
0.622
] | To evaluate the effectiveness and generalizability of STEL-LAR, we consider two case studies related to two domains and three AUTs, whose configurations are summarized in Table I. | 4 | 6 | 12 | [
"images/doc_000953/doc_000953_page0001.png",
"images/doc_000953/doc_000953_page0010.png",
"images/doc_000953/doc_000953_page0011.png",
"images/doc_000953/doc_000953_page0012.png",
"images/doc_000953/doc_000953_page0002.png",
"images/doc_000953/doc_000953_page0003.png",
"images/doc_000953/doc_000953_page... | English | 0.020889 | 180 | images/doc_000953/doc_000953_page0004.png |
sample_0000036 | doc_000912 | multi | 1,152 | 1,566 | [
0.726,
0.238,
0.937,
0.257
] | (b) Detector with proposed rectifier | 3 | 10 | 18 | [
"images/doc_000912/doc_000912_page0001.png",
"images/doc_000912/doc_000912_page0010.png",
"images/doc_000912/doc_000912_page0011.png",
"images/doc_000912/doc_000912_page0012.png",
"images/doc_000912/doc_000912_page0013.png",
"images/doc_000912/doc_000912_page0014.png",
"images/doc_000912/doc_000912_page... | English | 0.003995 | 36 | images/doc_000912/doc_000912_page0002.png |
sample_0000037 | doc_000965 | multi | 1,224 | 1,584 | [
0.073,
0.559,
0.495,
0.596
] | Fig. 4. Measurement points' locations. The arrows represent the platform's orientation. | 3 | 3 | 5 | [
"images/doc_000965/doc_000965_page0001.png",
"images/doc_000965/doc_000965_page0002.png",
"images/doc_000965/doc_000965_page0003.png",
"images/doc_000965/doc_000965_page0004.png",
"images/doc_000965/doc_000965_page0005.png"
] | English | 0.015643 | 87 | images/doc_000965/doc_000965_page0004.png |
sample_0000038 | doc_000985 | multi | 1,224 | 1,584 | [
0.758,
0.509,
0.893,
0.528
] | Cutter motion dynamics | 2 | 7 | 13 | [
"images/doc_000985/doc_000985_page0001.png",
"images/doc_000985/doc_000985_page0010.png",
"images/doc_000985/doc_000985_page0011.png",
"images/doc_000985/doc_000985_page0012.png",
"images/doc_000985/doc_000985_page0013.png",
"images/doc_000985/doc_000985_page0002.png",
"images/doc_000985/doc_000985_page... | English | 0.002553 | 22 | images/doc_000985/doc_000985_page0004.png |
sample_0000039 | doc_000889 | multi | 1,224 | 1,584 | [
0.289,
0.314,
0.455,
0.346
] | helium ignites violently
and highly off-center | 3 | 6 | 14 | [
"images/doc_000889/doc_000889_page0001.png",
"images/doc_000889/doc_000889_page0010.png",
"images/doc_000889/doc_000889_page0011.png",
"images/doc_000889/doc_000889_page0012.png",
"images/doc_000889/doc_000889_page0013.png",
"images/doc_000889/doc_000889_page0014.png",
"images/doc_000889/doc_000889_page... | English | 0.005386 | 46 | images/doc_000889/doc_000889_page0002.png |
sample_0000040 | doc_000980 | multi | 1,224 | 1,584 | [
0.063,
0.23,
0.496,
0.272
] | Fig. 5: Ablation study of node importance, with the bar plot showing the distribution of the most important nodes across the radars. | 4 | 4 | 5 | [
"images/doc_000980/doc_000980_page0001.png",
"images/doc_000980/doc_000980_page0002.png",
"images/doc_000980/doc_000980_page0003.png",
"images/doc_000980/doc_000980_page0004.png",
"images/doc_000980/doc_000980_page0005.png"
] | English | 0.01811 | 132 | images/doc_000980/doc_000980_page0005.png |
sample_0000041 | doc_000993 | multi | 1,224 | 1,584 | [
0.072,
0.274,
0.495,
0.307
] | their experimental validation is presented in the subsequent sections. | 3 | 9 | 14 | [
"images/doc_000993/doc_000993_page0001.png",
"images/doc_000993/doc_000993_page0010.png",
"images/doc_000993/doc_000993_page0011.png",
"images/doc_000993/doc_000993_page0012.png",
"images/doc_000993/doc_000993_page0013.png",
"images/doc_000993/doc_000993_page0014.png",
"images/doc_000993/doc_000993_page... | English | 0.014013 | 70 | images/doc_000993/doc_000993_page0005.png |
sample_0000042 | doc_000931 | multi | 1,191 | 1,684 | [
0.507,
0.63,
0.943,
0.725
] | To explore temporal asymmetry in the torus variability for the first time, we analyzed the MIR and optical continuum variability of AGNs using NEOWISE (W1 and W2 bands) and ZTF (g-band) light curves. We compared ensemble SFs between brightening and dimming phases for subsamples divided by various properties. The follow... | 4 | 7 | 9 | [
"images/doc_000931/doc_000931_page0001.png",
"images/doc_000931/doc_000931_page0002.png",
"images/doc_000931/doc_000931_page0003.png",
"images/doc_000931/doc_000931_page0004.png",
"images/doc_000931/doc_000931_page0005.png",
"images/doc_000931/doc_000931_page0006.png",
"images/doc_000931/doc_000931_page... | English | 0.041598 | 371 | images/doc_000931/doc_000931_page0008.png |
sample_0000043 | doc_000891 | multi | 953 | 1,384 | [
0.273,
0.271,
0.564,
0.287
] | the evolution of the magnetic flux. | 3 | 3 | 8 | [
"images/doc_000891/doc_000891_page0001.png",
"images/doc_000891/doc_000891_page0002.png",
"images/doc_000891/doc_000891_page0003.png",
"images/doc_000891/doc_000891_page0004.png",
"images/doc_000891/doc_000891_page0005.png",
"images/doc_000891/doc_000891_page0006.png",
"images/doc_000891/doc_000891_page... | English | 0.004797 | 35 | images/doc_000891/doc_000891_page0004.png |
sample_0000044 | doc_001016 | multi | 1,224 | 1,584 | [
0.174,
0.848,
0.689,
0.871
] | where t is the evolution time (branch length) of the child node. | 3 | 12 | 17 | [
"images/doc_001016/doc_001016_page0001.png",
"images/doc_001016/doc_001016_page0010.png",
"images/doc_001016/doc_001016_page0011.png",
"images/doc_001016/doc_001016_page0012.png",
"images/doc_001016/doc_001016_page0013.png",
"images/doc_001016/doc_001016_page0014.png",
"images/doc_001016/doc_001016_page... | English | 0.011779 | 64 | images/doc_001016/doc_001016_page0005.png |
sample_0000045 | doc_000991 | multi | 1,224 | 1,584 | [
0.085,
0.229,
0.897,
0.266
] | Overall, as speech length increases, the attention distribution progressively moves from local (Stage 1) to global
(Stage 3) features, demonstrating that HFQ-Former adaptively adjusts its focus based on speech length. | 4 | 4 | 18 | [
"images/doc_000991/doc_000991_page0001.png",
"images/doc_000991/doc_000991_page0010.png",
"images/doc_000991/doc_000991_page0011.png",
"images/doc_000991/doc_000991_page0012.png",
"images/doc_000991/doc_000991_page0013.png",
"images/doc_000991/doc_000991_page0014.png",
"images/doc_000991/doc_000991_page... | English | 0.029472 | 217 | images/doc_000991/doc_000991_page0013.png |
sample_0000046 | doc_000930 | multi | 1,224 | 1,584 | [
0.133,
0.317,
0.854,
0.335
] | a correctly classified melanoma case in which the model's attention is focused away from | 3 | 9 | 17 | [
"images/doc_000930/doc_000930_page0001.png",
"images/doc_000930/doc_000930_page0010.png",
"images/doc_000930/doc_000930_page0011.png",
"images/doc_000930/doc_000930_page0012.png",
"images/doc_000930/doc_000930_page0013.png",
"images/doc_000930/doc_000930_page0014.png",
"images/doc_000930/doc_000930_page... | English | 0.013086 | 88 | images/doc_000930/doc_000930_page0002.png |
sample_0000047 | doc_000882 | multi | 1,224 | 1,584 | [
0.599,
0.3,
0.847,
0.321
] | Pareto Front: Why Perfection is Impossible | 3 | 2 | 5 | [
"images/doc_000882/doc_000882_page0001.png",
"images/doc_000882/doc_000882_page0002.png",
"images/doc_000882/doc_000882_page0003.png",
"images/doc_000882/doc_000882_page0004.png",
"images/doc_000882/doc_000882_page0005.png"
] | English | 0.005092 | 42 | images/doc_000882/doc_000882_page0003.png |
sample_0000048 | doc_001003 | multi | 1,191 | 1,684 | [
0.137,
0.24,
0.528,
0.26
] | Determining the percentage Research Progress. | 3 | 3 | 11 | [
"images/doc_001003/doc_001003_page0001.png",
"images/doc_001003/doc_001003_page0010.png",
"images/doc_001003/doc_001003_page0011.png",
"images/doc_001003/doc_001003_page0002.png",
"images/doc_001003/doc_001003_page0003.png",
"images/doc_001003/doc_001003_page0004.png",
"images/doc_001003/doc_001003_page... | English | 0.007945 | 45 | images/doc_001003/doc_001003_page0002.png |
sample_0000049 | doc_000958 | multi | 1,224 | 1,584 | [
0.072,
0.368,
0.494,
0.402
] | We define a new surrogate function for the relaxed algorithm as | 3 | 10 | 13 | [
"images/doc_000958/doc_000958_page0001.png",
"images/doc_000958/doc_000958_page0010.png",
"images/doc_000958/doc_000958_page0011.png",
"images/doc_000958/doc_000958_page0012.png",
"images/doc_000958/doc_000958_page0013.png",
"images/doc_000958/doc_000958_page0002.png",
"images/doc_000958/doc_000958_page... | English | 0.014312 | 63 | images/doc_000958/doc_000958_page0007.png |
sample_0000050 | doc_000886 | multi | 1,224 | 1,584 | [
0.627,
0.363,
0.815,
0.382
] | Different initial states | 2 | 5 | 8 | [
"images/doc_000886/doc_000886_page0001.png",
"images/doc_000886/doc_000886_page0002.png",
"images/doc_000886/doc_000886_page0003.png",
"images/doc_000886/doc_000886_page0004.png",
"images/doc_000886/doc_000886_page0005.png",
"images/doc_000886/doc_000886_page0006.png",
"images/doc_000886/doc_000886_page... | English | 0.003411 | 24 | images/doc_000886/doc_000886_page0006.png |
sample_0000051 | doc_000926 | multi | 1,224 | 1,584 | [
0.065,
0.4,
0.5,
0.462
] | The rest of the paper is organized as follows. Section II presents the system model, while in Section III the packet loss rate approximation is derived. Numerical results are given in Section IV, and Section V concludes. | 4 | 1 | 6 | [
"images/doc_000926/doc_000926_page0001.png",
"images/doc_000926/doc_000926_page0002.png",
"images/doc_000926/doc_000926_page0003.png",
"images/doc_000926/doc_000926_page0004.png",
"images/doc_000926/doc_000926_page0005.png",
"images/doc_000926/doc_000926_page0006.png"
] | English | 0.027122 | 220 | images/doc_000926/doc_000926_page0002.png |
sample_0000052 | doc_001020 | multi | 1,224 | 1,584 | [
0.506,
0.479,
0.919,
0.52
] | by few components, whereas high entropy reflects a more distributed, high-dimensional state. | 3 | 1 | 7 | [
"images/doc_001020/doc_001020_page0001.png",
"images/doc_001020/doc_001020_page0002.png",
"images/doc_001020/doc_001020_page0003.png",
"images/doc_001020/doc_001020_page0004.png",
"images/doc_001020/doc_001020_page0005.png",
"images/doc_001020/doc_001020_page0006.png",
"images/doc_001020/doc_001020_page... | English | 0.016605 | 92 | images/doc_001020/doc_001020_page0002.png |
sample_0000053 | doc_000921 | multi | 1,224 | 1,584 | [
0.163,
0.317,
0.483,
0.338
] | M-step (gating and expert functions): | 3 | 17 | 20 | [
"images/doc_000921/doc_000921_page0001.png",
"images/doc_000921/doc_000921_page0010.png",
"images/doc_000921/doc_000921_page0011.png",
"images/doc_000921/doc_000921_page0012.png",
"images/doc_000921/doc_000921_page0013.png",
"images/doc_000921/doc_000921_page0014.png",
"images/doc_000921/doc_000921_page... | English | 0.006811 | 37 | images/doc_000921/doc_000921_page0007.png |
sample_0000054 | doc_000928 | multi | 1,224 | 1,584 | [
0.246,
0.29,
0.38,
0.309
] | Pixel-wise Multiplication | 2 | 1 | 9 | [
"images/doc_000928/doc_000928_page0001.png",
"images/doc_000928/doc_000928_page0002.png",
"images/doc_000928/doc_000928_page0003.png",
"images/doc_000928/doc_000928_page0004.png",
"images/doc_000928/doc_000928_page0005.png",
"images/doc_000928/doc_000928_page0006.png",
"images/doc_000928/doc_000928_page... | English | 0.002538 | 25 | images/doc_000928/doc_000928_page0002.png |
sample_0000055 | doc_000963 | multi | 1,224 | 1,584 | [
0.36,
0.09,
0.581,
0.13
] | Scenario & Geometry | 2 | 6 | 13 | [
"images/doc_000963/doc_000963_page0001.png",
"images/doc_000963/doc_000963_page0010.png",
"images/doc_000963/doc_000963_page0011.png",
"images/doc_000963/doc_000963_page0012.png",
"images/doc_000963/doc_000963_page0013.png",
"images/doc_000963/doc_000963_page0002.png",
"images/doc_000963/doc_000963_page... | English | 0.008878 | 19 | images/doc_000963/doc_000963_page0003.png |
sample_0000056 | doc_000906 | multi | 1,224 | 1,584 | [
0.143,
0.377,
0.621,
0.394
] | which can be represented as the joint of a first order endotactic | 3 | 13 | 19 | [
"images/doc_000906/doc_000906_page0001.png",
"images/doc_000906/doc_000906_page0010.png",
"images/doc_000906/doc_000906_page0011.png",
"images/doc_000906/doc_000906_page0012.png",
"images/doc_000906/doc_000906_page0013.png",
"images/doc_000906/doc_000906_page0014.png",
"images/doc_000906/doc_000906_page... | English | 0.008161 | 65 | images/doc_000906/doc_000906_page0004.png |
sample_0000057 | doc_000973 | multi | 1,224 | 1,584 | [
0.348,
0.119,
0.429,
0.131
] | passive element | 2 | 1 | 6 | [
"images/doc_000973/doc_000973_page0001.png",
"images/doc_000973/doc_000973_page0002.png",
"images/doc_000973/doc_000973_page0003.png",
"images/doc_000973/doc_000973_page0004.png",
"images/doc_000973/doc_000973_page0005.png",
"images/doc_000973/doc_000973_page0006.png"
] | English | 0.001032 | 15 | images/doc_000973/doc_000973_page0002.png |
sample_0000058 | doc_000974 | multi | 1,224 | 1,584 | [
0.235,
0.203,
0.48,
0.228
] | Use the known semantic decoder to directly reconstruct privacy information | 3 | 3 | 10 | [
"images/doc_000974/doc_000974_page0001.png",
"images/doc_000974/doc_000974_page0010.png",
"images/doc_000974/doc_000974_page0002.png",
"images/doc_000974/doc_000974_page0003.png",
"images/doc_000974/doc_000974_page0004.png",
"images/doc_000974/doc_000974_page0005.png",
"images/doc_000974/doc_000974_page... | English | 0.006189 | 74 | images/doc_000974/doc_000974_page0003.png |
sample_0000059 | doc_000968 | multi | 1,224 | 1,584 | [
0.069,
0.213,
0.499,
0.247
] | all subsequent Transformer block weights), the patch embedding and the Transformer block operations are: | 3 | 6 | 10 | [
"images/doc_000968/doc_000968_page0001.png",
"images/doc_000968/doc_000968_page0010.png",
"images/doc_000968/doc_000968_page0002.png",
"images/doc_000968/doc_000968_page0003.png",
"images/doc_000968/doc_000968_page0004.png",
"images/doc_000968/doc_000968_page0005.png",
"images/doc_000968/doc_000968_page... | English | 0.014589 | 104 | images/doc_000968/doc_000968_page0006.png |
sample_0000060 | doc_000971 | multi | 1,191 | 1,684 | [
0.501,
0.395,
0.944,
0.414
] | Carrying out this procedure on the interFoam logs gives | 3 | 4 | 8 | [
"images/doc_000971/doc_000971_page0001.png",
"images/doc_000971/doc_000971_page0002.png",
"images/doc_000971/doc_000971_page0003.png",
"images/doc_000971/doc_000971_page0004.png",
"images/doc_000971/doc_000971_page0005.png",
"images/doc_000971/doc_000971_page0006.png",
"images/doc_000971/doc_000971_page... | English | 0.008308 | 55 | images/doc_000971/doc_000971_page0005.png |
sample_0000061 | doc_000969 | multi | 1,224 | 1,584 | [
0.799,
0.569,
0.907,
0.599
] | Squeeze excited feature maps | 2 | 2 | 8 | [
"images/doc_000969/doc_000969_page0001.png",
"images/doc_000969/doc_000969_page0002.png",
"images/doc_000969/doc_000969_page0003.png",
"images/doc_000969/doc_000969_page0004.png",
"images/doc_000969/doc_000969_page0005.png",
"images/doc_000969/doc_000969_page0006.png",
"images/doc_000969/doc_000969_page... | English | 0.003332 | 28 | images/doc_000969/doc_000969_page0003.png |
sample_0000062 | doc_000881 | multi | 1,224 | 1,584 | [
0.516,
0.315,
0.928,
0.387
] | Figure 3. Validation of a compact soliton core in an idealized merger. Radial density profile of the simulated core (red points) plotted against the analytic soliton solution (black line). The inner region is well described by the soliton profile. | 4 | 3 | 6 | [
"images/doc_000881/doc_000881_page0001.png",
"images/doc_000881/doc_000881_page0002.png",
"images/doc_000881/doc_000881_page0003.png",
"images/doc_000881/doc_000881_page0004.png",
"images/doc_000881/doc_000881_page0005.png",
"images/doc_000881/doc_000881_page0006.png"
] | English | 0.02988 | 247 | images/doc_000881/doc_000881_page0004.png |
sample_0000063 | doc_001024 | multi | 1,191 | 1,684 | [
0.504,
0.56,
0.921,
0.588
] | 3. Depopulation may be a Great Filter lying in front of both humanity and all intelligent life forms | 3 | 6 | 8 | [
"images/doc_001024/doc_001024_page0001.png",
"images/doc_001024/doc_001024_page0002.png",
"images/doc_001024/doc_001024_page0003.png",
"images/doc_001024/doc_001024_page0004.png",
"images/doc_001024/doc_001024_page0005.png",
"images/doc_001024/doc_001024_page0006.png",
"images/doc_001024/doc_001024_page... | English | 0.011716 | 102 | images/doc_001024/doc_001024_page0007.png |
sample_0000064 | doc_000994 | multi | 1,191 | 1,684 | [
0.513,
0.782,
0.705,
0.806
] | Taking the expectation, | 2 | 10 | 14 | [
"images/doc_000994/doc_000994_page0001.png",
"images/doc_000994/doc_000994_page0010.png",
"images/doc_000994/doc_000994_page0011.png",
"images/doc_000994/doc_000994_page0012.png",
"images/doc_000994/doc_000994_page0013.png",
"images/doc_000994/doc_000994_page0014.png",
"images/doc_000994/doc_000994_page... | English | 0.004506 | 23 | images/doc_000994/doc_000994_page0006.png |
sample_0000065 | doc_001030 | multi | 1,224 | 1,584 | [
0.095,
0.367,
0.758,
0.396
] | Diffusion metrics to identify abnormality in individual patients | 3 | 1 | 20 | [
"images/doc_001030/doc_001030_page0001.png",
"images/doc_001030/doc_001030_page0010.png",
"images/doc_001030/doc_001030_page0011.png",
"images/doc_001030/doc_001030_page0012.png",
"images/doc_001030/doc_001030_page0013.png",
"images/doc_001030/doc_001030_page0014.png",
"images/doc_001030/doc_001030_page... | English | 0.018858 | 64 | images/doc_001030/doc_001030_page0010.png |
sample_0000066 | doc_000949 | multi | 1,224 | 1,584 | [
0.07,
0.307,
0.503,
0.349
] | Fig. 9. The resonant write driver has 4 additional transistors (M5 to M8) and a shared inductor over the traditional write driver to enable energy recycling during every write operation. | 4 | 10 | 14 | [
"images/doc_000949/doc_000949_page0001.png",
"images/doc_000949/doc_000949_page0010.png",
"images/doc_000949/doc_000949_page0011.png",
"images/doc_000949/doc_000949_page0012.png",
"images/doc_000949/doc_000949_page0013.png",
"images/doc_000949/doc_000949_page0014.png",
"images/doc_000949/doc_000949_page... | English | 0.018452 | 186 | images/doc_000949/doc_000949_page0006.png |
sample_0000067 | doc_000944 | multi | 1,224 | 1,584 | [
0.24,
0.08,
0.776,
0.112
] | Table 4. Impact of backbone on object detection methods with different architectures. | 3 | 4 | 8 | [
"images/doc_000944/doc_000944_page0001.png",
"images/doc_000944/doc_000944_page0002.png",
"images/doc_000944/doc_000944_page0003.png",
"images/doc_000944/doc_000944_page0004.png",
"images/doc_000944/doc_000944_page0005.png",
"images/doc_000944/doc_000944_page0006.png",
"images/doc_000944/doc_000944_page... | English | 0.016924 | 85 | images/doc_000944/doc_000944_page0005.png |
sample_0000068 | doc_000933 | multi | 1,191 | 1,684 | [
0.546,
0.492,
0.87,
0.53
] | Understanding Translation Collapse: A Two-Fold Failure | 3 | 6 | 10 | [
"images/doc_000933/doc_000933_page0001.png",
"images/doc_000933/doc_000933_page0010.png",
"images/doc_000933/doc_000933_page0002.png",
"images/doc_000933/doc_000933_page0003.png",
"images/doc_000933/doc_000933_page0004.png",
"images/doc_000933/doc_000933_page0005.png",
"images/doc_000933/doc_000933_page... | English | 0.01214 | 54 | images/doc_000933/doc_000933_page0006.png |
sample_0000069 | doc_000976 | multi | 1,224 | 1,584 | [
0.53,
0.411,
0.913,
0.427
] | addressing significant class imbalance in failure datasets. | 3 | 1 | 7 | [
"images/doc_000976/doc_000976_page0001.png",
"images/doc_000976/doc_000976_page0002.png",
"images/doc_000976/doc_000976_page0003.png",
"images/doc_000976/doc_000976_page0004.png",
"images/doc_000976/doc_000976_page0005.png",
"images/doc_000976/doc_000976_page0006.png",
"images/doc_000976/doc_000976_page... | English | 0.006346 | 59 | images/doc_000976/doc_000976_page0002.png |
sample_0000070 | doc_000917 | multi | 1,224 | 1,584 | [
0.31,
0.075,
0.49,
0.093
] | Backbone & Evidential Network | 2 | 2 | 7 | [
"images/doc_000917/doc_000917_page0001.png",
"images/doc_000917/doc_000917_page0002.png",
"images/doc_000917/doc_000917_page0003.png",
"images/doc_000917/doc_000917_page0004.png",
"images/doc_000917/doc_000917_page0005.png",
"images/doc_000917/doc_000917_page0006.png",
"images/doc_000917/doc_000917_page... | English | 0.003322 | 29 | images/doc_000917/doc_000917_page0003.png |
sample_0000071 | doc_000929 | multi | 1,224 | 1,584 | [
0.52,
0.404,
0.926,
0.425
] | The probability that a user transmits a total of K replicas, | 3 | 2 | 6 | [
"images/doc_000929/doc_000929_page0001.png",
"images/doc_000929/doc_000929_page0002.png",
"images/doc_000929/doc_000929_page0003.png",
"images/doc_000929/doc_000929_page0004.png",
"images/doc_000929/doc_000929_page0005.png",
"images/doc_000929/doc_000929_page0006.png"
] | English | 0.008638 | 60 | images/doc_000929/doc_000929_page0003.png |
sample_0000072 | doc_000945 | multi | 1,191 | 1,684 | [
0.14,
0.759,
0.481,
0.781
] | Convergent Cross Mapping (CCM) | 2 | 9 | 15 | [
"images/doc_000945/doc_000945_page0001.png",
"images/doc_000945/doc_000945_page0010.png",
"images/doc_000945/doc_000945_page0011.png",
"images/doc_000945/doc_000945_page0012.png",
"images/doc_000945/doc_000945_page0013.png",
"images/doc_000945/doc_000945_page0014.png",
"images/doc_000945/doc_000945_page... | English | 0.007735 | 30 | images/doc_000945/doc_000945_page0004.png |
sample_0000073 | doc_000901 | multi | 1,224 | 1,584 | [
0.161,
0.083,
0.422,
0.109
] | A Redshift Prediction Plots | 2 | 10 | 11 | [
"images/doc_000901/doc_000901_page0001.png",
"images/doc_000901/doc_000901_page0010.png",
"images/doc_000901/doc_000901_page0011.png",
"images/doc_000901/doc_000901_page0002.png",
"images/doc_000901/doc_000901_page0003.png",
"images/doc_000901/doc_000901_page0004.png",
"images/doc_000901/doc_000901_page... | English | 0.006808 | 27 | images/doc_000901/doc_000901_page0009.png |
sample_0000074 | doc_000955 | multi | 1,224 | 1,584 | [
0.514,
0.322,
0.911,
0.343
] | Public Transit Time-to-Arrive (Apps Compare w/ Time-to-Drive) | 3 | 0 | 10 | [
"images/doc_000955/doc_000955_page0001.png",
"images/doc_000955/doc_000955_page0010.png",
"images/doc_000955/doc_000955_page0002.png",
"images/doc_000955/doc_000955_page0003.png",
"images/doc_000955/doc_000955_page0004.png",
"images/doc_000955/doc_000955_page0005.png",
"images/doc_000955/doc_000955_page... | English | 0.00813 | 61 | images/doc_000955/doc_000955_page0001.png |
sample_0000075 | doc_001007 | multi | 1,224 | 1,584 | [
0.119,
0.477,
0.494,
0.523
] | Workflow Consistency and Logical Order: Verify whether the agent maintains correct dependency relationships and stable task sequencing. | 3 | 2 | 4 | [
"images/doc_001007/doc_001007_page0001.png",
"images/doc_001007/doc_001007_page0002.png",
"images/doc_001007/doc_001007_page0003.png",
"images/doc_001007/doc_001007_page0004.png"
] | English | 0.017282 | 135 | images/doc_001007/doc_001007_page0003.png |
sample_0000076 | doc_001029 | multi | 1,224 | 1,584 | [
0.175,
0.662,
0.83,
0.715
] | These results indicate that AntigenLM learns a well-structured, subtype-aware latent space, supporting applications such as automated influenza surveillance and real-time strain tracking without retraining. | 4 | 1 | 20 | [
"images/doc_001029/doc_001029_page0001.png",
"images/doc_001029/doc_001029_page0010.png",
"images/doc_001029/doc_001029_page0011.png",
"images/doc_001029/doc_001029_page0012.png",
"images/doc_001029/doc_001029_page0013.png",
"images/doc_001029/doc_001029_page0014.png",
"images/doc_001029/doc_001029_page... | English | 0.035128 | 206 | images/doc_001029/doc_001029_page0010.png |
sample_0000077 | doc_000940 | multi | 1,224 | 1,584 | [
0.217,
0.132,
0.339,
0.153
] | Queue state backlog, service share | 2 | 3 | 9 | [
"images/doc_000940/doc_000940_page0001.png",
"images/doc_000940/doc_000940_page0002.png",
"images/doc_000940/doc_000940_page0003.png",
"images/doc_000940/doc_000940_page0004.png",
"images/doc_000940/doc_000940_page0005.png",
"images/doc_000940/doc_000940_page0006.png",
"images/doc_000940/doc_000940_page... | English | 0.002678 | 34 | images/doc_000940/doc_000940_page0004.png |
sample_0000078 | doc_000938 | multi | 987 | 1,401 | [
0.138,
0.32,
0.848,
0.339
] | Summary: Pathways to answer the many open questions about massive stars | 3 | 6 | 8 | [
"images/doc_000938/doc_000938_page0001.png",
"images/doc_000938/doc_000938_page0002.png",
"images/doc_000938/doc_000938_page0003.png",
"images/doc_000938/doc_000938_page0004.png",
"images/doc_000938/doc_000938_page0005.png",
"images/doc_000938/doc_000938_page0006.png",
"images/doc_000938/doc_000938_page... | English | 0.013372 | 71 | images/doc_000938/doc_000938_page0007.png |
sample_0000079 | doc_001004 | multi | 1,191 | 1,588 | [
0.067,
0.771,
0.497,
0.805
] | Lastly, we also include normalized confusion matrices for qualitative error analysis, defined as: | 3 | 1 | 17 | [
"images/doc_001004/doc_001004_page0001.png",
"images/doc_001004/doc_001004_page0010.png",
"images/doc_001004/doc_001004_page0011.png",
"images/doc_001004/doc_001004_page0012.png",
"images/doc_001004/doc_001004_page0013.png",
"images/doc_001004/doc_001004_page0014.png",
"images/doc_001004/doc_001004_page... | English | 0.014904 | 97 | images/doc_001004/doc_001004_page0010.png |
sample_0000080 | doc_000900 | multi | 1,224 | 1,584 | [
0.672,
0.217,
0.77,
0.232
] | Integrating sphere | 2 | 1 | 9 | [
"images/doc_000900/doc_000900_page0001.png",
"images/doc_000900/doc_000900_page0002.png",
"images/doc_000900/doc_000900_page0003.png",
"images/doc_000900/doc_000900_page0004.png",
"images/doc_000900/doc_000900_page0005.png",
"images/doc_000900/doc_000900_page0006.png",
"images/doc_000900/doc_000900_page... | English | 0.00147 | 18 | images/doc_000900/doc_000900_page0002.png |
sample_0000081 | doc_000946 | multi | 1,191 | 1,684 | [
0.316,
0.285,
0.764,
0.327
] | Information on increased adoption in the city and a story of an early adopter. | 3 | 16 | 19 | [
"images/doc_000946/doc_000946_page0001.png",
"images/doc_000946/doc_000946_page0010.png",
"images/doc_000946/doc_000946_page0011.png",
"images/doc_000946/doc_000946_page0012.png",
"images/doc_000946/doc_000946_page0013.png",
"images/doc_000946/doc_000946_page0014.png",
"images/doc_000946/doc_000946_page... | English | 0.018925 | 78 | images/doc_000946/doc_000946_page0007.png |
sample_0000082 | doc_000880 | multi | 1,224 | 1,584 | [
0.156,
0.667,
0.803,
0.708
] | Group convolutions to reduce parameter redundancy while maintaining
representational power. | 3 | 4 | 9 | [
"images/doc_000880/doc_000880_page0001.png",
"images/doc_000880/doc_000880_page0002.png",
"images/doc_000880/doc_000880_page0003.png",
"images/doc_000880/doc_000880_page0004.png",
"images/doc_000880/doc_000880_page0005.png",
"images/doc_000880/doc_000880_page0006.png",
"images/doc_000880/doc_000880_page... | English | 0.026569 | 91 | images/doc_000880/doc_000880_page0005.png |
sample_0000083 | doc_001000 | multi | 1,191 | 1,684 | [
0.522,
0.3,
0.644,
0.327
] | Train Memory Head
as a Policy Network | 3 | 8 | 14 | [
"images/doc_001000/doc_001000_page0001.png",
"images/doc_001000/doc_001000_page0010.png",
"images/doc_001000/doc_001000_page0011.png",
"images/doc_001000/doc_001000_page0012.png",
"images/doc_001000/doc_001000_page0013.png",
"images/doc_001000/doc_001000_page0014.png",
"images/doc_001000/doc_001000_page... | English | 0.003181 | 37 | images/doc_001000/doc_001000_page0004.png |
sample_0000084 | doc_000966 | multi | 1,224 | 1,584 | [
0.098,
0.746,
0.49,
0.803
] | This dataset serves as the foundation for the forecasting models assessed in subsequent stages. | 3 | 9 | 14 | [
"images/doc_000966/doc_000966_page0001.png",
"images/doc_000966/doc_000966_page0010.png",
"images/doc_000966/doc_000966_page0011.png",
"images/doc_000966/doc_000966_page0012.png",
"images/doc_000966/doc_000966_page0013.png",
"images/doc_000966/doc_000966_page0014.png",
"images/doc_000966/doc_000966_page... | English | 0.022282 | 95 | images/doc_000966/doc_000966_page0005.png |
sample_0000085 | doc_000920 | multi | 1,224 | 1,584 | [
0.231,
0.456,
0.765,
0.487
] | Fig. 1. Overview of the proposed model structure based on Swin-Transformer. | 3 | 8 | 14 | [
"images/doc_000920/doc_000920_page0001.png",
"images/doc_000920/doc_000920_page0010.png",
"images/doc_000920/doc_000920_page0011.png",
"images/doc_000920/doc_000920_page0012.png",
"images/doc_000920/doc_000920_page0013.png",
"images/doc_000920/doc_000920_page0014.png",
"images/doc_000920/doc_000920_page... | English | 0.016827 | 75 | images/doc_000920/doc_000920_page0004.png |
sample_0000086 | doc_000937 | multi | 1,191 | 1,588 | [
0.073,
0.456,
0.491,
0.512
] | which maps each logit to [0, 1] and stabilizes inter-branch
scale differences. We then aggregate these scores across
branches using an element-wise Noisy-OR operator: | 4 | 14 | 16 | [
"images/doc_000937/doc_000937_page0001.png",
"images/doc_000937/doc_000937_page0010.png",
"images/doc_000937/doc_000937_page0011.png",
"images/doc_000937/doc_000937_page0012.png",
"images/doc_000937/doc_000937_page0013.png",
"images/doc_000937/doc_000937_page0014.png",
"images/doc_000937/doc_000937_page... | English | 0.023345 | 166 | images/doc_000937/doc_000937_page0008.png |
sample_0000087 | doc_000903 | multi | 1,224 | 1,584 | [
0.54,
0.496,
0.897,
0.518
] | FIG. 1. Schematic diagram of Taiji constellation. | 3 | 2 | 10 | [
"images/doc_000903/doc_000903_page0001.png",
"images/doc_000903/doc_000903_page0010.png",
"images/doc_000903/doc_000903_page0002.png",
"images/doc_000903/doc_000903_page0003.png",
"images/doc_000903/doc_000903_page0004.png",
"images/doc_000903/doc_000903_page0005.png",
"images/doc_000903/doc_000903_page... | English | 0.007898 | 49 | images/doc_000903/doc_000903_page0002.png |
sample_0000088 | doc_001015 | multi | 1,224 | 1,584 | [
0.174,
0.78,
0.553,
0.799
] | Models were pretrained on large scale datasets | 3 | 14 | 19 | [
"images/doc_001015/doc_001015_page0001.png",
"images/doc_001015/doc_001015_page0010.png",
"images/doc_001015/doc_001015_page0011.png",
"images/doc_001015/doc_001015_page0012.png",
"images/doc_001015/doc_001015_page0013.png",
"images/doc_001015/doc_001015_page0014.png",
"images/doc_001015/doc_001015_page... | English | 0.007176 | 3 | images/doc_001015/doc_001015_page0005.png |
sample_0000089 | doc_000899 | multi | 1,224 | 1,584 | [
0.36,
0.416,
0.464,
0.445
] | Few-Shot Online Adaptation | 2 | 7 | 11 | [
"images/doc_000899/doc_000899_page0001.png",
"images/doc_000899/doc_000899_page0010.png",
"images/doc_000899/doc_000899_page0011.png",
"images/doc_000899/doc_000899_page0002.png",
"images/doc_000899/doc_000899_page0003.png",
"images/doc_000899/doc_000899_page0004.png",
"images/doc_000899/doc_000899_page... | English | 0.002959 | 26 | images/doc_000899/doc_000899_page0006.png |
sample_0000090 | doc_001013 | multi | 1,190 | 1,564 | [
0.531,
0.501,
0.922,
0.567
] | Graph 2 extends Graph 1 by adding gene-phenotype associations, enabling the model to learn relationships between genes and phenotypes while still lacking disease information. Note that the set of genes is always known and does not change, making this approach also inductive. | 4 | 4 | 9 | [
"images/doc_001013/doc_001013_page0001.png",
"images/doc_001013/doc_001013_page0002.png",
"images/doc_001013/doc_001013_page0003.png",
"images/doc_001013/doc_001013_page0004.png",
"images/doc_001013/doc_001013_page0005.png",
"images/doc_001013/doc_001013_page0006.png",
"images/doc_001013/doc_001013_page... | English | 0.02604 | 275 | images/doc_001013/doc_001013_page0005.png |
sample_0000091 | doc_000877 | multi | 1,224 | 1,584 | [
0.225,
0.205,
0.353,
0.231
] | Fourier Positional Encoding | 2 | 11 | 15 | [
"images/doc_000877/doc_000877_page0001.png",
"images/doc_000877/doc_000877_page0010.png",
"images/doc_000877/doc_000877_page0011.png",
"images/doc_000877/doc_000877_page0012.png",
"images/doc_000877/doc_000877_page0013.png",
"images/doc_000877/doc_000877_page0014.png",
"images/doc_000877/doc_000877_page... | English | 0.003324 | 27 | images/doc_000877/doc_000877_page0006.png |
sample_0000092 | doc_001005 | multi | 1,191 | 1,684 | [
0.544,
0.791,
0.89,
0.871
] | While it is assumed that train-test source input similarity leads to patterns of contamination, we demonstrate that a not-so-similar input may still lead to a model calling memorized text. | 4 | 0 | 14 | [
"images/doc_001005/doc_001005_page0001.png",
"images/doc_001005/doc_001005_page0010.png",
"images/doc_001005/doc_001005_page0011.png",
"images/doc_001005/doc_001005_page0012.png",
"images/doc_001005/doc_001005_page0013.png",
"images/doc_001005/doc_001005_page0014.png",
"images/doc_001005/doc_001005_page... | English | 0.027559 | 194 | images/doc_001005/doc_001005_page0001.png |
sample_0000093 | doc_000935 | multi | 1,191 | 1,684 | [
0.504,
0.697,
0.942,
0.723
] | The approach leverages a targeted-cut strategy within isolation trees, aggregated into an isolation forest, to localize suspicious regions, | 3 | 5 | 7 | [
"images/doc_000935/doc_000935_page0001.png",
"images/doc_000935/doc_000935_page0002.png",
"images/doc_000935/doc_000935_page0003.png",
"images/doc_000935/doc_000935_page0004.png",
"images/doc_000935/doc_000935_page0005.png",
"images/doc_000935/doc_000935_page0006.png",
"images/doc_000935/doc_000935_page... | English | 0.011452 | 139 | images/doc_000935/doc_000935_page0006.png |
sample_0000094 | doc_001021 | multi | 1,224 | 1,584 | [
0.509,
0.659,
0.929,
0.692
] | Similarly, social welfare is defined as the expected total population payoff under the stationary distribution, | 3 | 1 | 15 | [
"images/doc_001021/doc_001021_page0001.png",
"images/doc_001021/doc_001021_page0010.png",
"images/doc_001021/doc_001021_page0011.png",
"images/doc_001021/doc_001021_page0012.png",
"images/doc_001021/doc_001021_page0013.png",
"images/doc_001021/doc_001021_page0014.png",
"images/doc_001021/doc_001021_page... | English | 0.01358 | 111 | images/doc_001021/doc_001021_page0010.png |
sample_0000095 | doc_000884 | multi | 1,191 | 1,684 | [
0.104,
0.438,
0.647,
0.46
] | which follows from the commutativity of the diagram below | 3 | 5 | 8 | [
"images/doc_000884/doc_000884_page0001.png",
"images/doc_000884/doc_000884_page0002.png",
"images/doc_000884/doc_000884_page0003.png",
"images/doc_000884/doc_000884_page0004.png",
"images/doc_000884/doc_000884_page0005.png",
"images/doc_000884/doc_000884_page0006.png",
"images/doc_000884/doc_000884_page... | English | 0.011867 | 57 | images/doc_000884/doc_000884_page0006.png |
sample_0000096 | doc_000904 | multi | 1,224 | 1,584 | [
0.178,
0.065,
0.409,
0.085
] | Cardiac conduction pathway to ECG | 3 | 3 | 11 | [
"images/doc_000904/doc_000904_page0001.png",
"images/doc_000904/doc_000904_page0010.png",
"images/doc_000904/doc_000904_page0011.png",
"images/doc_000904/doc_000904_page0002.png",
"images/doc_000904/doc_000904_page0003.png",
"images/doc_000904/doc_000904_page0004.png",
"images/doc_000904/doc_000904_page... | English | 0.004553 | 33 | images/doc_000904/doc_000904_page0002.png |
sample_0000097 | doc_000984 | multi | 1,224 | 1,584 | [
0.495,
0.693,
0.936,
0.771
] | Scaling law is essential for improving model performance and generalization. In our research, grasping its influence on the CSI-MAE framework is key to optimization. To fully verify CSI-MAE's scalability, we conducted experiments on data and model scaling. | 4 | 4 | 6 | [
"images/doc_000984/doc_000984_page0001.png",
"images/doc_000984/doc_000984_page0002.png",
"images/doc_000984/doc_000984_page0003.png",
"images/doc_000984/doc_000984_page0004.png",
"images/doc_000984/doc_000984_page0005.png",
"images/doc_000984/doc_000984_page0006.png"
] | English | 0.034467 | 256 | images/doc_000984/doc_000984_page0005.png |
sample_0000098 | doc_001019 | multi | 1,224 | 1,584 | [
0.718,
0.338,
0.826,
0.356
] | Predicted Paratope | 2 | 6 | 12 | [
"images/doc_001019/doc_001019_page0001.png",
"images/doc_001019/doc_001019_page0010.png",
"images/doc_001019/doc_001019_page0011.png",
"images/doc_001019/doc_001019_page0012.png",
"images/doc_001019/doc_001019_page0002.png",
"images/doc_001019/doc_001019_page0003.png",
"images/doc_001019/doc_001019_page... | English | 0.001906 | 18 | images/doc_001019/doc_001019_page0004.png |
sample_0000099 | doc_000978 | multi | 1,224 | 1,584 | [
0.385,
0.272,
0.606,
0.294
] | Real-time Wireless Channel State | 2 | 10 | 15 | [
"images/doc_000978/doc_000978_page0001.png",
"images/doc_000978/doc_000978_page0010.png",
"images/doc_000978/doc_000978_page0011.png",
"images/doc_000978/doc_000978_page0012.png",
"images/doc_000978/doc_000978_page0013.png",
"images/doc_000978/doc_000978_page0014.png",
"images/doc_000978/doc_000978_page... | English | 0.004897 | 32 | images/doc_000978/doc_000978_page0005.png |
sample_0000100 | doc_001027 | multi | 1,224 | 1,584 | [
0.077,
0.674,
0.495,
0.753
] | This provides a scalable and data-driven approach to disease classification. This dual approach seeks not only to assess predictive performance, but also to contribute to a deeper understanding of the molecular underpinnings of T2D and to identify potential gene biomarkers for future research. | 4 | 3 | 11 | [
"images/doc_001027/doc_001027_page0001.png",
"images/doc_001027/doc_001027_page0010.png",
"images/doc_001027/doc_001027_page0011.png",
"images/doc_001027/doc_001027_page0002.png",
"images/doc_001027/doc_001027_page0003.png",
"images/doc_001027/doc_001027_page0004.png",
"images/doc_001027/doc_001027_page... | English | 0.032961 | 294 | images/doc_001027/doc_001027_page0002.png |
MMTR-Bench: Multimodal Masked Text Reconstruction Benchmark
π Abstract
We present MMTR-Bench (Multimodal Masked Text Reconstruction Benchmark) to evaluate native visual context reconstruction in complex multimodal inputs. Unlike traditional question-answering tasks, MMTR-Bench presents models with masked single- or multi-image inputs from diverse real-world scenarios, such as documents and webpages.
To solve the task, models must recover the hidden text by relying on the remaining layout structure, visual cues, and relevant world knowledge. By removing question-based guidance, this task challenges models to autonomously parse and reason over complex visual structures, testing their fundamental capacity for end-to-end document parsing and structured reading. The benchmark contains 2,771 test samples spanning multiple languages and varying target lengths. To fairly assess this diversity, we introduce a level-aware scoring mechanism. Extensive experiments on representative models demonstrate that MMTR-Bench remains highly challenging, particularly for sentence- and paragraph-level recovery.
π Dataset Overview
MMTR-Bench evaluates a model's ability to maintain a continuous, structured reading flow across complex multimodal layouts. The dataset is rigorously balanced across various dimensions to ensure a comprehensive evaluation of current Multimodal Large Language Models (MLLMs).
The distributions in the dataset highlight our multi-faceted evaluation strategy:
- Difficulty Level & Context Mode (a): The dataset is categorized into four distinct difficulty levels (L1 to L4), scaling from word-level completion to complex paragraph-level reconstruction. It incorporates both Single Context (single-page) and Multi Context (multi-page) scenarios, demanding robust cross-page reasoning.
- Answer Length Distribution (b): Target texts span a wide spectrum of character lengths, ensuring models are tested on both concise factual recall and extended, coherent text generation based on visual context.
- Mask Ratio Distribution (c): The proportion of masked content varies dynamically across difficulty levels, pushing the boundaries of how much missing information a model can infer purely from surrounding document structures and visual semantics.
π Leaderboard & Evaluation
The benchmark assesses models using a specialized level-aware scoring mechanism to account for the varying complexities of L1 through L4 tasks. The inclusion of explicit reasoning ("Thinking") models reveals a significant paradigm shift in how MLLMs approach visual text reconstruction.
Main Results
Note: "Think" marks models with explicit reasoning capabilities, except for variants explicitly marked as "nothink" or "Instruct". All numbers are reported as percentages.
| Models | Think | Single-page | Multi-page | L1 | L2 | L3 | L4 | Final |
|---|---|---|---|---|---|---|---|---|
| Gemini-3.1-Pro | β | 42.57 | 38.70 | 64.17 | 44.64 | 37.50 | 31.86 | 41.87 |
| GPT5.4-High | β | 41.00 | 30.98 | 57.46 | 41.20 | 35.72 | 30.92 | 39.18 |
| Gemini-3-Flash | β | 38.49 | 34.90 | 56.75 | 38.51 | 34.86 | 29.46 | 37.84 |
| GPT5.2-High | β | 36.64 | 37.62 | 51.49 | 38.61 | 34.02 | 29.42 | 36.81 |
| Doubao-Seed2-Medium | β | 37.06 | 31.96 | 52.46 | 36.10 | 33.63 | 31.28 | 36.13 |
| GPT5.2-Medium | β | 35.39 | 36.61 | 50.27 | 37.22 | 32.72 | 30.51 | 35.61 |
| Qwen3.5-397B-A17B | β | 34.67 | 30.10 | 48.39 | 34.67 | 31.46 | 26.68 | 33.84 |
| Qwen3.5-122B-A10B | β | 30.37 | 23.94 | 43.91 | 27.23 | 27.84 | 23.92 | 29.20 |
| Doubao-Seed1.6-Thinking | β | 25.50 | 23.01 | 33.81 | 22.10 | 24.74 | 25.02 | 25.04 |
| Qwen3.5-397B-A17B | 24.25 | 18.96 | 31.94 | 20.75 | 22.91 | 22.37 | 23.29 | |
| Qwen3.5-112B-A10B | 18.56 | 15.47 | 18.79 | 13.62 | 19.31 | 23.40 | 18.00 | |
| Qwen3-VL-8B-Instruct | 12.16 | 11.38 | 7.94 | 7.12 | 14.19 | 20.11 | 12.02 |
Key Observations
- The Power of Explicit Reasoning: Models utilizing a "Think" mechanism consistently outperform their standard instruction-tuned counterparts. For instance, the reasoning-enabled
Qwen3.5-397B-A17Bachieves a Final score of 33.84%, compared to 23.29% without it. This underscores the necessity of chain-of-thought processing when parsing end-to-end multimodal documents. - Multi-page Degradation: Across almost all models, performance drops significantly in the Multi-page setting compared to Single-page, highlighting a critical gap in current architectures' ability to sustain long-context visual reasoning.
- Difficulty Scaling: Performance steeply declines as the difficulty progresses from L1 (word-level) to L4 (paragraph-level). Even the leading model, Gemini-3.1-Pro, struggles at L4 (31.86%), proving that MMTR-Bench leaves ample headroom for future research in multimodal document understanding.
π How to Use
MMTR-Bench is an evaluation-only benchmark designed to test Multimodal LLMs. There is no training set.
The dataset annotations (including mask bounding boxes, ground truth answers, and image paths) are stored in the metadata file, and the images are located in the images/ directory.
1. Installation
Ensure you have the required libraries installed:
pip install datasets
2. Loading the Benchmark
from datasets import load_dataset
# Load the benchmark dataset
# Note: Hugging Face maps single data files to the 'train' split by default.
dataset = load_dataset(
"MMTR-Bench/MMTR_Bench_Dataset",
data_files="metadata.json" # or metadata.jsonl
)
benchmark_data = dataset["train"]
# Inspect the first evaluation sample
sample = benchmark_data[0]
print(f"Sample ID: {sample['sample_id']}")
print(f"Difficulty Level: L{sample['level']}")
print(f"Ground Truth Answer: {sample['answer']}")
print(f"Mask Bounding Box: {sample['bbox']}")
print(f"Target Image: {sample['image_path']}")
π Citation
If you find our benchmark, models, or data useful in your research, please consider citing our paper:
@article{mmtrbench2026,
title={MMTR-Bench: Multimodal Masked Text Reconstruction Benchmark},
author={Anonymous Authors},
journal={Under Review},
year={2026}
}
- Downloads last month
- 2,172

