File size: 97,148 Bytes
40d081b 80e203f dc736c1 80e203f baa9ad4 80e203f dc736c1 80e203f 40d081b 80e203f dc736c1 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b b7f8c04 40d081b 80e203f dc736c1 80e203f 40d081b 80e203f 40d081b b7f8c04 40d081b b7f8c04 40d081b b7f8c04 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b dc736c1 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b dc736c1 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 661a70d 80e203f 40d081b 80e203f 40d081b 80e203f dc736c1 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f dc736c1 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f dc736c1 80e203f 40d081b 80e203f 8a11f9b 80e203f 8a11f9b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f dc736c1 80e203f dc736c1 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f dc736c1 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 8a11f9b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 80e203f 40d081b 8a11f9b 80e203f 40d081b 8a11f9b 40d081b 8a11f9b 367c2bc 80e203f 367c2bc 80e203f 367c2bc 80e203f 367c2bc 80e203f dc736c1 367c2bc 80e203f 367c2bc 661a70d 367c2bc 80e203f 367c2bc 80e203f dc736c1 80e203f 367c2bc 80e203f 661a70d 80e203f 367c2bc d63d185 80e203f dc736c1 80e203f 367c2bc 80e203f 661a70d 80e203f 367c2bc 80e203f 367c2bc 80e203f 367c2bc 80e203f 661a70d dc736c1 80e203f 661a70d dc736c1 661a70d 80e203f 367c2bc 80e203f dc736c1 80e203f 367c2bc 80e203f 661a70d 80e203f 661a70d 80e203f 367c2bc 80e203f 367c2bc 80e203f 367c2bc 80e203f 661a70d 80e203f 661a70d 80e203f 661a70d 80e203f 367c2bc 80e203f 367c2bc dc736c1 80e203f 367c2bc 80e203f 661a70d 80e203f 367c2bc 80e203f 367c2bc 80e203f 367c2bc 80e203f 661a70d 80e203f 661a70d 80e203f 661a70d 80e203f 367c2bc 80e203f 367c2bc 80e203f dc736c1 80e203f 367c2bc 80e203f 661a70d 80e203f 367c2bc 80e203f 661a70d 80e203f 661a70d 80e203f 661a70d 80e203f 367c2bc 80e203f 367c2bc 80e203f dc736c1 80e203f 661a70d 367c2bc 80e203f 661a70d 80e203f 661a70d 80e203f 367c2bc 80e203f 367c2bc 80e203f 367c2bc 80e203f 367c2bc dc736c1 367c2bc 80e203f 367c2bc 661a70d 367c2bc 80e203f 367c2bc 80e203f 661a70d 80e203f 661a70d 80e203f 661a70d 80e203f 22d58cf 80e203f dc736c1 80e203f 661a70d 22d58cf 80e203f 661a70d 80e203f 22d58cf 80e203f 661a70d 80e203f 661a70d 80e203f 22d58cf 80e203f 22d58cf 80e203f dc736c1 80e203f 22d58cf 80e203f 661a70d 80e203f 22d58cf 80e203f 661a70d 80e203f 661a70d 80e203f 22d58cf 80e203f 22d58cf 80e203f 661a70d 80e203f 661a70d 80e203f 661a70d 80e203f 22d58cf 80e203f 22d58cf dc736c1 80e203f 22d58cf 80e203f 661a70d 80e203f 661a70d 80e203f 661a70d 80e203f 77dc289 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438 1439 1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 | # Comprehensive Test Cases for SciVisAgentBench Main Tasks
# This test evaluates the ability to complete specific visualization tasks
# with detailed requirements and evaluation criteria
# Case 1: ABC
- vars:
question: |
Load the ABC (Arnold-Beltrami-Childress) flow vector field from "ABC/data/ABC_128x128x128_float32_scalar3.raw", the information about this dataset:
ABC Flow (Vector)
Data Scalar Type: float
Data Byte Order: Little Endian
Data Extent: 128x128x128
Number of Scalar Components: 3
Data loading is very important, make sure you correctly load the dataset according to their features.
Create streamlines using a "Stream Tracer" filter with "Point Cloud" seed type. Set the seed center to [73.77, 63.25, 71.65], with 150 seed points and a radius of 75.0. Set integration direction to "BOTH" and maximum streamline length to 150.0.
Add a "Tube" filter on the stream tracer to enhance visualization. Set tube radius to 0.57 with 12 sides.
Color the tubes by Vorticity magnitude using the 'Cool to Warm (Diverging)' colormap.
Show the dataset bounding box as an outline.
Use a white background. Render at 1024x1024.
Set the viewpoint parameters as: [-150.99, 391.75, 219.64] to position; [32.38, 120.41, 81.63] to focal point; [0.23, -0.31, 0.92] to camera up direction.
Save the visualization image as "ABC/results/{agent_mode}/ABC.png".
(Optional, but must save if use paraview) Save the paraview state as "ABC/results/{agent_mode}/ABC.pvsm".
(Optional, but must save if use python script) Save the python script as "ABC/results/{agent_mode}/ABC.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Streamline Density: Are the streamline tubes densely distributed throughout the volume, similar to the ground truth?
2. Color Mapping: Are the tubes colored by vorticity magnitude using a blue-white-red diverging colormap, with a similar color distribution as the ground truth?
3. Tube Appearance: Do the streamline tubes have a similar thickness and smooth appearance as the ground truth?
# Case 2: argon-bubble
- vars:
question: |
Task:
Load the Argon Bubble dataset from "argon-bubble/data/argon-bubble_128x128x256_float32.vtk".
Generate a visualization image of the Argon Bubble scalar field dataset with the following visualization settings:
1) Create volume rendering
2) Set the opacity transfer function as a ramp function across values of the volumetric data, assigning opacity 0 to value 0 and assigning opacity 1 to value 1.
3) Set the color transfer function to assign a warm red color [0.71, 0.02, 0.15] to the highest value, a cool color [0.23, 0.29, 0.75] to the lowest value, and a grey color[0.87, 0.87, 0.87] to the midrange value
4) Set the viewpoint parameters as: [0, 450, 0] to position; [0, 0, -15] to focal point; [0, 0, -1] to camera up direction
5) Visualization image resolution is 1024x1024. White background. Shade turned off. Volume rendering ray casting sample distance is 0.1
6) Don't show color/scalar bar or coordinate axes.
Save the visualization image as "argon-bubble/results/{agent_mode}/argon-bubble.png".
(Optional, but must save if use paraview) Save the paraview state as "argon-bubble/results/{agent_mode}/argon-bubble.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "argon-bubble/results/{agent_mode}/argon-bubble.py".
(Optional, but must save if use VTK) Save the cxx code script as "argon-bubble/results/{agent_mode}/argon-bubble.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Does the visualization image clearly show the regions of cool, warm, and mild regions?
2. Does the blueish region show areas with low opacity?
3. Does the reddish region show areas with high opacity?
# Case 3: Bernard
- vars:
question: |
Load the Rayleigh-Benard convection vector field from "Bernard/data/Bernard_128x32x64_float32_scalar3.raw", the information about this dataset:
Rayleigh-Benard Convection (Vector)
Data Scalar Type: float
Data Byte Order: Little Endian
Data Extent: 128x32x64
Number of Scalar Components: 3
Data loading is very important, make sure you correctly load the dataset according to their features.
Create four streamline sets using "Stream Tracer" filters with "Point Cloud" seed type, each with 100 seed points and radius 12.7:
- Streamline 1: Seed center at [30.69, 14.61, 47.99]. Apply a "Tube" filter (radius 0.3, 12 sides). Color with solid blue (RGB: 0.0, 0.67, 1.0).
- Streamline 2: Seed center at [91.10, 14.65, 45.70]. Apply a "Tube" filter (radius 0.3, 12 sides). Color with solid orange (RGB: 1.0, 0.33, 0.0).
- Streamline 3: Seed center at [31.87, 12.76, 15.89]. Apply a "Tube" filter (radius 0.3, 12 sides). Color by velocity magnitude using the 'Cool to Warm (Diverging)' colormap.
- Streamline 4: Seed center at [92.09, 10.50, 15.32]. Apply a "Tube" filter (radius 0.3, 12 sides). Color with solid green (RGB: 0.33, 0.67, 0.0).
In the pipeline browser panel, hide all stream tracers and only show the tube filters.
Use a white background. Render at 1280x1280. Do not show a color bar.
Set the viewpoint parameters as: [-81.99, -141.45, 89.86] to position; [65.58, 26.29, 28.48] to focal point; [0.18, 0.20, 0.96] to camera up direction.
Save the visualization image as "Bernard/results/{agent_mode}/Bernard.png".
(Optional, but must save if use paraview) Save the paraview state as "Bernard/results/{agent_mode}/Bernard.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "Bernard/results/{agent_mode}/Bernard.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Streamline Grouping: Are there four visually separate streamline clusters arranged in a 2x2 grid pattern, similar to the ground truth?
2. Color Assignment: Are the four streamline groups colored in distinct colors (blue, orange, magnitude-mapped, and green), matching the ground truth color scheme?
3. Convection Cell Structure: Do the streamlines within each group show circular or helical looping patterns characteristic of convection cells?
# Case 4: bonsai
- vars:
question: |
Task:
Load the bonsai dataset from "bonsai/data/bonsai_256x256x256_uint8.raw", the information about this dataset:
Bonsai (Scalar)
Data Scalar Type: unsigned char
Data Byte Order: little Endian
Data Spacing: 1x1x1
Data Extent: 256x256x256
Then visualize it with volume rendering, modify the transfer function and reach the visualization goal as: "A potted tree with brown pot silver branch and golden leaves."
Please think step by step and make sure to fulfill all the visualization goals mentioned above.
Use a white background. Render at 1280x1280. Do not show a color bar or coordinate axes.
Set the viewpoint parameters as: [-765.09, 413.55, 487.84] to position; [-22.76, 153.30, 157.32] to focal point; [0.30, 0.95, -0.07] to camera up direction.
Save the visualization image as "bonsai/results/{agent_mode}/bonsai.png".
(Optional, but must save if use paraview) Save the paraview state as "bonsai/results/{agent_mode}/bonsai.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "bonsai/results/{agent_mode}/bonsai.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Visualization Goal: How well does the result achieve the overall goal of showing a potted tree with the specified colors?
2. Brown Pot Visualization: Does the result show the pot portion in brown color?
3. Silver Branch Visualization: Does the result show the branch/trunk portion in silver color?
4. Golden Leaves Visualization: Does the result show the leaves portion in golden color?
# Case 5: carp
- vars:
question: |
Task:
Load the carp dataset from "carp/data/carp_256x256x512_uint16.raw", the information about this dataset:
Carp (Scalar)
Data Scalar Type: unsigned short
Data Byte Order: little Endian
Data Spacing: 0.78125x0.390625x1
Data Extent: 256x256x512
Instructions:
1. Load the dataset into ParaView.
2. Apply volume rendering to visualize the carp skeleton.
3. Adjust the transfer function to highlight only the bony structures with the original bone color.
4. Optimize the viewpoint to display the full skeleton, ensuring the head, spine, and fins are all clearly visible in a single frame.
5. Analyze the visualization and answer the following questions:
Q1: Which of the following options correctly describes the fins visible in the carp skeleton visualization?
A. 5 fins: 1 dorsal, 2 pectoral, 2 pelvic
B. 6 fins: 1 dorsal, 2 pectoral, 2 pelvic, 1 caudal
C. 7 fins: 1 dorsal, 2 pectoral, 2 pelvic, 1 anal, 1 caudal
D. 8 fins: 2 dorsal, 2 pectoral, 2 pelvic, 1 anal, 1 caudal
Q2: Based on the visualization, what is the approximate ratio of skull length to total body length?
A. ~15%
B. ~22%
C. ~30%
D. ~40%
6. Use a white background. Find an optimal view. Render at 1280x1280. Do not show a color bar or coordinate axes.
7. Set the viewpoint parameters as: [265.81, 1024.69, 131.23] to position; [141.24, 216.61, 243.16] to focal point; [0.99, -0.14, 0.07] to camera up direction.
8. Save your work:
Save the visualization image as "carp/results/{agent_mode}/carp.png".
Save the answers to the analysis questions in plain text as "carp/results/{agent_mode}/answers.txt".
(Optional, but must save if use paraview) Save the paraview state as "carp/results/{agent_mode}/carp.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "carp/results/{agent_mode}/carp.py".
(Optional, but must save if use VTK) Save the cxx code script as "carp/results/{agent_mode}/carp.cxx"
Do not save any other files, and always save the visualization image and the text file.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Visualization Goal: Does the result match the ground truth visualization of the carp skeleton?
2. Bone Visibility: Are the bones clearly visible, similar to the ground truth? Are thin fin rays distinguishable?
3. Skeletal Structure: Is the entire carp skeleton (head, spine, ribs, fins, tail) visible and similar in appearance to the ground truth?
- type: llm-rubric
subtype: text
value: |
1. Q1 correct answer: C. 7 fins: 1 dorsal, 2 pectoral, 2 pelvic, 1 anal, 1 caudal
2. Q2 correct answer: B. ~22%
# Case 6: chameleon_isosurface
- vars:
question: |
Task:
Load the chameleon dataset from "chameleon_isosurface/data/chameleon_isosurface_256x256x256_float32.vtk".
Generate a visualization image of 2 isosurfaces of the Chameleon scalar field dataset with the following visualization settings:
1) Create isosurfaces of Iso_1 with a value of 0.12 and Iso_2 with a value of 0.45
2) Assign RGB color of [0.0, 1.0, 0.0] to Iso_1, and color of [1.0, 1.0, 1.0] to Iso_2
3) Assign opacity of 0.1 to Iso_1, and opacity of 0.99 to Iso_2
4) Set the lighting parameter as: 0.1 to Ambient; 0.7 to Diffuse; 0.6 to Specular
5) Set the viewpoint parameters as: [600, 0, 0] to position; [0, 0, 0] to focal point; [0, -1, 0] to camera up direction
6) White background
7) Visualization image resolution is 1024x1024
8) Don't show color/scalar bar or coordinate axes.
Save the visualization image as "chameleon_isosurface/results/{agent_mode}/chameleon_isosurface.png".
(Optional, but must save if use paraview) Save the paraview state as "chameleon_isosurface/results/{agent_mode}/chameleon_isosurface.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "chameleon_isosurface/results/{agent_mode}/chameleon_isosurface.py".
(Optional, but must save if use VTK) Save the cxx code script as "chameleon_isosurface/results/{agent_mode}/chameleon_isosurface.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Does the result present 2 isosurfaces, one showing the inner part of the chameleon and one showing the outer part of the chameleon?
2. Is the skin of the Chameleon object of green color?
3. Is the bone of the Chameleon object of white color?
# Case 7: crayfish_streamline
- vars:
question: |
Load the Crayfish flow vector field from "crayfish_streamline/data/crayfish_streamline_322x162x119_float32_scalar3.raw", the information about this dataset:
Crayfish Flow (Vector)
Data Scalar Type: float
Data Byte Order: Little Endian
Data Extent: 322x162x119
Number of Scalar Components: 3
Data loading is very important, make sure you correctly load the dataset according to their features.
Create two streamline sets using "Stream Tracer" filters with "Point Cloud" seed type, each with 100 seed points and radius 32.2:
- Streamline 1: Seed center at [107.33, 81.0, 59.5]. Apply a "Tube" filter (radius 0.5, 12 sides). Color by Vorticity magnitude using a diverging colormap with the following RGB control points:
- Value 0.0 -> RGB(0.231, 0.298, 0.753) (blue)
- Value 0.02 -> RGB(0.865, 0.865, 0.865) (white)
- Value 0.15 -> RGB(0.706, 0.016, 0.149) (red)
- Streamline 2: Seed center at [214.67, 81.0, 59.5]. Apply a "Tube" filter (radius 0.5, 12 sides). Color by Vorticity magnitude using the same colormap.
Show the dataset bounding box as an outline (black).
In the pipeline browser panel, hide all stream tracers and only show the tube filters and the outline.
Use a white background. Render at 1280x1280.
Set the viewpoint parameters as: [435.04, -325.38, 567.82] to position; [111.64, 202.81, -21.96] to focal point; [-0.099, 0.714, 0.693] to camera up direction.
Save the paraview state as "crayfish_streamline/results/{agent_mode}/crayfish_streamline.pvsm".
Save the visualization image as "crayfish_streamline/results/{agent_mode}/crayfish_streamline.png".
(Optional, if use python script) Save the python script as "crayfish_streamline/results/{agent_mode}/crayfish_streamline.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Visualization Goal: Does the result show streamline tubes colored by vorticity magnitude within a rectangular bounding box, similar to the ground truth?
2. Streamline Clusters: Are there two distinct clusters matches the ground truth layout?
3. Color Mapping: Are the tubes colored by vorticity magnitude using a blue-white-red diverging colormap, with a similar distribution as the ground truth?
# Case 8: engine
- vars:
question: |
Task:
Load the vortex dataset from "engine/data/engine_256x256x128_uint8.raw", the information about this dataset:
engine (Scalar)
Data Scalar Type: float
Data Byte Order: little Endian
Data Extent: 256x256x128
Number of Scalar Components: 1
Instructions:
1. Load the dataset into ParaView.
2. Apply the volume rendering to visualize the engine dataset
3. Adjust the transfer function, let the outer part more transparent and the inner part more solid. Use light blue for the outer part and orange for the inner part.
4. Use a white background. Find an optimal view. Render at 1280x1280. Do not show a color bar or coordinate axes.
5. Set the viewpoint parameters as: [-184.58, 109.48, -431.72] to position; [134.05, 105.62, 88.92] to focal point; [0.01, 1.0, -0.001] to camera up direction.
6. Save your work:
Save the visualization image as "engine/results/{agent_mode}/engine.png".
(Optional, but must save if use paraview) Save the paraview state as "engine/results/{agent_mode}/engine.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "engine/results/{agent_mode}/engine.py".
(Optional, but must save if use VTK) Save the cxx code script as "engine/results/{agent_mode}/engine.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Visualization Goal: How well does the result use volume rendering to clearly present the internal and external structures of the engine dataset?
2. Structural Clarity: Does the visualization emphasize depth so that the outer layers do not obscure the inner structures?
3. Transfer Function Transparency: Is the outer region rendered with higher transparency and the inner region more solid, achieving a clear layering effect?
4. Transfer Function Color Mapping: Are colors correctly assigned so that the outer part is light blue and the inner part is orange, enhancing structural contrast?
# Case 9: foot
- vars:
question: |
Task:
Load the Foot dataset from "foot/data/foot_256x256x256_uint8.raw", the information about this dataset:
Foot
Description: Rotational C-arm x-ray scan of a human foot. Tissue and bone are present in the dataset.
Data Type: uint8
Data Byte Order: little Endian
Data Spacing: 1x1x1
Data Extent: 256x256x256
Data loading is very important, make sure you correctly load the dataset according to their features.
Visualize the anatomical structures:
1. Apply volume rendering with an X-ray transfer function that distinguishes soft tissues and bones. Bones with darker color, and soft tissue with lighter color.
2. Analyze the visualization and answer the following questions:
Q1: Based on the X-ray style volume rendering of the foot dataset, which of the following best describes the visibility of bony structures?
A. Both the phalanges and metatarsals are fully visible
B. The phalanges are fully visible, but the metatarsals are only partially visible
C. The metatarsals are fully visible, but the phalanges are only partially visible
D. Neither the phalanges nor the metatarsals are clearly visible
3. Use a white background. Find an optimal view. Render at 1280x1280. Do not show a color bar or coordinate axes.
4. Set the viewpoint parameters as: [-576.41, -264.41, -153.48] to position; [127.5, 127.5, 127.5] to focal point; [-0.52, 0.38, 0.76] to camera up direction.
5. Save your work:
Save the visualization image as "foot/results/{agent_mode}/foot.png".
Save the answers to the analysis questions in plain text as "foot/results/{agent_mode}/answers.txt".
(Optional, but must save if use paraview) Save the paraview state as "foot/results/{agent_mode}/foot.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "foot/results/{agent_mode}/foot.py".
(Optional, but must save if use VTK) Save the cxx code script as "foot/results/{agent_mode}/foot.cxx"
Do not save any other files, and always save the visualization image and the text file.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Goal: Does the visualization effectively distinguish between different tissue types in the foot dataset?
2. X-ray Appearance: Does the visualization resemble an X-ray (monochrome or grayscale, transparent look, consistent lighting)?
- type: llm-rubric
subtype: text
value: |
1. Q1 correct answer: B. The phalanges are fully visible, but the metatarsals are only partially visible
# Case 10: lobster
- vars:
question: |
Task:
Load the Lobster dataset from "lobster/data/lobster_301x324x56_uint8.raw", the information about this dataset:
Lobster
Description: CT scan of a lobster contained in a block of resin.
Data Type: uint8
Data Byte Order: little Endian
Data Spacing: 1x1x1.4
Data Extent: 301x324x56
Data loading is very important, make sure you correctly load the dataset according to their features.
Visualize the scanned specimen:
1. Create an isosurface at the specimen boundary, find a proper isovalue to show the whole structure.
2. Use natural colors appropriate for the specimen (red-orange for lobster)
3. Analyze the visualization and answer the following questions:
Q1: Based on the isosurface visualization of the lobster specimen, how many walking legs are visible?
A. 6 walking legs
B. 7 walking legs
C. 8 walking legs
D. 10 walking legs
4. Use a white background. Find an optimal view. Render at 1280x1280. Do not show a color bar or coordinate axes.
5. Set the viewpoint parameters as: [543.52, -957.0, 1007.87] to position; [150.0, 161.5, 38.5] to focal point; [-0.15, 0.62, 0.77] to camera up direction.
6. Save your work:
Save the visualization image as "lobster/results/{agent_mode}/lobster.png".
Save the answers to the analysis questions in plain text as "lobster/results/{agent_mode}/answers.txt".
(Optional, but must save if use paraview) Save the paraview state as "lobster/results/{agent_mode}/lobster.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "lobster/results/{agent_mode}/lobster.py".
(Optional, but must save if use VTK) Save the cxx code script as "lobster/results/{agent_mode}/lobster.cxx"
Do not save any other files, and always save the visualization image and the text file.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Goal: Does the visualization clearly show the structure and details of the Lobster?
2. Boundary Clearity: Are surface details and boundaries of the lobster well-defined?
3. Correct Color: Is the color of the lobster mimic a real one? (red-orange)
- type: llm-rubric
subtype: text
value: |
1. Q1 correct answer: B. 7 walking legs
# Case 11: mhd-magfield_streamribbon
- vars:
question: |
Load the MHD magnetic field dataset from "mhd-magfield_streamribbon/data/mhd-magfield_streamribbon.vti" (VTI format, 128x128x128 grid with components bx, by, bz).
Generate a stream ribbon seeded from a line source along the y-axis at x=64, z=64 (from y=20 to y=108), with 30 seed points.
The stream ribbon should be traced along the magnetic field lines.
Color the stream ribbon by magnetic field magnitude using the 'Cool to Warm' colormap. Enable surface lighting with specular reflection for better 3D perception.
Add a color bar labeled 'Magnetic Field Magnitude'.
Use a white background. Set an isometric camera view. Render at 1024x1024 resolution.
Set the viewpoint parameters as: [200.0, 200.0, 200.0] to position; [63.5, 63.5, 63.5] to focal point; [0.0, 0.0, 1.0] to camera up direction.
Save the visualization image as "mhd-magfield_streamribbon/results/{agent_mode}/mhd-magfield_streamribbon.png".
(Optional, but must save if use paraview) Save the paraview state as "mhd-magfield_streamribbon/results/{agent_mode}/mhd-magfield_streamribbon.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "mhd-magfield_streamribbon/results/{agent_mode}/mhd-magfield_streamribbon.py".
(Optional, but must save if use VTK) Save the cxx code script as "mhd-magfield_streamribbon/results/{agent_mode}/mhd-magfield_streamribbon.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Visualization Goal: Does the result match the ground truth stream ribbon visualization of the MHD magnetic field?
2. Surface Patterns: Does the stream ribbon show similar flow patterns and structures as the ground truth?
3. Surface Coverage: Is the spatial extent and shape of the stream ribbon similar to the ground truth?
4. Color Mapping: Is the color distribution across the surface visually similar to the ground truth?
# Case 12: mhd-turbulence_pathline
- vars:
question: |
Load the MHD turbulence velocity field time series "mhd-turbulence_pathline/data/mhd-turbulence_pathline_{timestep}.vti", where "timestep" in {0000, 0010, 0020, 0030, 0040} (5 timesteps, VTI format, 128x128x128 grid each).
Compute true pathlines by tracking particles through the time-varying velocity field using the ParticlePath filter. Apply TemporalShiftScale (scale=20) and TemporalInterpolator (interval=0.5) to extend particle travel and smooth trajectories.
Seed 26 points along a line on the z-axis at x=64, y=64 (from z=20 to z=108). Use static seeds with termination time 80.
Render pathlines as tubes with radius 0.3. Color by velocity magnitude using the 'Viridis (matplotlib)' colormap.
Add a color bar for velocity magnitude. Set the viewpoint parameters as: [200.0, 200.0, 200.0] to position; [63.5, 63.5, 63.5] to focal point; [0.0, 0.0, 1.0] to camera up direction.
Use a white background. Set an isometric camera view. Render at 1024x1024.
Save the visualization image as "mhd-turbulence_pathline/results/{agent_mode}/mhd-turbulence_pathline.png".
(Optional, but must save if use paraview) Save the paraview state as "mhd-turbulence_pathline/results/{agent_mode}/mhd-turbulence_pathline.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "mhd-turbulence_pathline/results/{agent_mode}/mhd-turbulence_pathline.py".
(Optional, but must save if use VTK) Save the cxx code script as "mhd-turbulence_pathline/results/{agent_mode}/mhd-turbulence_pathline.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Visualization Goal: Does the result match the ground truth pathline visualization of the MHD turbulence velocity field?
2. Pathline Patterns: Do the pathlines show similar particle trajectories and flow structures as the ground truth?
3. Pathline Coverage: Is the spatial extent and distribution of pathlines similar to the ground truth?
4. Color Mapping: Is the color distribution along pathlines visually similar to the ground truth?
# Case 13: mhd-turbulence_pathribbon
- vars:
question: |
Load the MHD turbulence velocity field time series "mhd-turbulence_pathribbon/data/mhd-turbulence_pathribbon_{timestep}.vti", where "timestep" in {0000, 0010, 0020, 0030, 0040} (5 timesteps, VTI format, 128x128x128 grid each).
Compute true pathlines by tracking particles through the time-varying velocity field using the ParticlePath filter. Apply TemporalShiftScale (scale=20) and TemporalInterpolator (interval=0.1) for dense, smooth trajectories.
Seed 26 points along a line on the z-axis at x=64, y=64 (from z=20 to z=108). Use static seeds with termination time 80.
Create ribbon surfaces from the pathlines using the Ribbon filter with width 1.5 and a fixed default normal to prevent twisting. Apply Smooth filter (500 iterations) and generate surface normals for smooth shading.
Set surface opacity to 0.85. Color by velocity magnitude using the 'Cool to Warm' colormap (range 0.1-0.8). Add specular highlights (0.5).
Add a color bar for velocity magnitude. Use a white background. Set an isometric camera view. Render at 1024x1024.
Set the viewpoint parameters as: [200.0, 200.0, 200.0] to position; [63.5, 63.5, 63.5] to focal point; [0.0, 0.0, 1.0] to camera up direction.
Save the visualization image as "mhd-turbulence_pathribbon/results/{agent_mode}/mhd-turbulence_pathribbon.png".
(Optional, but must save if use paraview) Save the paraview state as "mhd-turbulence_pathribbon/results/{agent_mode}/mhd-turbulence_pathribbon.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "mhd-turbulence_pathribbon/results/{agent_mode}/mhd-turbulence_pathribbon.py".
(Optional, but must save if use VTK) Save the cxx code script as "mhd-turbulence_pathribbon/results/{agent_mode}/mhd-turbulence_pathribbon.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Surface Patterns: Does the path ribbon show similar flow patterns and structures as the ground truth?
2. Surface Coverage: Is the spatial extent and shape of the path ribbon similar to the ground truth?
3. Color Mapping: Is the color distribution across the surface visually similar to the ground truth?
# Case 14: mhd-turbulence_streamline
- vars:
question: |
Load the MHD turbulence velocity field dataset "mhd-turbulence_streamline/data/mhd-turbulence_streamline.vti" (VTI format, 128x128x128 grid).
Generate 3D streamlines seeded from a line source along the z-axis at x=64, y=64 (from z=0 to z=127), with 50 seed points.
Color the streamlines by velocity magnitude using the 'Turbo' colormap. Set streamline tube radius to 0.3 using the Tube filter.
Add a color bar labeled 'Velocity Magnitude'. Use a white background. Set an isometric camera view. Render at 1024x1024.
Set the viewpoint parameters as: [200.0, 200.0, 200.0] to position; [63.5, 63.5, 63.5] to focal point; [0.0, 0.0, 1.0] to camera up direction.
Save the visualization image as "mhd-turbulence_streamline/results/{agent_mode}/mhd-turbulence_streamline.png".
(Optional, but must save if use paraview) Save the paraview state as "mhd-turbulence_streamline/results/{agent_mode}/mhd-turbulence_streamline.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "mhd-turbulence_streamline/results/{agent_mode}/mhd-turbulence_streamline.py".
(Optional, but must save if use VTK) Save the cxx code script as "mhd-turbulence_streamline/results/{agent_mode}/mhd-turbulence_streamline.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Visualization Goal: Does the result match the ground truth streamline visualization of the MHD turbulence velocity field?
2. Streamline Patterns: Do the streamlines show similar flow patterns and structures as the ground truth?
3. Streamline Coverage: Is the spatial distribution and density of streamlines similar to the ground truth?
4. Color Mapping: Is the color distribution along streamlines visually similar to the ground truth?
# Case 15: miranda
- vars:
question: |
Task:
Load the Rayleigh-Taylor Instability dataset from "miranda/data/miranda_256x256x256_float32.vtk".
Generate a visualization image of the Rayleigh-Taylor Instability dataset, a time step of a density field in a simulation of the mixing transition in Rayleigh-Taylor instability, with the following visualization settings:
1) Create volume rendering
2) Set the opacity transfer function as a ramp function from value 0 to 1 of the volumetric data, assigning opacity 0 to value 0 and assigning opacity 1 to value 1.
3) Set the color transfer function following the 7 rainbow colors and assign a red color [1.0, 0.0, 0.0] to the highest value, a purple color [0.5, 0.0, 1.0] to the lowest value.
4) Set the viewpoint parameters as: [650, 650, 650] to position; [128, 128, 128] to focal point; [1, 0, 0] to camera up direction
5) Volume rendering ray casting sample distance is 0.1
6) White background
7) Visualization image resolution is 1024x1024
8) Don't show color/scalar bar or coordinate axes.
Save the visualization image as "miranda/results/{agent_mode}/miranda.png".
(Optional, but must save if use paraview) Save the paraview state as "miranda/results/{agent_mode}/miranda.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "miranda/results/{agent_mode}/miranda.py".
(Optional, but must save if use VTK) Save the cxx code script as "miranda/results/{agent_mode}/miranda.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Does the visualization image clearly show the regions from low to high intensity?
2. Does the purple region show areas with low opacity?
3. Does the red region show areas with high opacity?
# Case 16: richtmyer
- vars:
question: |
Task:
Load the Richtmyer dataset from "richtmyer/data/richtmyer_256x256x240_float32.vtk".
Generate a visualization image of the Richtmyer dataset, Entropy field (timestep 160) of Richtmyer-Meshkov instability simulation, with the following visualization settings:
1) Create volume rendering
2) Set the opacity transfer function as a ramp function from value 0.05 to 1 of the volumetric data, assigning opacity 0 to value less than 0.05 and assigning opacity 1 to value 1.
3) Set the color transfer function following the 7 rainbow colors and assign a red color [1.0, 0.0, 0.0] to the highest value, a purple color [0.5, 0.0, 1.0] to the lowest value.
4) Visualization image resolution is 1024x1024
5) Set the viewpoint parameters as: [420, 420, -550] to position; [128, 128, 150] to focal point; [-1, -1, 1] to camera up direction
6) Turn on the shade and set the ambient, diffuse and specular as 1.0
7) White background. Volume rendering ray casting sample distance is 0.1
8) Don't show color/scalar bar or coordinate axes.
Save the visualization image as "richtmyer/results/{agent_mode}/richtmyer.png".
(Optional, but must save if use paraview) Save the paraview state as "richtmyer/results/{agent_mode}/richtmyer.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "richtmyer/results/{agent_mode}/richtmyer.py".
(Optional, but must save if use VTK) Save the cxx code script as "richtmyer/results/{agent_mode}/richtmyer.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Does the visualization show a clear surface with peaks and valleys?
2. Are the peaks highlighted with the reddish color?
3. Are the valleys highlighted with the bluish color?
# Case 17: rotstrat
- vars:
question: |
Task:
Load the rotstrat dataset from "rotstrat/data/rotstrat_256x256x256_float32.vtk".
Generate a visualization image of the Rotstrat dataset, temperature field of a direct numerical simulation of rotating stratified turbulence, with the following visualization settings:
1) Create volume rendering
2) Set the opacity transfer function as a step function jumping from 0 to 1 at value 0.12
3) Set the color transfer function to assign a warm red color [0.71, 0.02, 0.15] to the highest value, a cool color [0.23, 0.29, 0.75] to the lowest value, and a grey color[0.87, 0.87, 0.87] to the midrange value
4) Set the viewpoint parameters as: [800, 128, 128] to position; [0, 128, 128] to focal point; [0, 1, 0] to camera up direction
5) Volume rendering ray casting sample distance is 0.1
6) White background
7) Visualization image resolution is 1024x1024
8) Don't show color/scalar bar or coordinate axes.
Save the visualization image as "rotstrat/results/{agent_mode}/rotstrat.png".
(Optional, but must save if use paraview) Save the paraview state as "rotstrat/results/{agent_mode}/rotstrat.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "rotstrat/results/{agent_mode}/rotstrat.py".
(Optional, but must save if use VTK) Save the cxx code script as "rotstrat/results/{agent_mode}/rotstrat.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Does the visualization image clearly show the shape of turbulence compared to ground truth?
2. Does the visualization show the shape of a vortex in the upper right part of the image?
3. Does the visualization show the shape of a vortex in the bottom left corner of the image?
# Case 18: rti-velocity_glyph
- vars:
question: |
Load the Rayleigh-Taylor instability velocity field dataset from "rti-velocity_glyph/data/rti-velocity_glyph.vti" (VTI format, 128x128x128 grid).
Create a slice at y=64 through the volume.
Place arrow glyphs on the slice, oriented by the velocity vector. Use uniform arrow size (no magnitude scaling, scale factor 3.0).
Color the arrows by velocity magnitude using the 'Viridis (matplotlib)' colormap. Use a sampling stride of 3.
Add a color bar labeled 'Velocity Magnitude'.
Use a white background. Set the camera to view along the negative y-axis. Render at 1024x1024.
Set the viewpoint parameters as: [63.5, 250.0, 63.5] to position; [63.5, 64.0, 63.5] to focal point; [0.0, 0.0, 1.0] to camera up direction.
Save the visualization image as "rti-velocity_glyph/results/{agent_mode}/rti-velocity_glyph.png".
(Optional, but must save if use paraview) Save the paraview state as "rti-velocity_glyph/results/{agent_mode}/rti-velocity_glyph.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "rti-velocity_glyph/results/{agent_mode}/rti-velocity_glyph.py".
(Optional, but must save if use VTK) Save the cxx code script as "rti-velocity_glyph/results/{agent_mode}/rti-velocity_glyph.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Visualization Goal: Does the result match the ground truth glyph visualization of the RTI velocity field?
2. Glyph Patterns: Do the arrow glyphs show similar orientation and spatial patterns as the ground truth?
3. Glyph Appearance: Do the glyphs appear with similar uniform sizing as the ground truth?
4. Color Mapping: Is the color distribution across glyphs visually similar to the ground truth?
# Case 19: rti-velocity_slices
- vars:
question: |
Load the Rayleigh-Taylor instability velocity field from "rti-velocity_slices/data/rti-velocity_slices.vti" (VTI format, 128x128x128).
Create three orthogonal slices: at x=64 (YZ-plane), y=64 (XZ-plane), and z=64 (XY-plane).
Color all three slices by velocity magnitude using the 'Turbo' colormap.
Add a color bar labeled 'Velocity Magnitude'.
Use a white background. Set an isometric camera view that shows all three slices. Render at 1024x1024.
Set the viewpoint parameters as: [200.0, 200.0, 200.0] to position; [63.5, 63.5, 63.5] to focal point; [0.0, 0.0, 1.0] to camera up direction.
Save the visualization image as "rti-velocity_slices/results/{agent_mode}/rti-velocity_slices.png".
(Optional, but must save if use paraview) Save the paraview state as "rti-velocity_slices/results/{agent_mode}/rti-velocity_slices.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "rti-velocity_slices/results/{agent_mode}/rti-velocity_slices.py".
(Optional, but must save if use VTK) Save the cxx code script as "rti-velocity_slices/results/{agent_mode}/rti-velocity_slices.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Slice Count and Orientation: Are there exactly three perpendicular slices (one horizontal XY-plane and two vertical XZ and YZ planes), matching the ground truth arrangement?
2. Color Mapping: Are the slices colored using a Turbo-like colormap (blue to green to yellow to red) mapped to velocity magnitude, with a similar color distribution as the ground truth?
3. Mixing Zone Pattern: Does the horizontal (XY) slice show a chaotic, high-velocity-magnitude mixing pattern in its center region, similar to the ground truth?
# Case 20: rti-velocity_streakline
- vars:
question: |
Load the Rayleigh–Taylor instability velocity field time series from "rti-velocity_streakline/data/rti-velocity_streakline_{timestep}.nc", where "timestep" in {0030, 0031, 0032, 0033, 0034, 0035, 0036, 0037, 0038, 0039, 0040} (11 timesteps, NetCDF format, 128×128×128 grid each, with separate vx, vy, vz arrays).
Construct the time-varying velocity field u(x,t) by merging vx, vy, vz into a single vector field named "velocity", and compute the velocity magnitude "magnitude" = |velocity| for coloring.
Compute streaklines as a discrete approximation of continuous particle injection: continuously release particles from fixed seed points at every sub-timestep into the time-varying velocity field using the StreakLine filter. Apply TemporalShiftScale (scale=20) to extend particle travel time, and apply TemporalInterpolator with a sub-timestep interval of 0.25 (or smaller) to approximate continuous injection over time.
Seed 26 static points along a line on the z-axis at x=64, y=64 (from z=20 to z=108). Use StaticSeeds=True, ForceReinjectionEveryNSteps=1 (reinjection at every sub-timestep), and set TerminationTime=200.
Render the resulting streaklines as tubes with radius 0.3. Color the tubes by velocity magnitude ("magnitude") using the 'Cool to Warm (Extended)' colormap. Add a color bar for velocity magnitude.
Use a white background. Set an isometric camera view and render at 1024×1024.
Set the viewpoint parameters as: [200.0, 200.0, 200.0] to position; [63.5, 63.5, 63.5] to focal point; [0.0, 0.0, 1.0] to camera up direction.
Save the visualization image as "rti-velocity_streakline/results/{agent_mode}/rti-velocity_streakline.png".
(Optional, but must save if use paraview) Save the paraview state as "rti-velocity_streakline/results/{agent_mode}/rti-velocity_streakline.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "rti-velocity_streakline/results/{agent_mode}/rti-velocity_streakline.py".
(Optional, but must save if use VTK) Save the cxx code script as "rti-velocity_streakline/results/{agent_mode}/rti-velocity_streakline.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Streak Line Patterns: Do the streak lines show similar flow patterns and structures as the ground truth?
2. Streak Line Coverage: Is the spatial extent and distribution of streak lines similar to the ground truth?
3. Color Mapping: Is the color distribution along streak lines visually similar to the ground truth?
# Case 21: solar-plume
- vars:
question: |
Task:
Load the solar plume dataset from "solar-plume/data/solar-plume_126x126x512_float32_scalar3.raw", the information about this dataset:
solar-plume (Vector)
Data Scalar Type: float
Data Byte Order: little Endian
Data Extent: 126x126x512
Number of Scalar Components: 3
Data loading is very important, make sure you correctly load the dataset according to their features.
Add a "stream tracer" filter under the solar plume data to display streamline, set the "Seed type" to "Point Cloud" and set the center of point cloud to 3D position [50, 50, 320] with a radius 30, then hide the point cloud sphere.
Add a "tube" filter under the "stream tracer" filter to enhance the streamline visualization. Set the radius to 0.5. In the pipeline browser panel, hide everything except the "tube" filter.
Please think step by step and make sure to fulfill all the visualization goals mentioned above.
Set the viewpoint parameters as: [62.51, -984.78, 255.45] to position; [62.51, 62.46, 255.45] to focal point; [0, 0, 1] to camera up direction.
Use a white background. Find an optimal view. Render at 1280x1280. Do not show a color bar or coordinate axes.
Save the visualization image as "solar-plume/results/{agent_mode}/solar-plume.png".
(Optional, but must save if use paraview) Save the paraview state as "solar-plume/results/{agent_mode}/solar-plume.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "solar-plume/results/{agent_mode}/solar-plume.py".
(Optional, but must save if use VTK) Save the cxx code script as "solar-plume/results/{agent_mode}/solar-plume.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Visualization Goal: Does the result match the ground truth streamline visualization of solar-plume flow structures?
2. Streamline Patterns: Do the streamlines show similar flow patterns and structures as the ground truth, particularly in the plume region?
3. Streamline Coverage: Is the spatial distribution and density of streamlines similar to the ground truth?
4. Visual Appearance: Do the streamline tubes appear similar in thickness and visibility to the ground truth?
# Case 22: supernova_isosurface
- vars:
question: |
Task:
Load the supernova dataset from "supernova_isosurface/data/supernova_isosurface_256x256x256_float32.raw", the information about this dataset:
supernova (Scalar)
Data Scalar Type: float
Data Byte Order: little Endian
Data Spacing: 1x1x1
Data Extent: 256x256x256
Data loading is very important, make sure you correctly load the dataset according to their features.
Then visualize it and extract two isosurfaces. One of them use color red, showing areas with low density (isovalue 40 and opacity 0.2), while the other use color light blue, showing areas with high density (isovalue 150 and opacity 0.4).
Please think step by step and make sure to fulfill all the visualization goals mentioned above. Only make the two isosurfaces visible.
Use a white background. Find an optimal view. Render at 1280x1280. Do not show a color bar or coordinate axes.
Set the viewpoint parameters as: [567.97, 80.17, 167.28] to position; [125.09, 108.83, 121.01] to focal point; [-0.11, -0.86, 0.50] to camera up direction.
Save the visualization image as "supernova_isosurface/results/{agent_mode}/supernova_isosurface.png".
(Optional, but must save if use paraview) Save the paraview state as "supernova_isosurface/results/{agent_mode}/supernova_isosurface.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "supernova_isosurface/results/{agent_mode}/supernova_isosurface.py".
(Optional, but must save if use VTK) Save the cxx code script as "supernova_isosurface/results/{agent_mode}/supernova_isosurface.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Visualization Goal: How well does the result achieve the overall goal of showing the supernova structure with two distinct isosurfaces representing different density regions?
2. Does the red isosurface show low density areas (outside regions) with lower opacity?
3. Does the blue isosurface show high density areas (inside regions) with higher opacity?
# Case 23: supernova_streamline
- vars:
question: |
Load the Supernova velocity vector field from "supernova_streamline/data/supernova_streamline_100x100x100_float32_scalar3.raw", the information about this dataset:
Supernova Velocity (Vector)
Data Scalar Type: float
Data Byte Order: Little Endian
Data Extent: 100x100x100
Number of Scalar Components: 3
Data loading is very important, make sure you correctly load the dataset according to their features.
Create streamlines using a "Stream Tracer" filter with "Point Cloud" seed type. Set the seed center to [50, 50, 50], with 200 seed points and a radius of 45.0. Set maximum streamline length to 100.0.
Add a "Tube" filter on the stream tracer. Set tube radius to 0.3 with 12 sides.
Color the tubes by Vorticity magnitude using a diverging colormap with the following RGB control points:
- Value 0.0 -> RGB(0.231, 0.298, 0.753) (blue)
- Value 0.05 -> RGB(0.865, 0.865, 0.865) (white)
- Value 0.5 -> RGB(0.706, 0.016, 0.149) (red)
Show the dataset bounding box as an outline (black).
In the pipeline browser panel, hide the stream tracer and only show the tube filter and the outline.
Use a white background. Render at 1280x1280.
Set the viewpoint parameters as: [41.38, 73.91, -282.0] to position; [49.45, 49.50, 49.49] to focal point; [0.01, 1.0, 0.07] to camera up direction.
Save the visualization image as "supernova_streamline/results/{agent_mode}/supernova_streamline.png".
(Optional, but must save if use paraview) Save the paraview state as "supernova_streamline/results/{agent_mode}/supernova_streamline.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "supernova_streamline/results/{agent_mode}/supernova_streamline.py".
(Optional, but must save if use VTK) Save the cxx code script as "supernova_streamline/results/{agent_mode}/supernova_streamline.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Central Structure: Is there a dense, chaotic cluster of streamlines near the center of the volume, matching the ground truth?
2. Radial Extensions: Are there long, straight streamline tubes extending radially outward from the central region, similar to the ground truth?
3. Color Mapping: Are the tubes colored by vorticity magnitude using a blue-white-red diverging colormap, with warm colors concentrated near the center and cool colors on the extended lines?
# Case 24: tangaroa_streamribbon
- vars:
question: |
Task:
Load the tangaroa dataset from "tangaroa_streamribbon_300x180x120_float32_scalar3.raw", the information about this dataset:
tangaroa (Vector)
Data Scalar Type: float
Data Byte Order: little Endian
Data Extent: 300x180x120
Number of Scalar Components: 3
Data loading is very important, make sure you correctly load the dataset according to their features.
Apply "streamline tracer" filter, set the "Seed Type" to point cloud, turn off the "show sphere", set the center to [81.6814, 80.708, 23.5093], and radius to 29.9
Add "Ribbon" filter to the streamline tracer results and set width to 0.3, set the Display representation to Surface.
In pipeline browser panel, hide everything except the ribbon filter results.
Please think step by step and make sure to fulfill all the visualization goals mentioned above.
Use a white background. Find an optimal view. Render at 1280x1280. Do not show a color bar or coordinate axes.
Set the viewpoint parameters as: [372.27, 278.87, 214.44] to position; [169.85, 76.46, 12.02] to focal point; [-0.41, 0.82, -0.41] to camera up direction.
Save the visualization image as "tangaroa_streamribbon/results/{agent_mode}/tangaroa_streamribbon.png".
(Optional, but must save if use paraview) Save the paraview state as "tangaroa_streamribbon/results/{agent_mode}/tangaroa_streamribbon.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "tangaroa_streamribbon/results/{agent_mode}/tangaroa_streamribbon.py".
(Optional, but must save if use VTK) Save the cxx code script as "tangaroa_streamribbon/results/{agent_mode}/tangaroa_streamribbon.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Visualization Goal: Does the result match the ground truth visualization of tangaroa flow structures using ribbon surfaces?
2. Flow Surface Patterns: Do the ribbon surfaces show similar flow patterns and structures as the ground truth?
3. Surface Coverage: Is the spatial distribution and coverage of the flow surfaces similar to the ground truth?
4. Visual Appearance: Do the ribbon surfaces appear similar in width and structure to the ground truth?
# Case 25: tgc-velocity_contour
- vars:
question: |
Load the turbulence-gravity-cooling velocity field dataset from "tgc-velocity_contour/data/tgc-velocity_contour.vti" (VTI format, 64x64x64).
Extract a slice at z=32 and color it by velocity magnitude using 'Viridis (matplotlib)' colormap.
Also add contour lines of velocity magnitude on the same slice at values [0.3, 0.6, 0.9, 1.2] using the Contour filter on the slice output.
Display contour lines in white. Add a color bar labeled 'Velocity Magnitude'.
Light gray background (RGB: 0.9, 0.9, 0.9). Top-down camera. Render at 1024x1024.
Set the viewpoint parameters as: [31.5, 31.5, 100.0] to position; [31.5, 31.5, 32.0] to focal point; [0.0, 1.0, 0.0] to camera up direction.
Save the visualization image as "tgc-velocity_contour/results/{agent_mode}/tgc-velocity_contour.png".
(Optional, but must save if use paraview) Save the paraview state as "tgc-velocity_contour/results/{agent_mode}/tgc-velocity_contour.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "tgc-velocity_contour/results/{agent_mode}/tgc-velocity_contour.py".
(Optional, but must save if use VTK) Save the cxx code script as "tgc-velocity_contour/results/{agent_mode}/tgc-velocity_contour.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Visualization Goal: Does the result match the ground truth slice and contour visualization of the TGC velocity field?
2. Slice Pattern: Does the colored slice show similar patterns and structures as the ground truth?
3. Contour Lines: Are the contour lines positioned and shaped similarly to the ground truth?
4. Color Mapping: Is the color distribution on the slice visually similar to the ground truth?
# Case 26: tornado
- vars:
question: |
Load the Tornado vector field from "tornado/data/tornado_64x64x64_float32_scalar3.raw", the information about this dataset:
Tornado (Vector)
Data Scalar Type: float
Data Byte Order: Little Endian
Data Extent: 64x64x64
Number of Scalar Components: 3
Data loading is very important, make sure you correctly load the dataset according to their features.
Create a streamline visualization using a "Stream Tracer" filter with "Point Cloud" seed type. Set the seed center to [31.5, 31.5, 47.25], radius 12.6, and maximum streamline length to 512.0. Add a "Tube" filter (radius 0.25) on the stream tracer. Color the tubes by Velocity magnitude using the 'Cool to Warm (Diverging)' colormap. Also display the stream tracer directly with line width 5.0 and "Render Lines As Tubes" enabled.
Add a "Glyph" filter on the original data using Arrow glyph type. Orient arrows by the Velocity vector and scale by Velocity magnitude with a scale factor of 25.0. Set maximum number of sample points to 2500. Color glyphs by Velocity magnitude using the same colormap.
Add an "Outline" filter to display the dataset bounding box (black).
Use a white background (RGB: 1.0, 1.0, 1.0). Find an optimal view. Render at 1280x1280. Show both color bar and coordinate axes.
Set the viewpoint parameters as: [142.01, -36.46, 93.96] to position; [31.5, 31.5, 31.5] to focal point; [-0.35, 0.25, 0.90] to camera up direction.
Save the visualization image as "tornado/results/{agent_mode}/tornado.png".
(Optional, but must save if use paraview) Save the paraview state as "tornado/results/{agent_mode}/tornado.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "tornado/results/{agent_mode}/tornado.py".
(Optional, but must save if use VTK) Save the cxx code script as "tornado/results/{agent_mode}/tornado.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Vortex Structure: Is a funnel-shaped vortex core visible with streamlines spiraling around a central vertical axis, matching the ground truth?
2. Glyph Presence: Are arrow glyphs distributed throughout the volume showing velocity direction, similar to the ground truth?
3. Color Mapping: Are both the streamline tubes and arrow glyphs colored by velocity magnitude using a blue-to-red diverging colormap, matching the ground truth color distribution?
# Case 27: twoswirls_streamribbon
- vars:
question: |
Load the Two Swirls vector field from "twoswirls_streamribbon/data/twoswirls_streamribbon_64x64x64_float32_scalar3.raw", the information about this dataset:
Two Swirls (Vector)
Data Scalar Type: float
Data Byte Order: Little Endian
Data Extent: 64x64x64
Number of Scalar Components: 3
Data loading is very important, make sure you correctly load the dataset according to their features.
Create two stream ribbons using "Stream Tracer" filters with "Line" seed type (resolution 25 points each), and apply a "Ribbon" filter (width 2.5) to each:
- Stream Ribbon 1: Line seed from [16, 10, 32] to [16, 54, 32]. Ribbon colored solid green (RGB: 0.2, 0.7, 0.3) with opacity 0.35.
- Stream Ribbon 2: Line seed from [48, 10, 32] to [48, 54, 32]. Ribbon colored solid blue (RGB: 0.2, 0.4, 0.85) with opacity 0.35.
Show the dataset bounding box as an outline (black, opacity 0.3).
In the pipeline browser panel, hide all stream tracers and only show the ribbon filters and the outline.
Use a white background (RGB: 1.0, 1.0, 1.0). Find an optimal view. Render at 1280x1280. Do not show a color bar or coordinate axes.
Set the viewpoint parameters as: [30.51, -154.18, 144.99] to position; [30.51, 31.5, 30.91] to focal point; [0.0, 0.53, 0.85] to camera up direction.
Save the visualization image as "twoswirls_streamribbon/results/{agent_mode}/twoswirls_streamribbon.png".
(Optional, but must save if use paraview) Save the paraview state as "twoswirls_streamribbon/results/{agent_mode}/twoswirls_streamribbon.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "twoswirls_streamribbon/results/{agent_mode}/twoswirls_streamribbon.py".
(Optional, but must save if use VTK) Save the cxx code script as "twoswirls_streamribbon/results/{agent_mode}/twoswirls_streamribbon.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Swirl Separation: Are there two visually distinct swirl structures (one on the left and one on the right), matching the spatial arrangement in the ground truth?
2. Stream Ribbon Shape: Do the ribbon surfaces show wrapped, swirling sheet-like structures similar to the ground truth?
3. Color and Transparency: Are the stream ribbons rendered with distinct colors (green and blue) and semi-transparency, similar to the ground truth?
# Case 28: vortex
- vars:
question: |
Task:
Load the vortex dataset from "vortex/data/vortex_128x128x128_float32.raw", the information about this dataset:
vortex (Scalar)
Data Scalar Type: float
Data Byte Order: little Endian
Data Extent: 128x128x128
Number of Scalar Components: 1
Instructions:
1. Load the dataset into ParaView.
2. Leverage "contour" filter to achieve iso-surface rendering. In pipeline browser panel, hide everything except the "contour" fileter.
3. In properties panel of "contour" filter, set isosurface value to -0.2, use Solid Color and set the color as beige.
4. Enable Ambient occlusion by toggle the "Use Ambient Occlusion" button in the Render Passes.
5. Add head light with light inspector, set "Coords" as Camera, "Intentsity" to 0.2, Type to "Directional".
6. Use a white background. Find an optimal view. Render at 1280x1280. Do not show a color bar or coordinate axes.
7. Set the viewpoint parameters as: [308.85, 308.85, 308.85] to position; [63.5, 63.5, 63.5] to focal point; [-0.41, 0.82, -0.41] to camera up direction.
8. Save your work:
Save the visualization image as "vortex/results/{agent_mode}/vortex.png".
(Optional, but must save if use paraview) Save the paraview state as "vortex/results/{agent_mode}/vortex.pvsm".
(Optional, but must save if use pvpython script) Save the python script as "vortex/results/{agent_mode}/vortex.py".
(Optional, but must save if use VTK) Save the cxx code script as "vortex/results/{agent_mode}/vortex.cxx"
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Overall Visualization Goal: Does the result match the ground truth isosurface rendering of the vortex scalar field?
2. Isosurface Structure: Does the isosurface show the same vortex structure and topology as the ground truth?
3. Surface Appearance: Does the surface color and shading appear similar to the ground truth?
4. Visualization Clarity: Are the vortex features clearly visible and comparable to the ground truth?
# Case 29: line-plot
- vars:
question: |
Read the dataset in the file "line-plot/data/line-plot.ex2", and print the number of components and the range of all the variables.
Show a default view of the dataset, colored by the variable Pres.
Create a line plot over all the variables in the dataset, from (0,0,0) to (0,0,10).
Write the values of the line plot in the file "line-plot/results/{agent_mode}/line-plot.csv", and save a screenshot of the line plot in "line-plot/results/{agent_mode}/line-plot.png".
(Optional, but must save if use paraview) Save the paraview state as "line-plot/results/{agent_mode}/line-plot.pvsm".
(Optional, but must save if use python script) Save the python script as "line-plot/results/{agent_mode}/line-plot.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Line Visualization Quality: Are multiple distinct lines clearly visible and properly rendered showing the evolution of different variables along the specified path?
2. Variable Differentiation: Are all dataset variables visually distinguishable through distinct colors or line styles with clear separation between curves?
3. Axis and Scale Appropriateness: Do the plot axes display appropriate ranges and scaling that effectively show the data trends and variations?
4. Legend and Readability: Is there a clear legend identifying each variable line with readable labels and proper visual organization?
# Case 30: ml-dvr
- vars:
question: |
I would like to use ParaView to visualize a dataset.
Read in the file named "ml-dvr/data/ml-dvr.vtk".
Generate a volume rendering using the default transfer function.
Rotate the view to an isometric direction.
Save a screenshot of the result in the filename "ml-dvr/results/{agent_mode}/ml-dvr.png".
The rendered view and saved screenshot should be 1920 x 1080 pixels.
(Optional, but must save if use paraview) Save the paraview state as "ml-dvr/results/{agent_mode}/ml-dvr.pvsm".
(Optional, but must save if use python script) Save the python script as "ml-dvr/results/{agent_mode}/ml-dvr.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Volume Rendering Quality: Is the volume rendering properly generated with appropriate opacity and color mapping that reveals internal structures?
2. Transfer Function Application: Does the default transfer function effectively highlight meaningful data features and provide good visual contrast?
3. Isometric View Setup: Is the visualization displayed from an isometric viewpoint that provides a clear three-dimensional perspective of the volume?
4. Visual Clarity and Detail: Are the volume details clearly visible with proper lighting and shading that enhances depth perception?
# Case 31: ml-iso
- vars:
question: |
Read in the file named "ml-iso/data/ml-iso.vtk", and generate an isosurface of the variable var0 at value 0.5.
Use a white background color. Save a screenshot of the result, size 1920 x 1080 pixels, in "ml-iso/results/{agent_mode}/ml-iso.png".
(Optional, but must save if use paraview) Save the paraview state as "ml-iso/results/{agent_mode}/ml-iso.pvsm".
(Optional, but must save if use python script) Save the python script as "ml-iso/results/{agent_mode}/ml-iso.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Isosurface Generation: Is the isosurface properly generated at the specified value (0.5) with correct topology and continuity?
2. Surface Rendering Quality: Does the isosurface display smooth surfaces with appropriate shading and lighting that reveals the 3D structure?
3. Geometric Accuracy: Are the surface features geometrically correct and free from artifacts or discontinuities?
4. Visual Presentation: Is the isosurface clearly visible with good contrast and coloring that enhances the understanding of the data structure?
# Case 32: ml-slice-iso
- vars:
question: |
Please generate a ParaView Python script for the following operations.
Read in the file named "ml-slice-iso/data/ml-slice-iso.vtk".
Slice the volume in a plane parallel to the y-z plane at x=0.
Take a contour through the slice at the value 0.5.
Color the contour red. Use a white background.
Rotate the view to look at the +x direction.
Save a screenshot of the result in the filename "ml-slice-iso/results/{agent_mode}/ml-slice-iso.png".
The rendered view and saved screenshot should be 1920 x 1080 pixels.
(Optional, but must save if use paraview) Save the paraview state as "ml-slice-iso/results/{agent_mode}/ml-slice-iso.pvsm".
(Optional, but must save if use python script) Save the python script as "ml-slice-iso/results/{agent_mode}/ml-slice-iso.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Slice Generation: Is the y-z plane slice properly generated at x=0 position showing the correct cross-section of the volume?
2. Contour on Slice: Are the contour lines at value 0.5 correctly extracted from the slice and properly displayed?
3. Red Color Application: Is the contour visualization properly colored red as specified in the requirements?
4. View Direction: Is the visualization displayed from the correct +x direction view that provides clear visibility of the slice and contours?
# Case 33: points-surf-clip
- vars:
question: |
I would like to use ParaView to visualize a dataset.
Read in the file named "points-surf-clip/data/points-surf-clip.ex2".
Generate an 3d Delaunay triangulation of the dataset.
Clip the data with a y-z plane at x=0, keeping the -x half of the data and removing the +x half.
Render the image as a wireframe.
Save a screenshot of the result in the filename "points-surf-clip/results/{agent_mode}/points-surf-clip.png".
The rendered view and saved screenshot should be 1920 x 1080 pixels. Use a white background color.
(Optional, but must save if use paraview) Save the paraview state as "points-surf-clip/results/{agent_mode}/points-surf-clip.pvsm".
(Optional, but must save if use python script) Save the python script as "points-surf-clip/results/{agent_mode}/points-surf-clip.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Delaunay Triangulation Quality: Is the 3D Delaunay triangulation properly generated creating a valid mesh structure from the point data?
2. Clipping Accuracy: Is the mesh correctly clipped by the y-z plane at x=0, with only the -x half of the data remaining visible?
3. Wireframe Representation: Is the result displayed as a clear wireframe showing the triangulated mesh structure with visible edges?
4. Geometric Integrity: Does the clipped wireframe maintain proper connectivity and show the expected geometric features without artifacts?
# Case 34: shrink-sphere
- vars:
question: |
Create a default sphere and then hide it.
Create a shrink filter from the sphere.
Double the sphere's theta resolution.
Divide the shrink filter's shrink factor in half.
Extract a wireframe from the sphere.
Group the shrink filter and wireframe together and show them.
Save a screenshot of the result in the filename "shrink-sphere/results/{agent_mode}/shrink-sphere.png".
The rendered view and saved screenshot should be 1920 x 1080 pixels and have a white background.
(Optional, but must save if use paraview) Save the paraview state as "shrink-sphere/results/{agent_mode}/shrink-sphere.pvsm".
(Optional, but must save if use python script) Save the python script as "shrink-sphere/results/{agent_mode}/shrink-sphere.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Sphere Creation and Resolution: Is the sphere created with doubled theta resolution providing higher geometric detail and smoother curvature?
2. Shrink Filter Application: Is the shrink filter properly applied with halved shrink factor creating visible separation between mesh elements?
3. Dual Representation: Are both the wireframe sphere and shrink filter results simultaneously visible and properly grouped together?
4. Visual Quality: Does the visualization clearly show the contrast between the wireframe structure and the shrunken elements with appropriate white background?
# Case 35: stream-glyph
- vars:
question: |
I would like to use ParaView to visualize a dataset.
Read in the file named "stream-glyph/data/stream-glyph.ex2".
Trace streamlines of the V data array seeded from a default point cloud.
Render the streamlines with tubes.
Add cone glyphs to the streamlines.
Color the streamlines and glyphs by the Temp data array.
View the result in the +X direction.
Save a screenshot of the result in the filename "stream-glyph/results/{agent_mode}/stream-glyph.png".
The rendered view and saved screenshot should be 1920 x 1080 pixels.
(Optional, but must save if use paraview) Save the paraview state as "stream-glyph/results/{agent_mode}/stream-glyph.pvsm".
(Optional, but must save if use python script) Save the python script as "stream-glyph/results/{agent_mode}/stream-glyph.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Streamline Generation: Are streamlines properly traced following the V variable flow field with appropriate seeding from the point cloud?
2. Tube and Glyph Rendering: Are streamlines rendered as tubes with cone glyphs properly attached showing flow direction and magnitude?
3. Temperature Color Mapping: Are both streamlines and glyphs correctly colored by the Temp variable with appropriate color scaling?
4. View Configuration: Is the visualization displayed from the correct +x view direction providing clear visibility of the flow patterns and structures?
# Case 36: time-varying
- vars:
question: |
Read the dataset in the file "time-varying/data/time-varying.ex2", and color the data by the EQPS variable.
Viewing in the +y direction, play an animation through the time steps, with visible color bar legend.
Rescale the data range to last time step, and play the animation again.
Create a second linked render view to the right of the first, applying a temporal interpolator to the second view.
Play the animation simultaneously in both views, and save the animation of both views in "time-varying/results/{agent_mode}/time-varying.avi".
Print the following statistics: average value of EQPS over all locations and all time steps, average value of EQPS over all locations in the first half of the time steps, average value of EQPS over all locations in the even numbered time steps, and variance of EQPS over all locations and all the time steps.
Save the last frame of the visualization image as "time-varying/results/{agent_mode}/time-varying.png".
(Optional, but must save if use paraview) Save the paraview state as "time-varying/results/{agent_mode}/time-varying.pvsm".
(Optional, but must save if use python script) Save the python script as "time-varying/results/{agent_mode}/time-varying.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Temporal Animation Quality: Does the animation smoothly progress through all time steps showing the evolution of the EQPS variable over time?
2. Dual View Configuration: Are both render views properly configured with the second view showing temporal interpolation effects compared to the first?
3. Color Mapping and Legend: Is the EQPS variable properly color-mapped with an appropriate color bar legend visible throughout the animation?
4. View Direction and Layout: Is the +y direction view properly set and are both views arranged side-by-side in the correct layout configuration?
# Case 37: chart-opacity
- vars:
question: |
Create a wavelet object.
Create a plot over line chart from the wavelet with three paths: arc_length, Points_Z, and RTData variables with opacity for arc_length 1 and opacity for Points_Z and RTData 0.3.
Save the visualization image as "chart-opacity/results/{agent_mode}/chart-opacity.png".
(Optional, but must save if use paraview) Save the paraview state as "chart-opacity/results/{agent_mode}/chart-opacity.pvsm".
(Optional, but must save if use python script) Save the python script as "chart-opacity/results/{agent_mode}/chart-opacity.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Chart Generation: Is the plot over line chart properly created from the wavelet data?
2. Variable Display: Are arc_length, Points_Z, and RTData variables all correctly plotted, showing all three specified variables and distinguishable in the chart?
3. Opacity Settings: Is the arc_length variable displayed with full opacity (1.0) while Points_Z and RTData show reduced opacity (0.3)?
4. Chart Clarity: Does the chart provide clear visualization of the data trends with appropriate axis scaling and readable formatting?
# Case 38: color-blocks
- vars:
question: |
I would like to use ParaView to visualize a dataset.
Set the background to a blue-gray palette.
Read the file "color-blocks/data/color-blocks.ex2".
This is a multiblock dataset.
Color the dataset by the vtkBlockColors field.
Retrieve the color map for vtkBlockColors.
Retrieve the opacity transfer function for vtkBlockColors.
Retrieve the 2D transfer function for vtkBlockColors.
Set block coloring for the block at /IOSS/element_blocks/block_2 using the variable ACCL on the x component of the points.
Rescale the block's color and opacity maps to match the current data range of block_2.
Retrieve the color transfer function for the ACCL variable of block_2.
Enable the color bar for block_2.
Apply a cool to warm color preset to the color map for block_2.
Set the camera to look down the -y direction and to see the entire dataset.
Save the visualization image as "color-blocks/results/{agent_mode}/color-blocks.png".
(Optional, but must save if use paraview) Save the paraview state as "color-blocks/results/{agent_mode}/color-blocks.pvsm".
(Optional, but must save if use python script) Save the python script as "color-blocks/results/{agent_mode}/color-blocks.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Block Color Mapping: Is the dataset properly colored by vtkBlockColors field with distinct block visualization?
2. Individual Block Coloring: Is block_2 correctly colored using the x component of the ACCL variable with appropriate scaling?
3. Color Transfer Functions: Are the color transfer functions properly applied with cool to warm coloring for the ACCL variable?
4. View Configuration: Is the dataset displayed from the -y direction with blue-gray background and visible color bar legend?
# Case 39: color-data
- vars:
question: |
Create a wavelet object.
Create a new calculator with the function 'RTData*iHat + ln(RTData)*jHat + coordsZ*kHat'.
Get a color transfer function/color map and opacity transfer function/opacity map for the result of the calculation, scaling the color and/or opacity maps to the data range.
For a surface representation, color by the x coordinate of the result using a cool to warm color map, show the color bar/color legend, and save a screenshot of size 1158 x 833 pixels in "color-data/results/{agent_mode}/color-data.png".
(Optional, but must save if use paraview) Save the paraview state as "color-data/results/{agent_mode}/color-data.pvsm".
(Optional, but must save if use python script) Save the python script as "color-data/results/{agent_mode}/color-data.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Color Transfer Function: Is the color transfer function correctly applied with cool to warm color mapping scaled to the data range?
2. Surface Coloring: Is the surface representation properly colored by the x coordinate of the calculated result?
3. Color Bar Display: Is the color bar/legend visible and properly displaying the color mapping scale and values?
# Case 40: export-gltf
- vars:
question: |
Create a wavelet object.
Create a surface rendering of the wavelet object and color by RTData.
Scale the color map to the data, and don't display the color bar or the orientation axes.
Export the view to "export-gltf/results/{agent_mode}/ExportedGLTF.gltf".
Next load the file "export-gltf/results/{agent_mode}/ExportedGLTF.gltf" and display it as a surface.
Color this object by TEXCOORD_0.
Scale the color map to the data, and don't display the color bar or the orientation axes.
Use the 'Cool to Warm' colormap. Set the background color to white.
Save the visualization image as "export-gltf/results/{agent_mode}/export-gltf.png".
(Optional, but must save if use paraview) Save the paraview state as "export-gltf/results/{agent_mode}/export-gltf.pvsm".
(Optional, but must save if use python script) Save the python script as "export-gltf/results/{agent_mode}/export-gltf.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. GLTF Export Quality: Is the wavelet object properly exported to GLTF format with correct surface representation and RTData coloring?
2. GLTF Import and Display: Is the exported GLTF file successfully loaded and displayed as a surface with proper geometry?
3. Texture Coordinate Coloring: Is the imported GLTF object correctly colored by TEXCOORD_0 with Cool to Warm colormap?
4. Clean Presentation: Are the color bar and orientation axes properly hidden for a clean visualization appearance?
# Case 41: import-gltf
- vars:
question: |
Load the "BlueGrayBackground" palette.
Read the file "import-gltf/data/import-gltf.glb" and import the nodes "/assembly/Axle", "assembly/OuterRing/Torus002", and "assembly/OuterRing/MiddleRing/InnerRing".
Set the layout size to 300x300 pixels.
Point the camera in the positive Y direction and zoom to fit.
Make sure all views are rendered, then save a screenshot to "import-gltf/results/{agent_mode}/import-gltf.png".
(Optional, but must save if use paraview) Save the paraview state as "import-gltf/results/{agent_mode}/import-gltf.pvsm".
(Optional, but must save if use python script) Save the python script as "import-gltf/results/{agent_mode}/import-gltf.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. GLTF Import Success: Are the specified GLTF nodes properly imported and displayed as separate geometric components?
2. Node Selection: Are all three specified nodes (Axle, Torus002, InnerRing) correctly imported and visible?
3. Camera Positioning: Is the camera positioned in the positive Y direction with appropriate zoom to fit all imported geometry? Carefully compare the camera position of GT and result images.
4. Layout Configuration: Is the view properly sized to 300x300 pixels with correct rendering and background palette?
# Case 42: render-histogram
- vars:
question: |
Create a wavelet object and render it as a surface colored by RTDATA with a visible color bar.
Rescale the colors to the data range and use the 'Cool to Warm' color map.
Next, split the view horizontally to the right and create a histogram view from the wavelet RTDATA.
Apply the same 'Cool to Warm' color map to the histogram.
Save a screenshot of both views (wavelet rendering on the left and histogram on the right) in the file "render-histogram/results/{agent_mode}/render-histogram.png".
(Optional, but must save if use paraview) Save the paraview state as "render-histogram/results/{agent_mode}/render-histogram.pvsm".
(Optional, but must save if use python script) Save the python script as "render-histogram/results/{agent_mode}/render-histogram.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Wavelet Visualization: Is the wavelet object properly rendered with RTDATA coloring and visible color bar?
2. Split View Layout: Is the view correctly split with the wavelet visualization on the left and histogram on the right?
3. Histogram Generation: Is the histogram properly generated from RTDATA showing the data distribution?
4. Color Map Consistency: Are both the wavelet visualization and histogram using the same Cool to Warm color map?
# Case 43: reset-camera-direction
- vars:
question: |
Create a Wavelet object, set its representation to "Surface with Edges", and set the camera direction to [0.5, 1, 0.5].
Save a screenshot to the file "reset-camera-direction/results/{agent_mode}/reset-camera-direction.png".
(Optional, but must save if use paraview) Save the paraview state as "reset-camera-direction/results/{agent_mode}/reset-camera-direction.pvsm".
(Optional, but must save if use python script) Save the python script as "reset-camera-direction/results/{agent_mode}/reset-camera-direction.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Wavelet Creation: Is the Wavelet object properly created and displayed in the scene?
2. Surface with Edges Representation: Is the wavelet correctly displayed with "Surface with Edges" representation showing both surface and wireframe?
3. Camera Direction: Is the camera positioned according to the specified direction vector [0.5, 1, 0.5]?
4. View Quality: Does the visualization provide a clear view of the wavelet structure from the specified camera angle?
# Case 44: save-transparent
- vars:
question: |
I would like to use ParaView to visualize a dataset.
Create a wavelet object and show it. Color the rendering by the variable ‘RTData’.
Render the wavelet as a surface. Hide the color bar.
Next, set the layout size to be 300 pixels by 300 pixels.
Next, move the camera with the following settings. The camera position should be [30.273897726939246, 40.8733980301544, 43.48927935675712]. The camera view up should be [-0.3634544237682163, 0.7916848767068606, -0.49105594165731975]. The camera parallel scale should be 17.320508075688775.
Save a screenshot to the file “save-transparent/results/{agent_mode}/save-transparent.png”, set the image resolution to 300x300, and set the background to transparent.
(Optional, but must save if use paraview) Save the paraview state as “save-transparent/results/{agent_mode}/save-transparent.pvsm”.
(Optional, but must save if use python script) Save the python script as “save-transparent/results/{agent_mode}/save-transparent.py”.
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Object Creation: Is the wavelet object properly created and displayed in the scene? Looking similar to the GT image?
2. Transparent Background: Is the screenshot saved with a properly transparent background instead of solid color?
# Case 45: subseries-of-time-series
- vars:
question: |
Read the file "subseries-of-time-series/data/subseries-of-time-series.ex2". Load two element blocks: the first is called 'Unnamed block ID: 1 Type: HEX', the second is called 'Unnamed block ID: 2 Type: HEX'.
Next, slice this object with a plane with origin at [0.21706008911132812, 4.0, -5.110947132110596] and normal direction [1.0, 0.0, 0.0]. The plane should have no offset.
Next, save this time series to a collection of .vtm files. The base file name for the time series is "subseries-of-time-series/results/{agent_mode}/canslices.vtm" and the suffix is '_%d'. Only save time steps with index between 10 and 20 inclusive, counting by 3.
Next, load the files "subseries-of-time-series/results/{agent_mode}/canslices_10.vtm", "subseries-of-time-series/results/{agent_mode}/canslices_13.vtm", "subseries-of-time-series/results/{agent_mode}/canslices_16.vtm", and "subseries-of-time-series/results/{agent_mode}/canslices_19.vtm" in multi-block format.
Finally, show the multi-block data set you just loaded.
Save a screenshot to the file "subseries-of-time-series/results/{agent_mode}/subseries-of-time-series.png".
(Optional, but must save if use paraview) Save the paraview state as "subseries-of-time-series/results/{agent_mode}/subseries-of-time-series.pvsm".
(Optional, but must save if use python script) Save the python script as "subseries-of-time-series/results/{agent_mode}/subseries-of-time-series.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Data Loading and Block Selection: Are the specified element blocks properly loaded and the slice plane correctly applied?
2. Multi-block Loading: Are the exported VTM files successfully loaded back as a multi-block dataset?
3. Final Visualization: Is the multi-block dataset properly displayed showing the sliced geometry from the time series?
# Case 46: write-ply
- vars:
question: |
I would like to use ParaView to visualize a dataset.
Create a wavelet object. Change the view size to 400x400.
Show the wavelet object and reset the camera to fit the data.
Next, create a contour of wavelet object from the dataset "RTData".
The contour should have isosurfaces at the following values: 97.222075, 157.09105, 216.96002500000003, and 276.829.
Show the contour and color it with the same colormap that is used for coloring "RTData".
Finally, save the contour in PLY format to the file "write-ply/results/{agent_mode}/PLYWriterData.ply".
Save the visualization image as "write-ply/results/{agent_mode}/write-ply.png".
(Optional, but must save if use paraview) Save the paraview state as "write-ply/results/{agent_mode}/write-ply.pvsm".
(Optional, but must save if use python script) Save the python script as "write-ply/results/{agent_mode}/write-ply.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Cube Creation: Is the cube object properly created and displayed with correct geometry?
2. PLY Import: Is the exported PLY file correctly loaded back into ParaView maintaining geometric fidelity?
3. Visualization Quality: Does the imported cube display properly with correct surface representation and rendering?
# Case 47: climate
- vars:
question: |
I would like to use ParaView to visualize a dataset of ocean currents.
Read in the file named "climate/data/climate.vtp".
Apply a calculator filter to compute the following function:
(-velocity_X*sin(coordsX*0.0174533) + velocity_Y*cos(coordsX*0.0174533)) * iHat + (-velocity_X * sin(coordsY*0.0174533) * cos(coordsX*0.0174533) - velocity_Y * sin(coordsY*0.0174533) * sin(coordsX*0.0174533) + velocity_Z * cos(coordsY*0.0174533)) * jHat + 0*kHat
Render the computed values using a tube filter with 0.05 as the tube radius.
Color the tubes by the magnitude of the velocity.
Light the tubes with the maximum shininess and include normals in the lighting.
Add cone glyphs to show the direction of the velocity.
The glyphs are composed of 10 polygons, having a radius 0 0.15, a height of 0.5, and a scaling factor of 0.5.
View the result in the -z direction.
Adjust the view so that the tubes occupy the 90% of the image.
Save a screenshot of the result, 2294 x 1440 pixels, white background, in the filename "climate/results/{agent_mode}/climate.png".
(Optional, but must save if use paraview) Save the paraview state as "climate/results/{agent_mode}/climate.pvsm".
(Optional, but must save if use python script) Save the python script as "climate/results/{agent_mode}/climate.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Tube Visualization: Are the tubes rendered with correct radius (0.05), colored by velocity magnitude, and proper lighting with maximum shininess?
2. Cone Glyph Direction: Are the cone glyphs properly configured with specified parameters and showing velocity direction accurately?
3. View Configuration: Is the visualization displayed from -z direction with appropriate scaling and white background as specified?
# Case 48: materials
- vars:
question: |
Compare two datasets in two views side by side each 900 pixels wide x 1400 pixels high.
Read the dataset "materials/data/materials_prediction.vtr" in the left view and "materials/data/materials_ground_truth.vtr" in the right view.
In both views, convert the "Intensity" and "Phase" variables from cell to point data.
In both views, take an isovolume of the "Intensity" variable in the range of [0.2, 1.0], clipped with a plane at (32.0, 32.0, 32.0) and +x normal direction.
Color both views with the Viridis (matplotlib) color map for the "Phase" variable, scaled to the data range, including a colormap legend in both views.
Label the left view "NN Prediction" and the right view "Ground Truth".
Orient the camera to look in the (-1, 0, -1) direction, with the datasets fitting in the views.
Save the visualization image as "materials/results/{agent_mode}/materials.png".
(Optional, but must save if use paraview) Save the paraview state as "materials/results/{agent_mode}/materials.pvsm".
(Optional, but must save if use python script) Save the python script as "materials/results/{agent_mode}/materials.py".
Do not save any other files, and always save the visualization image.
assert:
- type: llm-rubric
subtype: vision
value: |
1. Side-by-Side Comparison: Are both datasets properly displayed in side-by-side views with correct dimensions and labeling?
2. Data Conversion and Filtering: Are the Intensity and Phase variables correctly converted to point data and isovolume filtering applied?
3. Clipping and Color Mapping: Is the plane clipping correctly applied and Viridis colormap properly used for Phase variable?
4. Camera and Layout: Is the camera positioned correctly in (-1, 0, -1) direction with appropriate fitting and legends visible? |