| --- |
| pretty_name: "DyT Composition Study Artifacts" |
| license: cc-by-4.0 |
| language: |
| - en |
| tags: |
| - machine-learning |
| - transformers |
| - layernorm |
| - dynamic-tanh |
| - activation-bounding |
| - reproducibility |
| size_categories: |
| - n<1K |
| configs: |
| - config_name: artifact_index |
| default: true |
| data_files: |
| - split: train |
| path: data/artifact_index.jsonl |
| --- |
| |
| # DyT Composition Study Artifacts |
|
|
| [](https://arxiv.org/abs/2604.23434) |
| [](https://github.com/lucky-verma/dyt-composition-study) |
| [](https://creativecommons.org/licenses/by/4.0/) |
|
|
| This dataset contains sanitized result manifests and analysis outputs for |
| **[When Does Removing LayerNorm Help? Activation Bounding as a Regime-Dependent Implicit Regularizer](https://arxiv.org/abs/2604.23434)**. |
|
|
| DOI: https://doi.org/10.48550/arXiv.2604.23434 |
|
|
| ## Contents |
|
|
| The artifacts include aggregate training metrics, saturation measurements, statistical-test summaries, predictor-validation outputs, table-source manifests, and selected aggregate analysis files used by the public code repository. |
|
|
| The Dataset Viewer table is an index of the artifact files. The machine-readable result artifacts are stored under `results/`. |
|
|
| This is not a natural-language training dataset. It does not redistribute WikiText, OpenWebText, LAMBADA, BLIMP, model checkpoints, or raw training logs. |
|
|
| ## Intended Use |
|
|
| Use this artifact bundle to: |
|
|
| - inspect the machine-readable results behind the paper; |
| - reproduce paper tables and consistency checks; |
| - compare DyT, LayerNorm, RMSNorm, HardTanh, DiffAttn, and related controls at the reported scales; |
| - audit provenance for reported quantitative claims. |
|
|
| ## Limitations |
|
|
| - The experiments are compute-limited and below Chinchilla-optimal training. |
| - The included files are result artifacts, not full raw training traces or checkpoints. |
| - The saturation diagnostic should be treated as a per-deployment calibration cue, not a universal rule. |
| - Raw public datasets retain their original licenses and are not mirrored here. |
|
|
| ## Licensing |
|
|
| The result artifacts in this dataset are released under CC BY 4.0. |
|
|
| The associated GitHub code is released under the MIT License. |
|
|
| ## Citation |
|
|
| ```bibtex |
| @misc{verma2026dytcomposition, |
| title = {When Does Removing LayerNorm Help? Activation Bounding as a Regime-Dependent Implicit Regularizer}, |
| author = {Verma, Lucky}, |
| year = {2026}, |
| publisher = {arXiv}, |
| doi = {10.48550/arXiv.2604.23434}, |
| url = {https://arxiv.org/abs/2604.23434}, |
| eprint = {2604.23434}, |
| archivePrefix = {arXiv}, |
| primaryClass = {cs.LG} |
| } |
| ``` |
|
|