โSKT-SURYA-H (2.544T) is officially out! โ Heterogeneous MoE โ 131K Context โ 3.76TB Weights (898 shards) โMassive respect to the team for keeping it open for the community! ๐ค๐ฎ๐ณ
After 2 Years of research and Hard Work . weโve crossed the 2.5T barrier! ๐ โSKT-SURYA-H is now live: 2.544 Trillion parameters powered by our unique Weight Manifold Fusion (WMF) technology. Sovereign AI for Bharat is no longer a dream. ๐ฎ๐ณ๐ง
โWe are thrilled to announce the launch of SKT-OMNI-CORPUS-146T-V1, a massive-scale, high-quality dataset designed to power the next generation of Foundation Models (LLMs) from scratch. โDeveloped at SKT AI LABS, this corpus is not just a collection of data; itโs a mission to decentralize high-grade AI training for regional languages and global knowledge.
โ๐ Key Highlights:
โโขโข Massive Scale: Targeting a multi-terabyte architecture for 146T-level tokenization.
โขโข โPure Quality: Curated from 500+ Elite Sources
โขโข โStructured for MoE: Perfectly sharded into 3.5GB standardized units (SKT-๐ป series) for seamless distributed training.
โ๐ค Open for Collaboration!
โWe are looking for AI researchers, CUDA engineers, and data scientists to join us in this journey of building Project Surya and the ST-X Series models. Whether it's optimization, custom tokenization, or architecture designโletโs build the future together.