sijieli's picture
Upload ScaleBench dataset
691a758 verified
"""Scaling laws for parallel (tensor/pipeline) scaling.
X columns: [num_params (N), parallel_size (S)]
"""
from typing import Literal
import benchmark.dataset.utils as utils
_EPS = 1e-12
# sl_1 (6p): c0 + cN * N^(-α) + cS * S^(-β) + cNS * N^(-α) * S^(-β)
# theta: [c0, cN, alpha, cS, beta, cNS]
def sl_1(theta, X, backend: Literal["numpy", "jax", "torch"] = "jax"):
ops = utils.get_ops(backend)
xp = ops.xp
X = ops.asarray(X, atleast_2d=True)
theta = ops.asarray(theta, atleast_2d=True)
N = ops.clamp_min(X[:, 0], _EPS)
S = ops.clamp_min(X[:, 1], _EPS)
c0, cN, alpha, cS, beta, cNS = [theta[:, i] for i in range(6)]
Na = N[None, :] ** (-alpha[:, None]) # (B, M)
Sb = S[None, :] ** (-beta[:, None]) # (B, M)
NaSb = Na * Sb
pred = c0[:, None] + cN[:, None] * Na + cS[:, None] * Sb + cNS[:, None] * NaSb
# Jacobian: d/d[c0, cN, alpha, cS, beta, cNS]
ones = pred * 0.0 + 1.0
logN = xp.log(ops.clamp_min(N[None, :], _EPS)) # (1, M) or (B, M) after broadcast
logS = xp.log(ops.clamp_min(S[None, :], _EPS))
d_c0 = ones
d_cN = Na
# d(Na)/d(alpha) = d(N^(-alpha))/d(alpha) = -N^(-alpha)*log(N) = -Na*logN
d_alpha = (cN[:, None] * Na + cNS[:, None] * NaSb) * (-logN)
d_cS = Sb
# d(Sb)/d(beta) = -Sb*logS
d_beta = (cS[:, None] * Sb + cNS[:, None] * NaSb) * (-logS)
d_cNS = NaSb
jac = ops.stack([d_c0, d_cN, d_alpha, d_cS, d_beta, d_cNS], axis=-1)
if pred.shape[0] == 1:
return pred[0], jac[0]
return pred, jac
# sl_2 (5p): c0 + cN * N^(-α) + cS * S^(-β)
# theta: [c0, cN, alpha, cS, beta]
def sl_2(theta, X, backend: Literal["numpy", "jax", "torch"] = "jax"):
ops = utils.get_ops(backend)
xp = ops.xp
X = ops.asarray(X, atleast_2d=True)
theta = ops.asarray(theta, atleast_2d=True)
N = ops.clamp_min(X[:, 0], _EPS)
S = ops.clamp_min(X[:, 1], _EPS)
c0, cN, alpha, cS, beta = [theta[:, i] for i in range(5)]
Na = N[None, :] ** (-alpha[:, None])
Sb = S[None, :] ** (-beta[:, None])
pred = c0[:, None] + cN[:, None] * Na + cS[:, None] * Sb
# Jacobian: d/d[c0, cN, alpha, cS, beta]
ones = pred * 0.0 + 1.0
logN = xp.log(ops.clamp_min(N[None, :], _EPS))
logS = xp.log(ops.clamp_min(S[None, :], _EPS))
d_c0 = ones
d_cN = Na
d_alpha = cN[:, None] * Na * (-logN)
d_cS = Sb
d_beta = cS[:, None] * Sb * (-logS)
jac = ops.stack([d_c0, d_cN, d_alpha, d_cS, d_beta], axis=-1)
if pred.shape[0] == 1:
return pred[0], jac[0]
return pred, jac
# sl_3 (4p): a * N^b + c / (1 + S) + d
# theta: [a, b, c, d]
def sl_3(theta, X, backend: Literal["numpy", "jax", "torch"] = "jax"):
ops = utils.get_ops(backend)
xp = ops.xp
X = ops.asarray(X, atleast_2d=True)
theta = ops.asarray(theta, atleast_2d=True)
N = ops.clamp_min(X[:, 0], _EPS)
S = ops.clamp_min(X[:, 1], _EPS)
a, b, c, d = [theta[:, i] for i in range(4)]
Nb = N[None, :] ** b[:, None]
inv_1pS = 1.0 / (1.0 + S[None, :])
pred = a[:, None] * Nb + c[:, None] * inv_1pS + d[:, None]
# Jacobian: d/d[a, b, c, d]
ones = pred * 0.0 + 1.0
logN = xp.log(ops.clamp_min(N[None, :], _EPS))
d_a = Nb
d_b = a[:, None] * Nb * logN
d_c = inv_1pS + ones * 0.0 # broadcast
d_d = ones
jac = ops.stack([d_a, d_b, d_c, d_d], axis=-1)
if pred.shape[0] == 1:
return pred[0], jac[0]
return pred, jac
# sl_4 (4p): a * N^b + c * S^(-0.5) + d (fixed beta=0.5)
# theta: [a, b, c, d]
def sl_4(theta, X, backend: Literal["numpy", "jax", "torch"] = "jax"):
ops = utils.get_ops(backend)
xp = ops.xp
X = ops.asarray(X, atleast_2d=True)
theta = ops.asarray(theta, atleast_2d=True)
N = ops.clamp_min(X[:, 0], _EPS)
S = ops.clamp_min(X[:, 1], _EPS)
a, b, c, d = [theta[:, i] for i in range(4)]
Nb = N[None, :] ** b[:, None]
S_inv_half = S[None, :] ** (-0.5)
pred = a[:, None] * Nb + c[:, None] * S_inv_half + d[:, None]
# Jacobian: d/d[a, b, c, d]
ones = pred * 0.0 + 1.0
logN = xp.log(ops.clamp_min(N[None, :], _EPS))
d_a = Nb
d_b = a[:, None] * Nb * logN
d_c = S_inv_half + ones * 0.0 # broadcast
d_d = ones
jac = ops.stack([d_a, d_b, d_c, d_d], axis=-1)
if pred.shape[0] == 1:
return pred[0], jac[0]
return pred, jac
# sl_5 (4p): (A / (N * (k * log(S) + 1)))^α + E
# theta: [A, k, alpha, E]
def sl_5(theta, X, backend: Literal["numpy", "jax", "torch"] = "jax"):
ops = utils.get_ops(backend)
xp = ops.xp
X = ops.asarray(X, atleast_2d=True)
theta = ops.asarray(theta, atleast_2d=True)
N = ops.clamp_min(X[:, 0], _EPS)
S = ops.clamp_min(X[:, 1], 1.0)
A, k, alpha, E = [theta[:, i] for i in range(4)]
logS = xp.log(S[None, :])
denom = N[None, :] * (k[:, None] * logS + 1.0)
denom = ops.clamp_min(denom, _EPS)
base = A[:, None] / denom
base = ops.clamp_min(base, _EPS)
power = base ** alpha[:, None] # base^alpha
pred = power + E[:, None]
# Jacobian: d/d[A, k, alpha, E]
# Let f = base^alpha, pred = f + E
# base = A / denom, denom = N*(k*logS + 1)
ones = pred * 0.0 + 1.0
log_base = xp.log(ops.clamp_min(base, _EPS))
# d(pred)/dA: d(base^alpha)/dA = alpha * base^(alpha-1) * (1/denom)
# = alpha * base^alpha * (1/base) * (1/denom) = alpha * power / A[:, None]
# But safer: alpha * power * (1/base) * d(base)/dA = alpha * power / base * (1/denom)
# = alpha * power / (A/denom) * (1/denom) = alpha * power / A
d_A = alpha[:, None] * power / ops.clamp_min(A[:, None], _EPS)
# d(pred)/dk: d(base^alpha)/dk = alpha * base^(alpha-1) * d(base)/dk
# d(base)/dk = A * d(1/denom)/dk = -A * logS * N / denom^2 = -base * logS * N / denom
# Actually: d(base)/dk = -A * N * logS / denom^2 = -(A/denom) * (N*logS)/denom = -base * logS / (k*logS+1)
# Simpler: d(base)/dk = -A * N * logS / denom^2
# So d(f)/dk = alpha * power / base * (-A * N * logS / denom^2)
# = alpha * power * (-N * logS / denom) / base ... hmm
# Let's use: d(log(base))/dk = d(log(A) - log(denom))/dk = -d(log(denom))/dk
# d(log(denom))/dk = N*logS / denom ... but denom = N*(k*logS+1)
# d(denom)/dk = N*logS, so d(log(denom))/dk = N*logS/denom = logS/(k*logS+1)
# d(f)/dk = f * alpha * d(log(base))/dk = -power * alpha * logS / (k[:, None] * logS + 1.0)
k_logS_1 = ops.clamp_min(k[:, None] * logS + 1.0, _EPS)
d_k = -power * alpha[:, None] * logS / k_logS_1
# d(pred)/dalpha: d(base^alpha)/dalpha = base^alpha * log(base) = power * log(base)
d_alpha = power * log_base
d_E = ones
jac = ops.stack([d_A, d_k, d_alpha, d_E], axis=-1)
if pred.shape[0] == 1:
return pred[0], jac[0]
return pred, jac
# sl_6 (4p): c0 + c1 * (N^(-α) + S^(-β))
# theta: [c0, c1, alpha, beta]
def sl_6(theta, X, backend: Literal["numpy", "jax", "torch"] = "jax"):
ops = utils.get_ops(backend)
xp = ops.xp
X = ops.asarray(X, atleast_2d=True)
theta = ops.asarray(theta, atleast_2d=True)
N = ops.clamp_min(X[:, 0], _EPS)
S = ops.clamp_min(X[:, 1], _EPS)
c0, c1, alpha, beta = [theta[:, i] for i in range(4)]
Na = N[None, :] ** (-alpha[:, None])
Sb = S[None, :] ** (-beta[:, None])
sumNS = Na + Sb
pred = c0[:, None] + c1[:, None] * sumNS
# Jacobian: d/d[c0, c1, alpha, beta]
ones = pred * 0.0 + 1.0
logN = xp.log(ops.clamp_min(N[None, :], _EPS))
logS = xp.log(ops.clamp_min(S[None, :], _EPS))
d_c0 = ones
d_c1 = sumNS
d_alpha = c1[:, None] * Na * (-logN)
d_beta = c1[:, None] * Sb * (-logS)
jac = ops.stack([d_c0, d_c1, d_alpha, d_beta], axis=-1)
if pred.shape[0] == 1:
return pred[0], jac[0]
return pred, jac
# sl_7 (4p): L0 + A * N^(-α) / (1 + k * ln(S))
# theta: [L0, A, alpha, k]
def sl_7(theta, X, backend: Literal["numpy", "jax", "torch"] = "jax"):
ops = utils.get_ops(backend)
xp = ops.xp
X = ops.asarray(X, atleast_2d=True)
theta = ops.asarray(theta, atleast_2d=True)
N = ops.clamp_min(X[:, 0], _EPS)
S = ops.clamp_min(X[:, 1], 1.0)
L0, A, alpha, k = [theta[:, i] for i in range(4)]
Na = N[None, :] ** (-alpha[:, None])
logS = xp.log(S[None, :])
denom = 1.0 + k[:, None] * logS
denom = ops.clamp_min(denom, _EPS)
numer = A[:, None] * Na
frac = numer / denom
pred = L0[:, None] + frac
# Jacobian: d/d[L0, A, alpha, k]
ones = pred * 0.0 + 1.0
logN = xp.log(ops.clamp_min(N[None, :], _EPS))
d_L0 = ones
d_A = Na / denom
# d(frac)/d(alpha) = A * d(Na)/d(alpha) / denom = A * (-Na*logN) / denom = -frac * logN
d_alpha = -frac * logN
# d(frac)/dk = -numer * logS / denom^2 = -frac * logS / denom
d_k = -frac * logS / denom
jac = ops.stack([d_L0, d_A, d_alpha, d_k], axis=-1)
if pred.shape[0] == 1:
return pred[0], jac[0]
return pred, jac
# sl_8 (4p): (a * N^b + c) / (1 + d * log(S))
# theta: [a, b, c, d]
def sl_8(theta, X, backend: Literal["numpy", "jax", "torch"] = "jax"):
ops = utils.get_ops(backend)
xp = ops.xp
X = ops.asarray(X, atleast_2d=True)
theta = ops.asarray(theta, atleast_2d=True)
N = ops.clamp_min(X[:, 0], _EPS)
S = ops.clamp_min(X[:, 1], 1.0)
a, b, c, d = [theta[:, i] for i in range(4)]
Nb = N[None, :] ** b[:, None]
logS = xp.log(S[None, :])
numer = a[:, None] * Nb + c[:, None]
denom = 1.0 + d[:, None] * logS
denom = ops.clamp_min(denom, _EPS)
pred = numer / denom
# Jacobian: d/d[a, b, c, d]
logN = xp.log(ops.clamp_min(N[None, :], _EPS))
inv_denom = 1.0 / denom
d_a = Nb * inv_denom # d(numer)/da = N^b
d_b = a[:, None] * Nb * logN * inv_denom # d(numer)/db = a * N^b * logN
d_c = inv_denom # d(numer)/dc = 1
d_d = -numer * logS * inv_denom ** 2 # d(denom)/dd = logS → quotient rule
jac = ops.stack([d_a, d_b, d_c, d_d], axis=-1)
if pred.shape[0] == 1:
return pred[0], jac[0]
return pred, jac
# sl_9 (3p): A * N^(-α) * S^(-β)
# theta: [A, alpha, beta]
def sl_9(theta, X, backend: Literal["numpy", "jax", "torch"] = "jax"):
ops = utils.get_ops(backend)
xp = ops.xp
X = ops.asarray(X, atleast_2d=True)
theta = ops.asarray(theta, atleast_2d=True)
N = ops.clamp_min(X[:, 0], _EPS)
S = ops.clamp_min(X[:, 1], _EPS)
A, alpha, beta = [theta[:, i] for i in range(3)]
Na = N[None, :] ** (-alpha[:, None])
Sb = S[None, :] ** (-beta[:, None])
pred = A[:, None] * Na * Sb
# Jacobian: d/d[A, alpha, beta]
logN = xp.log(ops.clamp_min(N[None, :], _EPS))
logS = xp.log(ops.clamp_min(S[None, :], _EPS))
d_A = Na * Sb # pred / A
d_alpha = pred * (-logN) # d(N^(-alpha))/dalpha = -N^(-alpha)*logN
d_beta = pred * (-logS)
jac = ops.stack([d_A, d_alpha, d_beta], axis=-1)
if pred.shape[0] == 1:
return pred[0], jac[0]
return pred, jac
# sl_10 (4p): (A * N^(-α) + E) * S^(-β)
# theta: [A, alpha, E, beta]
def sl_10(theta, X, backend: Literal["numpy", "jax", "torch"] = "jax"):
ops = utils.get_ops(backend)
xp = ops.xp
X = ops.asarray(X, atleast_2d=True)
theta = ops.asarray(theta, atleast_2d=True)
N = ops.clamp_min(X[:, 0], _EPS)
S = ops.clamp_min(X[:, 1], _EPS)
A, alpha, E, beta = [theta[:, i] for i in range(4)]
Na = N[None, :] ** (-alpha[:, None])
Sb = S[None, :] ** (-beta[:, None])
bracket = A[:, None] * Na + E[:, None] # (A*N^(-alpha) + E)
pred = bracket * Sb
# Jacobian: d/d[A, alpha, E, beta]
logN = xp.log(ops.clamp_min(N[None, :], _EPS))
logS = xp.log(ops.clamp_min(S[None, :], _EPS))
d_A = Na * Sb
# d(bracket)/dalpha = A * (-Na * logN), then * Sb
d_alpha = A[:, None] * Na * (-logN) * Sb
d_E = Sb
# d(pred)/dbeta = bracket * d(Sb)/dbeta = bracket * (-Sb * logS) = -pred * logS
d_beta = -pred * logS
jac = ops.stack([d_A, d_alpha, d_E, d_beta], axis=-1)
if pred.shape[0] == 1:
return pred[0], jac[0]
return pred, jac
PARAM_BOUNDS = {
# Dataset: N ∈ [5.4e8, 4.4e9], log(N) ∈ [20.1, 22.2]; S ∈ {1,2,4,8}; loss ∈ [0.98, 2.11]
# Groups (pile/stack) have different loss floors; fit per group.
# sl_1: [c0, cN, alpha, cS, beta, cNS] — c0+cN*N^(-alpha)+cS*S^(-beta)+cNS*N^(-alpha)*S^(-beta)
# Fit pile: [1.38, 112.7, 0.260, 0.097, 0.399, 5.69]; stack: [0.763, 66.0, 0.263, 0.049, 0.470, 5.39]
"sl_1": [(-5, 5), (0, 500), (0, 2), (-50, 50), (0, 2), (-100, 100)],
# sl_2: [c0, cN, alpha, cS, beta] — c0 + cN*N^(-alpha) + cS*S^(-beta)
# Fit pile: [1.364, 117.8, 0.261, 0.121, 0.399]; stack: [0.750, 70.7, 0.264, 0.070, 0.472]
"sl_2": [(-5, 5), (0, 500), (0, 2), (-50, 50), (0, 2)],
# sl_3: [a, b, c, d] — a*N^b + c/(1+S) + d; b<0 required (more params → lower loss)
# Fit pile: [117.9, -0.261, 0.174, 1.398]; stack: [70.7, -0.264, 0.112, 0.764]
"sl_3": [(0, 500), (-2, 0), (-5, 5), (-5, 5)],
# sl_4: [a, b, c, d] — a*N^b + c*S^(-0.5) + d; b<0 required
# Fit pile: [117.9, -0.261, 0.105, 1.381]; stack: [70.7, -0.264, 0.068, 0.753]
"sl_4": [(0, 500), (-2, 0), (-5, 5), (-5, 5)],
# sl_5: [A, k, alpha, E] — (A/(N*(k*log(S)+1)))^alpha + E
# A must be ~N*constant: A/N_min ∈ [0.02, 1.85] for physical predictions
# Fit pile: [1.95e8, 0.334, 0.196, 1.291]; stack: [1.14e7, 0.396, 0.199, 0.709]
"sl_5": [(0, 1e9), (-100, 100), (0, 3), (-5, 5)],
# sl_6: [c0, c1, alpha, beta] — c0 + c1*(N^(-alpha) + S^(-beta))
# Fit pile: [-115.8, 117.3, 0.261, ~0]; stack: [-69.4, 70.2, 0.264, ~0]
# c0 hits -115 for pile (old -100 bound was too tight); beta≈0 (structural degeneracy)
"sl_6": [(-200, 200), (-200, 200), (0, 2), (0, 2)],
# sl_7: [L0, A, alpha, k] — L0 + A*N^(-alpha)/(1+k*ln(S))
# Fit pile: [1.315, 48.1, 0.204, 0.055]; stack: [0.728, 30.6, 0.211, 0.065]
"sl_7": [(-5, 5), (0, 500), (0, 2), (-50, 50)],
# sl_8: [a, b, c, d] — (a*N^b + c)/(1+d*log(S)); b<0 required
# Fit pile: [118.1, -0.260, 1.472, 0.017]; stack: [71.0, -0.263, 0.812, 0.020]
"sl_8": [(0, 500), (-2, 0), (-5, 5), (-10, 10)],
# sl_9: [A, alpha, beta] — A*N^(-alpha)*S^(-beta)
# Fit pile: [7.73, 0.065, 0.017]; stack: [4.45, 0.067, 0.020] (small exponents)
"sl_9": [(0, 5000), (0, 2), (0, 2)],
# sl_10: [A, alpha, E, beta] — (A*N^(-alpha) + E)*S^(-beta)
# Fit pile: [118.1, 0.260, 1.471, 0.017]; stack: [71.0, 0.263, 0.812, 0.020]
"sl_10": [(0, 500), (0, 2), (-5, 5), (0, 2)],
}
LAW_REGISTRY = {
"sl_1": sl_1, "sl_2": sl_2, "sl_3": sl_3, "sl_4": sl_4, "sl_5": sl_5,
"sl_6": sl_6, "sl_7": sl_7, "sl_8": sl_8, "sl_9": sl_9, "sl_10": sl_10,
}
PARAM_COUNTS = {
"sl_1": 6, "sl_2": 5, "sl_3": 4, "sl_4": 4, "sl_5": 4,
"sl_6": 4, "sl_7": 4, "sl_8": 4, "sl_9": 3, "sl_10": 4,
}