Josh Warner
JDWarner
AI & ML interests
None yet
Recent Activity
new activity 1 day ago
z-lab/Qwen3.5-27B-DFlash:dflash with quantize model new activity 1 day ago
z-lab/Qwen3.5-27B-DFlash:FP8 work for base model or is 16-bit of 27B required? liked a model 2 days ago
arcee-ai/Trinity-Large-ThinkingOrganizations
None yet
dflash with quantize model
1
#5 opened 1 day ago
by
Shimon324
FP8 work for base model or is 16-bit of 27B required?
12
#2 opened 13 days ago
by
unoid
pruned version
๐๐ฅ 1
2
#16 opened 19 days ago
by
pirola
There's got to be a better way.
23
#6 opened 24 days ago
by
phil111
Recall from embed documents not as good as the original
5
#4 opened 26 days ago
by
o0Linny0o
A wild idea / suggestion...
๐ฅ 3
2
#4 opened about 1 month ago
by
MrDevolver
Consider releasing full BF16 weights
2
#1 opened about 1 month ago
by
JDWarner
good model
5
#1 opened about 1 month ago
by
Roman1111111
Work great on 3090 except for weird (...) generation
โค๏ธ 1
5
#1 opened about 1 month ago
by
ortegaalfredo
Qwopus with visual capabilities?
2
#19 opened about 1 month ago
by
AQLabs
Security/Compliance Audit: EU AI Act & NIST Exposure
๐ฅ 1
3
#8 opened about 1 month ago
by
tradeapollo
FP8 models
3
#1 opened about 1 month ago
by
ecopoiesis
IQ5_K 136.891 GiB
๐ฅ 2
30
#9 opened 2 months ago
by
Hunterx
Request: GGUF / quantized weights for Intern-S1-Pro
1
#7 opened 2 months ago
by
gileneo
INT8 quantization for KVCache on DGX Spark/GB10
4
#6 opened 2 months ago
by
JDWarner
This just trades general performance for domain specific gains.
๐ฅ๐ 16
11
#3 opened 8 months ago
by
phil111
Disable thinking mode in Jan-v1-4B model
2
#9 opened 8 months ago
by
vuhaix95