Rayen
Lissanro
AI & ML interests
None yet
Recent Activity
new activity about 2 months ago
zai-org/GLM-5:Please release a model with native 4-bit quantization new activity 5 months ago
ubergarm/Ling-1T-GGUF:128K context does not work (possibly because YaRN meta information is missing?) new activity 7 months ago
Jinx-org/Jinx-gpt-oss-20b-GGUF:MXFP4_MOEOrganizations
None yet
Please release a model with native 4-bit quantization
👀➕ 11
4
#4 opened about 2 months ago
by
calycekr
128K context does not work (possibly because YaRN meta information is missing?)
➕ 1
2
#8 opened 5 months ago
by
Lissanro
MXFP4_MOE
🔥🚀 1
11
#1 opened 8 months ago
by
marcelone
Incorrect Model Uploaded
🤗👍 18
6
#8 opened 7 months ago
by
noteventhrice
Context length: is it 128K (as mentioned in the model card) or 160K (as specified in config.json)?
1
#17 opened 8 months ago
by
Lissanro
Please consider creating ik_llama.cpp compatible quants (without llama.cpp-specific MLA tensors)
1
#1 opened 11 months ago
by
Lissanro
chat_template.json is missing
2
#1 opened about 1 year ago
by
Lissanro
chat_template.json is missing
2
#1 opened about 1 year ago
by
Lissanro
Tell me how do you feel about this model without telling me how do you feel about this model
4
#5 opened about 1 year ago
by
MrDevolver
Is this model native 128K context length, or YaRN extended?
7
#28 opened about 1 year ago
by
danielhanchen
Doesn't Generate `<think>` tags
3
#25 opened about 1 year ago
by
bingw5
Works great on oobabooga, but always ends with assistant
1
#3 opened almost 2 years ago
by
Noodlz
This could likely be dewokefied and possible even improved using mergekit's new 'Model Stock' method!
🔥 1
31
#5 opened almost 2 years ago
by
jukofyork