LagPixelLOL
v2ray
AI & ML interests
Looking for compute sponsors, please contact me through my email 2282688304@qq.com!
Recent Activity
updated a model 1 day ago
x2ray/stuffs updated a dataset 10 days ago
v2ray/r-chatgpt-general-dump updated a dataset 12 days ago
v2ray/r-chatgpt-general-dumpOrganizations
It's so unsafe that it can generate inappropriate adult content easily
🤯👍 2
52
#21 opened 8 months ago
by
liougehooa
running with flashmla on A100s
1
#1 opened 10 months ago
by
ehartford
could you give me a reason why you ignore kv_a_proj_with_mqa layer when quantizing this model?
1
#10 opened 10 months ago
by
superahn
gshshdhdhhdididhxbcbdhdudhbdbdbdbehe
4
#3 opened 10 months ago
by
MYNK337
H800 has errors
1
#35 opened 10 months ago
by
yiyepialing
update vllm to 0.8.x and meet some trouble
3
#34 opened 11 months ago
by
HuggingLianWang
Stuck when run on 8xH100
1
#8 opened 11 months ago
by
Thai
why tokenizer_config.json changed for AWQ model.
2
#7 opened 11 months ago
by
rockcat-miao
Does FlashMLA support kv cache fp8 dtype and how to enable FlashMLA ?
10
#6 opened 12 months ago
by
CharlesLincoln
test
👀 3
158
#2 opened about 1 year ago
by
adamm-hf
How to Resolve "GLIBC_2.32 Not Found" Error When Deploying vLLM Environment?
8
#32 opened 12 months ago
by
lastsummerLi
Can the 4090 device run this model?
3
#3 opened 12 months ago
by
jinzhongwei
vllm crach with a slightly longer prompt
1
#4 opened 12 months ago
by
rockcat-miao
可以添加一下LICENSE文件吗?
2
#2 opened 12 months ago
by
adol-ch
Are there any updates to the recommended commands?
5
#27 opened about 1 year ago
by
NaiveYan
Why hasn't the MTP layer of the 61st layer been quantized?
1
#30 opened 12 months ago
by
yang001002
Is there any testing on the support for running on other memory capacities
1
#29 opened 12 months ago
by
HRan2004
Any one can run this model with SGlang framework?
5
#13 opened about 1 year ago
by
muziyongshixin
DeepSeek-R1-AWQ quantized model missing one layer of experts
4
#28 opened about 1 year ago
by
virilo