Steve Li
CHNtentes
AI & ML interests
None yet
Recent Activity
new activity
1 day ago
Qwen/Qwen2.5-Coder-32B-Instruct:Thieves!
new activity
5 days ago
unsloth/DeepSeek-V3-GGUF:why use q5 for key cache?
new activity
5 days ago
bullerwins/DeepSeek-V3-GGUF:这我也跑不动啊你们都是在拿什么玩意玩啊。我勒个公鸡母鸡战斗鸡啊
Organizations
None yet
CHNtentes's activity
Thieves!
6
#36 opened 2 days ago
by
supercharge19
why use q5 for key cache?
1
#7 opened 5 days ago
by
CHNtentes
这我也跑不动啊你们都是在拿什么玩意玩啊。我勒个公鸡母鸡战斗鸡啊
5
#8 opened 8 days ago
by
ss996
Can we use "temporal tiling support" with these gguf models?
1
#10 opened 16 days ago
by
CHNtentes
minimum vram?
11
#9 opened 20 days ago
by
CHNtentes
GGUF weights?
7
#1 opened 22 days ago
by
luijait
How was r7b?
6
#3 opened about 1 month ago
by
MRU4913
transformers version?
1
#5 opened about 2 months ago
by
CHNtentes
Q4_0, Q4_1, Q5_0, Q5_1 can be dropped?
1
#1 opened 2 months ago
by
CHNtentes
Where is 't5xxl.safetensors' ?
4
#12 opened 2 months ago
by
ajavamind
Hardware requirements
6
#10 opened 4 months ago
by
ZahirHamroune
T4 - bfloat 16 not support
10
#2 opened 4 months ago
by
SylvainV
🚩 Report: Spam
#150 opened 4 months ago
by
CHNtentes
Is it using ggml to compute?
1
#30 opened 5 months ago
by
CHNtentes
For the fastest inference on 12GB VRAM, are the following GGUF models appropriate to use?
3
#4 opened 5 months ago
by
ViratX
Inquiry on Minimum Configuration and Cost for Running Gemma-2-9B Model Efficiently
3
#39 opened 5 months ago
by
ltkien2003
Error in readme?
1
#6 opened 5 months ago
by
CHNtentes
Good work!
1
#1 opened 5 months ago
by
CHNtentes
Compared to the regular FP8 model, what is the better performance of the 8BIT model here
4
#16 opened 5 months ago
by
demo001s