Quan Nguyen PRO
qnguyen3
AI & ML interests
None yet
Organizations
Moderation block via XiaoMiMo API Platform
2
#9 opened 3 months ago
by
qnguyen3
how to use onnx
#17 opened 10 months ago
by
qnguyen3
VSCODE + Cline + Ollama + Qwen2.5-Coder-32B-Instruct.Q8_0
3
#20 opened over 1 year ago
by
BigDeeper
Adding Evaluation Results
#2 opened over 1 year ago
by
leaderboard-pr-bot
Open LLM Leaderboard results
1
#3 opened over 1 year ago
by
SaisExperiments
thank you for making quants
1
#1 opened over 1 year ago
by
qnguyen3
Evaluate output results
1
#3 opened over 1 year ago
by
Quy1004
Why dataset tag?
7
#1 opened over 1 year ago
by
rombodawg
Transformers doesn't support it yet?
➕ 2
6
#2 opened over 1 year ago
by
mahiatlinux
Missing configuration_llava_qwen2.py and configuration_llava_qwen2.py ??
1
#1 opened over 1 year ago
by
nicolollo
Handling `flash_attn` Dependency for Non-GPU Environments
👍 ❤️ 11
20
#4 opened over 1 year ago
by
giacomopedemonte
Sample code for inference in Google Colab? RuntimeError: "slow_conv2d_cuda" not implemented for 'Byte'
1
#12 opened over 1 year ago
by
sanjeev-bhandari01
This model is amazing!
👍 1
3
#1 opened over 1 year ago
by
nicolollo
Leaderboard
1
#6 opened over 1 year ago
by
Stark2008
Multi-round conversation w/ PKV cache example code
4
#5 opened almost 2 years ago
by
Xenova
vilm/VinaLlama2-14B-arxiv vs vilm/VinaLlama2-14B
1
#1 opened almost 2 years ago
by
anhnh2002
Approach to reduce hallucination
8
#1 opened almost 2 years ago
by
LoneRanger44
Gặp vấn đề khi finetune
2
#2 opened about 2 years ago
by
104-wonohfor
Run on Macbook without flash_attn?
2
#1 opened almost 2 years ago
by
palebluewanders