Model does not run with VLLM
2
#3 opened 18 days ago
by
aswad546
Any idea when the evaluation data will be in for this model? would like to know how the performance differ from unquantized version of the model
#2 opened 2 months ago
by
jahhs0n
Any chance your team is working on a 4-bit Llama-3.2-90B-Vision-Instruct-quantized.w4a16 version?
1
#1 opened 3 months ago
by
mrhendrey