You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Inference with bitsandbytes for Llama 3.2 vision model as per the blog https://huggingface.co/blog/llama32 is taking more time than the default model load.
#2510
Open
aabbhishekksr opened this issue
Dec 4, 2024
· 1 comment
The text was updated successfully, but these errors were encountered:
aabbhishekksr
changed the title
Inference with bitsandbytes for Llama 3.2 vision model as per the blog https://huggingface.co/blog/llama32
Inference with bitsandbytes for Llama 3.2 vision model as per the blog https://huggingface.co/blog/llama32 is taking more time than the default model load.
Dec 4, 2024
import torch
from transformers import MllamaForConditionalGeneration, AutoProcessor
from transformers import BitsAndBytesConfig
import time
from PIL import Image
import requests
No description provided.
The text was updated successfully, but these errors were encountered: