Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Cannot reproduce Playground demo on the Web vs result on a locally running model #251

Open
jthecodemonk opened this issue Feb 20, 2025 · 2 comments

Comments

@jthecodemonk
Copy link

I see a difference in performance between the Playground model and when I run the model (moondream-2b-int8.mf) locally. I use the same prompt 'head'.

Issue 1: I don't even get the same number of detections when I compare the Playground and local model.
Playground output:
Image

Locally running model:
Image

Issue 2: There is a slight difference in the size of bounding boxes between the Playground and local model.
Playground output:
Image

Locally running model:
Image

In all cases, the Playground model seems to perform better.

  1. Which model is used in the Playground?
  2. Why is there a difference in size of the bounding boxes?
@pbtaffi
Copy link

pbtaffi commented Feb 21, 2025

I'm facing the same issue with the query. I tried all models locally and the results are incorrect while the playground web version and the API work exactly as expected. I would also like to know the model and configuration (if any) used in the playground and API.

Thanks.

@autmoate
Copy link

I also noticed that the results were different, which is why I tried to set up moondream2 with transformers, because the model here: huggingface moondream2
is larger and I had suspected that this might be a reason (maybe higher precision which may give better results?), but I also had to contribute to this issue issue 235 because there were problems getting it running.
In the end it worked (thanks to the help of @parsakhaz), I describe my solution in the issue, but it takes a very long time for inference (CPU) and I can‘t tell that results are much better especially compared to API?

In the meantime I use the API again, because there are also very generous daily request rates (thanks to @vikhyat).

Nevertheless, I would also be very interested in how this can be explained and which model is actually behind the API. That would be exciting to know and it would also be good to know whether this model will also be available for a local usage?

Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants