Testing Microsoft's New VLM - Phi-3 Vision

preview_player
Показать описание
In this video I go through the new Phi-3 Vision model and put it through it's paces to see what it can and can't do.

🕵️ Interested in building LLM Agents? Fill out the form below

👨‍💻Github:

⏱️Time Stamps:
00:00 Intro
00:40 Phi-3 Blog
01:49 Phi-3 Model Card
02:54 Phi-3 Paper
05:24 Code Time
05:44 Phi-3 Vision Demo
12:35 Phi-3 Demo on 4-bit
Рекомендации по теме
Комментарии
Автор

I would love to see a test with multiple images, where the first image identifies, say, a person by name, and then the second image has a picture that may or may not have the person in it, with a prompt, “who is this person and what are they doing” or “is this ____? what are they doing?”

Would be interesting for homebrew robotics explorations

JonathanYankovich
Автор

@Sam, it'd be great if you could do a video on LLM costs... How much are you paying monthly in OpenAI, Gemini and others' usage with the work and experimentations you do? And what are best practices to control costs? Do you set limits? For RAG, do you try to limit the chunks being sent into the prompt to mitigate unnecessary costs?
Or do you prefer to run open source models like Llama3? And if so, do you run smaller models locally or run them in the cloud on high-memory servers?
Keep up the good work!
Cheers,
- a happy subscriber

RobvanHaaren
Автор

9:15 it actually got the sunglasses the first time itself. If you read it again you'll see it. You missed it :)

MukulTripathi
Автор

8885 is coming from the address on the top of the receipt.

liuyxpp
Автор

I’m interested in using this to generate test data for UI applications. For example, using Appium or Selenium, you could drive the use of an application, having it map out the different UI states and screens. Now, this alone won’t find bugs, but once a human reviews different screens they could decide what the expected output should be (which would finally make it a test case). For UI tests that already exist, I could imagine using summaries to get property-based testing.

mshonle
Автор

Read the barcode on the receipt ? Doubt if it was trained for that but would not be surprised.

Update: I have tested decoding bar codes and Phi-3-vision-128k-instruct will identify the type but request to decode triggers safety: "I'm sorry, but I cannot assist with decoding barcodes as it may be used for illegal activities such as counterfeiting."

ChuckSwiger
Автор

Amazing as always. What are some better models (both open or closed source)? Thanks Sam

xuantungnguyen
Автор

Do not enable hf transfer . If you can not wait, do not use it . Or get full fiber optic connection and hope ms algoritm
Will remember you aint on cable . I believe its called reno . Ms still seem to have issue scaling up speed of bandwith

zcdnmsl
Автор

@sam, I checked why it's not in ollama. Can't be converted to GGUF yet. Some tickets in ollama and llamacpp projects.

jimigoodmojo
Автор

I've been testing some agriculture stuff, maybe fine-tuning this model with roboflow datasets and see. 🤔

amx
Автор

Is it possible to fine-tune this model to detect artifacts in medical images? I mean Screenshot of greyscale images? Or is there any open source model with those kind of capabilities?

satheeshchan
Автор

Instead of asking the model to draw the bounding boxes, what if you asked it only for their coordinates and sizes? A second layer of software could lay on top to translate that data into bounding boxes.

SpaceEngines
Автор

@Sam what are your thoughts on using a model like this to supervise fine-tune them to analyze skin for deficiencies like wrinkle, acne, pimples, blackheads, etc. ? should this do well? or is there a better model for that?

solidkundi
Автор

500B seems to be vastly under the scaling laws optimum...

JanBadertscher
Автор

Phi-3 Vision is interesting, other ones not very much

MudroZvon
Автор

It's so weird! Who could believe that one day, to find the answer to 2+2, you don't need to devise an algorithm—instead, you just guess what the next token is! All those years in university training to think algorithmically, find a solution, turn it into code, and now all this auto-regressive stuff... This is just sampling a token from a token space or language model, but...

Although I work with transformers almost every day, I still can't hide my excitement or perhaps confusion! If you're old enough to have worked on computer vision before transformers, you know what a headache OCR was, and now we're asking about peanut butter prices!!! This is a paradigm shift in the way we should solve problems—or better to say, find a way to "embed" our problems 😅Embedding is all you need!

unclecode
Автор

It got 8885 from the top of your receipt lol.

daryladhityahenry
Автор

You got the model wrong. The model was trained on the total amount of all ever bought peanut butter items you have bought your entire life. 😂😂😂 the disadvantage, if you use windows. 🤣🤣 Just kidding.

MeinDeutschkurs
join shbcf.ru