Just came after seeing post on LinkedIn as I follow you there - going to try on weekends
@airoundtable
Ай бұрын
I hope you enjoy the content!
@divye.ruhela
3 күн бұрын
Great video! Subbed! Can you direct me to the resources for how one could train llava to add new classes to it? For instance, teach it to recognize and describe traditional battle poses or describe dishes with their traditional names, etc.?
@airoundtable
2 күн бұрын
Thanks. From the technical stand point, what you want to do is very similar with what I did in the video. I also explained how you need to prepare your data for that scenario in the video. There is also a notebook that gives you the hints for data preparation. from there it is just passing the right data to the model and that's it. You have access to everything that you need with this video and the project in my github repository
@raminguyen7940
7 күн бұрын
I am currently working with this model: LLaVA-v1.6 Mistral 7B. I have my own image dataset, but the images are stored in array format. I would appreciate some guidance on how to convert these images into a suitable input for the model. Below is the code I am using: prompt = "What are the things I should be cautious about when I visit this place? What should I bring with me?" max_output_token = 500 prompt = f"[INST] {prompt} [/INST]" inputs = processor(prompt, image, return_tensors="pt").to("cuda:0") output = model.generate(**inputs, max_new_tokens=max_output_token) response = processor.decode(output[0], skip_special_tokens=True) pprint(response)
@airoundtable
6 күн бұрын
I responded to you on LinkedIn
@MuhammadAdnan-tq3fx
Ай бұрын
Thanks for this informative video. I have a question: how can we perform distributed model training on multiple GPUs? In this video, the training is performed on a single 80GB GPU. For example, if we want to perform the training on multiple GPUs (48,48GB) than what should we do?
@airoundtable
28 күн бұрын
The concept is called model sharding where the architecture will be distributed over multiple GPUs. I haven't done it with LLAVA but to understand it, you can have a look at this pytorch blog: pytorch.org/blog/introducing-pytorch-fully-sharded-data-parallel-api/ In pytorch the class that does this is called `FullyShardedDataParallel`. You can find more info about it here: pytorch.org/docs/stable/fsdp.html
@PareshPawar-y5w
28 күн бұрын
What do you suggest for that making Python GUI app using tkinkter? or do you prefer other one? do you have any video for it? Thank you in advance!!! Big fan of your teaching!!!
@airoundtable
26 күн бұрын
Thanks! I haven't used thinkter and I don't have any videos for it in the channel
Пікірлер: 12