Fine tune Llava model on VQA pairs using NIH XRAY images

The Goal very simple: to fine-tune a vision-language model (LLaVA or CogVLM2) on a meaningful(but as small as feasibnle) subset of the NIH ChestX-ray14 dataset. The goal is to build a system capable of Visual Question Answering (VQA) tailored to medical diagnostics… (Budget: ₹600 – ₹1500 INR, Jobs: AI Chatbot Development, AI Development, AI Model Development, AI Research, AI Text-to-text, Computer Vision, Data Processing, Data Science, Machine Learning (ML), Natural Language Processing)

Comments

Leave a Reply

Your email address will not be published. Required fields are marked *