Cradawx@alien.topB to LocalLLaMA@poweruser.forumEnglish · 10 months agoShareGPT4V - New multi-modal model, improves on LLaVAsharegpt4v.github.ioexternal-linkmessage-square17fedilinkarrow-up11arrow-down10
arrow-up11arrow-down1external-linkShareGPT4V - New multi-modal model, improves on LLaVAsharegpt4v.github.ioCradawx@alien.topB to LocalLLaMA@poweruser.forumEnglish · 10 months agomessage-square17fedilink
minus-squarejustletmefuckinggo@alien.topBlinkfedilinkEnglisharrow-up1·10 months agoim new here. but is this true multimodality, or is it the llm communicating with a vision model? and what are those 4 models being benchmark tested here for exactly?
im new here. but is this true multimodality, or is it the llm communicating with a vision model?
and what are those 4 models being benchmark tested here for exactly?