• 0 Posts
  • 14 Comments
Joined 1 year ago
cake
Cake day: October 27th, 2023

help-circle




  • Don’t be sad, it’s just a part of how things are you just have to choose a method and stick to it.

    I personally use Notion. I’ve created a database and added properties like date, venue, authors, organizations, etc.

    For example, the other day I needed to recap what the BLIP paper was about so I just searched the paper in the database and took a look at the page. On that page I’ve highlighted different text with different colors depending on when I came back to read it.

    Took me a while to get this working and into the habit of it though.




  • TL;DR The more constraints on the model, the more time should spend analyzing your data and formulating your problem.

    I’ll agree with the top comment. I’ve also had to deal with a problem at work where we were trying to perform product name classification for our e-commerce product. The problem was that we couldn’t afford to have anything too large or increase infrastructure costs (i.e., if possible we didn’t want to use any more GPU computing resources than we already were).

    It turns out that extensive EDA was what saved us. We were able to come up with a string-matching algorithm sophisticated enough that it achieved high precision with practically no latency concerns. Might not be as flexible as something like BERT but it got the job done.


  • TL;DR The more constraints on the model, the more time should spend analyzing your data and formulating your problem.

    I’ll agree with the top comment. I’ve also had to deal with a problem at work where we were trying to perform product name classification for our e-commerce product. The problem was that we couldn’t afford to have anything too large or increase infrastructure costs (i.e., if possible we didn’t want to use any more GPU computing resources than we already were).

    It turns out that extensive EDA was what saved us. We were able to come up with a string-matching algorithm sophisticated enough that it achieved high precision with practically no latency concerns. Might not be as flexible as something like BERT but it got the job done.





  • Is there any reason why you won’t just use a CLIP-based model and why you’re trying to use OpenAI’s GPT?

    I’m also in charge of a text-image (text-image, not multimodal in my case) model that my company’s trying to create a search product with. There have been talks about using “ChatGPT” from higher-ups but I just don’t see the reason why we’d have to do this. I figured that a simple NER model or something would work just as well, I mean how many people do online shopping while expecting textual responses from the website.