DeepMind released the Training Compute-Optimal Large Language Models paper in 2022 which describe some scaling laws for LLMs. As far as I understand this is the most accredited reference to estimate the optimal relation between dataset size, compute power and model size.

Recently a number of models have been developed using far less data, parameters and compute than the bigger LLMs. Yet these models achieved great results thanks to much better data quality. For instance models like WizardLM, TinyStories and phi-1. Similarly, a lot of research seems to imply that better data could offer huge improvements without any other changes.

I’m curious about what role the data quality plays in the training of LLMs.
Is the set of values estimated by the Chinchilla scaling laws optimal for these smaller models with optimized data too?
Do we have any model to estimate the quality of some datasets and some scaling laws that take it into account?
Are there any relevant projects or research I could check out, focused on creating big datasets to train larger LLMs with high-quality data?

  • thedabking123@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Measuring and improving quality of NLP datasets in a comprehensive way is probably the main migraine there.

    You can measure and improve quality by many dimensions that practitioners disagree on… ( accuracy, completeness, consistency, timeliness, validity, and uniqueness are common ways to slice data quality) and there’s no consistent single measure for some of those either.