The thing I have is randomly sampling days from the year however this runs into flaws of the day just not being optimal. Other research papers in related work has two years one for training and another for testing. I only have one year and asked the professor for another year of data and he said it isn’t available with no explanation behind it. I tried searching online for the data since it’s supposedly public data but can’t find it. It’s kind of hard to get an accurate display of evaluation when you don’t have a good test environment. Instead of complaining I have to figure out something. This is in the works in trying to get published to a smaller journal. Not sure if any of you all had this so curious how you would handle such situations?

At least an idea

  • Sample months instead of just time samples and create a pseudo environment and run it through there. Picking diverse set that has similar yearly trends.

I don’t have much experience so open ears when you all had similar limitations and how you all overcame it.

  • Altruistic-Skill8667@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 months ago

    When I hear: not a lot of data, i immediately think: overfitting danger.

    If it’s a Reinforcement Learning algorithm, then maybe pretraining with synthetic data that’s similar to the real one, so that you already have some rough Q values. And then k-fold crossvalidation. Train on a subset and test on another, and then rotate through.

  • maieutic@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 months ago

    If your dataset is small/medium then use cross validation instead of a test set. Just make sure your folds are stratified by time to avoid data leakage.

  • linearmodality@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 months ago

    Other research papers in related work has two years one for training and another for testing.

    Have you considered emailing the authors of these papers and asking them for their datasets?

  • ahf95@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 months ago

    You will likely want your split whatever data you have. Probably just use a standard train:test split, with cross validation, as said in another comment. Also, you said in a comment that this was in an RL context, but if that were the case then you’d most likely be generating the next dataset after training on what you already have, so you’d know that you have more data on the way. So, are you solving a markov decision problem here, or is this just a applied form of supervised learning?

    • I_will_delete_myself@alien.topOPB
      link
      fedilink
      English
      arrow-up
      1
      ·
      10 months ago

      so, are you solving a markov decision problem here

      Yes. I am thinking of just using a metric to see if it made the optimal decision by the amount of value it delivers per capita.

      The main flaw from my previous metric is that it had a bias towards naive algorithms because the way its calculated which leads results to be misleading from reality. Skipping turns is sometimes the optimal decision which the metric said it was bad, but reality it isn’t.

      When I dug closer into the data it turns out the AI was destroying the naive algorithms with this metric and the total results we were aiming for.

      • ahf95@alien.topB
        link
        fedilink
        English
        arrow-up
        1
        ·
        10 months ago

        Dude, okay, are you actually doing research or is this troll? Or do you have a history of mental illness?