Flieber, Syrup Tech, and SupChains Launch an AI-Driven Supply Chain Forecasting Competition
Hi, so yesterday I was trying to make submissions for the phase 1. Unfortunately, I kept receiving errors in the submission page flagging nulls, blanks etc. I checked my datasets and these weren't present so I emailed the team but unfortunately, no response.
Since the phase 2 datasets were released, I was able to test using those datasets and it seems that have the datasets ordered so the index would match appeared to be causing the issue.
I don't expect to have any late submissions but thought I would make everyone aware in case of similar issues with phase 2.
Hi, so yesterday I was trying to make submissions for the phase 1. Unfortunately, I kept receiving errors in the submission page flagging nulls, blanks etc. I checked my datasets and these weren't present so I emailed the team but unfortunately, no response.
Since the phase 2 datasets were released, I was able to test using those datasets and it seems that have the datasets ordered so the index would match appeared to be causing the issue.
I don't expect to have any late submissions but thought I would make everyone aware in case of similar issues with phase 2.
I have not submitted any result yet. Reason: late joining and time management.
Kindly tell me that I am eligible for phase 2 competition and if able to score good on leaderboard, then I will be eligible for price money or not.
I wanted to share a thought regarding the competition's evaluation metric and get your inputs.
Since the competition's metric resembles MAE and 1) we are dealing with sparse, zero-inflated data, and 2) we are not evaluating models at a higher level of the hierarchy, could we end up selecting models (as the best model) that systematically under-forecast for non-zero, high-volume items?
MAE tends to select models that best estimate the median of the data, and in zero-inflated datasets, the median is often zero (or close to it). While we do have a bias term in the evaluation metric to help balance things, I’m curious if it’s strong enough to prevent this issue.
Typically, the dollar value of those minor, high-volume items is as significant, if not greater, than the majority of low-volume items, making under-forecasting them particularly costly.
While the leaderboard does show some updated results, nothing I've uploaded works. That includes the submission example provided here… which really should work.
I've also tested my submissions and they meet all the specified formatting requirements as per the attached test code. So what's up?
Hi all,
I'm observing a large gap between my internal time-based cross validation, and the score shown on leaderboard.
I was expecting to find some bug in my code leading to leakage but I couldn't find anything relevant.
Are you also experiencing the same?
Keep up to date by participating in our global community of data scientists and AI enthusiasts. We discuss the latest developments in data science competitions, new techniques for solving complex challenges, AI and machine learning models, and much more!