Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Where do we get the result file for all evaluation #19

Open
DhavalRepo18 opened this issue Nov 29, 2024 · 4 comments
Open

Where do we get the result file for all evaluation #19

DhavalRepo18 opened this issue Nov 29, 2024 · 4 comments

Comments

@DhavalRepo18
Copy link

Assuming we do not want to run the benchmark, but want to run a new method and get the comparison done. What is the process?

@qhliu26
Copy link
Member

qhliu26 commented Nov 29, 2024

Hi! You can now access the evaluation results on the benchmark here https://github.com/TheDatumOrg/TSB-AD/tree/main/benchmark_exp/benchmark_eval_results. Let us know if any help is needed.

@DhavalRepo18
Copy link
Author

Thanks.

@DhavalRepo18
Copy link
Author

@qhliu26 thanks for uploading the (subset) of results. I have one question what if results are not reproducible? Do we need further investigation? or all care has been taken to ensure results are reproducible?

@qhliu26
Copy link
Member

qhliu26 commented Dec 18, 2024

Hi! Based on my experience, various factors such as hyperparameter settings, normalization techniques, random seeds, or even different Python versions can impact the performance and evaluation results to some extent. Some of these variations can be mitigated by running the algorithms multiple times and reporting the results with error bars. Additionally, comparing performance across the entire benchmark can provide a more reliable and robust evaluation. Specifically the evaluation results currently available on github are produced under numpy random seed 2024 and python 3.9. If there are specific algorithms for which you have observed significant deviations from our reported results, please let us know. We will be happy to investigate further.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants