Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reasoning behind Alapca's default split #364

Open
macsz opened this issue Feb 1, 2024 · 1 comment
Open

Reasoning behind Alapca's default split #364

macsz opened this issue Feb 1, 2024 · 1 comment
Labels

Comments

@macsz
Copy link

macsz commented Feb 1, 2024

For the Alpaca dataset, the default split comprises 51,800 samples for training and 200 samples for testing [1]. What is the rationale behind such a small test set? I haven't been able to find any recommended split ratios for the Alpaca dataset.

Is the purpose of the small test set merely to serve as a reference point, suggesting that for more reliable testing, another dataset or framework, such as HELM, should be utilized?

[1] https://github.com/facebookresearch/llama-recipes/blob/main/src/llama_recipes/datasets/alpaca_dataset.py#L30

@HamidShojanazeri
Copy link
Contributor

@macsz I believe it was more of quick test rather than a recommended setting, agree with you it should be higher, will consider adding a fix.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants