Comments (4)
The code and data in this repo can reproduce the reported results.
We filtered out the samples that are too long to fit in the context length of our model. Therefore, the number of samples in the repo are smaller than that in the paper (12.4k v.s. 15k, 242 v.s. 252). For validation samples in dolly, we indeed use 1K samples for validation in our experiments.
We will fix and explain these information in our paper. Thanks for pointing it out!
from lmops.
The code and data in this repo can reproduce the reported results. We filtered out the samples that are too long to fit in the context length of our model. Therefore, the number of samples in the repo are smaller than that in the paper (12.4k v.s. 15k, 242 v.s. 252). For validation samples in dolly, we indeed use 1K samples for validation in our experiments. We will fix and explain these information in our paper. Thanks for pointing it out!
Thanks very much for your quick reply! So for dolly, you use 11.4k data for training, 1k data for validation (selecting the best ckpt) and 500 data for testing (reproducing the results in paper)?
from lmops.
yes
from lmops.
yes
Thanks!
from lmops.
Related Issues (20)
- Paper:ADAPTING LARGE LANGUAGE MODELS VIA READING COMPREHENSION HOT 3
- prompt_optimization HOT 2
- iS LLMA lossless? HOT 1
- top-p < 1 fails inf assertion HOT 1
- why is the mpu/cross_entropy missing a softmax_logits_t HOT 2
- [MiniLLM] sft of llama2-7b out of memory on V100 HOT 2
- [MiniLLM]LLama sft on Dolly hard to reproduce results in paper. HOT 2
- Questions about the free-law data used in the paper "Adapt LLM to domains" HOT 2
- [MiniLLM]Why dolly only has 12435 training samples? HOT 2
- [MiniLLM] About the gradient accumulation in finetune.py HOT 2
- [tuna] Libraries are conflicting and/or very aged HOT 5
- Missing Jailbreak dataset from protegi? HOT 2
- ImportError: cannot import name 'mpu' from 'transformers' HOT 4
- ModuleNotFoundError: No module named 'deepspeed'
- [llm_retriever] Questions about the dataset
- The update method in the UCB algorithm is inconsistent with the paper and code HOT 1
- RoBERTa Corpus HOT 1
- Questions about task datasets used in the paper "Adapt LLM to domains" HOT 2
- 【MiniLLM】is it normal to get negative loss at some step? HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from lmops.