Comments (5)
Thanks for your interest in our work!
The direct answer for your Q1 is YES. We found that the best way to train a pre-experienced model is to consider diversity. Thus we try to gain the embeddings for all the data and select by diversity.
However:
1 If your base model is already really powerful, you can try to neglect the pre-experienced model and directly run the cherry_analysis on the base model.
2 You can also randomly choose some data for the training of the pre-experienced model. Though not as good as considering diversity, it still works.
3 You can also use other quick methods to consider the diversity. For example, sentence_bert + K means.
For the second question, I don't know what base model and what SFT data you use, so I can not give a definite answer. But I think in most situations, you don't need to modify it.
from cherry_llm.
Thank you for your reply!
Because I saw the previous text saying "Learning from Brief Experience" by selecting a small amount of data, I'm not sure it's right to put all the data into it for training.
In addition, full data takes a long time to train.
I'll try it. Thank you.
from cherry_llm.
Ah, I am not sure if there is still a misunderstanding.
For the pre-experienced model, it indeed only needs a small amount of data. The code "pre_experience_analysis.sh" you were asking for is not "put all the data into it for training", it just tries to select a suitable small amount of the data for training the pre-experienced model.
from cherry_llm.
Thank you for your reply.
Maybe I'm not asking the question accurately.
The "pre_experience_analysis.sh" script does not perform training. It embeds all SFT data (that is, "get_perplexity_and_embedding_whole_text") and then uses the "pre_expeerience_selection.sh" script to perform clustering.
Is my understanding correct?
Thank you again
from cherry_llm.
Yes, I think you are correct~
from cherry_llm.
Related Issues (20)
- a confusion about Instruction-Following Difficulty (IFD) scores HOT 2
- a confusion about data_by_IFD HOT 3
- Logic behind IFD score HOT 1
- I plan to apply this method on Llama2, which part of this project needs to be changed to adapt to Llama2? HOT 1
- May I ask if this project is suitable for other large models, such as the Baichuan model, to filter high-quality datasets from other fields HOT 4
- about the paper HOT 1
- Multi-round conversation data set HOT 3
- GPT-4/ChatGPT Evaluation Code HOT 1
- How to filter code SFT data? HOT 2
- Could the Pre-Experienced Model be used in other different dataset? HOT 1
- Any report of time consuming? HOT 1
- Chinese SFT data cannot be displayed. HOT 3
- 'The training of pre-experienced models is discarded for more efficient usage': that means we can only use base model to do cherry analysis and selection? HOT 1
- batch? HOT 1
- 关于Direct Answer Score sθ(A) HOT 2
- Evaluation reproducibility on benchmarks HOT 4
- how many epochs to train on cherry data? HOT 2
- Question about the effect of labels[0, :start_token] = -100 HOT 1
- why is the process so slow HOT 5
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from cherry_llm.