Comments (8)
Did this occur on the default dataset we provided?
from openrlhf.
No, I use customized dataset. For SFT, it works fine.
from openrlhf.
No, I use customized dataset
Please refer to the
https://github.com/OpenRLHF/OpenRLHF/blob/main/openrlhf/datasets/reward_dataset.py
to organize the datasets.
from openrlhf.
Yes. I use similar template for the customized dataset and it works fine for SFT method.
from openrlhf.
It works well on my side (acc_mean > 0.6 )
Train step of epoch 0: 2%|_ | 229/12500 [03:19<2:52:35, 1.19it/s, preference_loss=0.763, chosen_reward=-1.03, reject_reward=-1.12, acc_mean=0.658, loss_mean=0.635]
please use a low learning rate: --learning_rate 5e-7 \
from openrlhf.
May not be relevant.
0.6931 is ln2, which means r(win) = r(lose) almost everywhere.
One probable reason is that the customized dataset is too hard, for example, y_w and y_l is too close and only differ a few tokens, or there are bugs in the dataset preparation, for example, y_w = y_l .
from openrlhf.
Hi, after I set the lr=1e-7, I can see some loss change. However, the dpo loss drops sharply and approach near 0. For this case, would it be possible that the positive and negative samples in the dataset are excessively dissimilar to each other?
from openrlhf.
There are a lot of possible reasons, such as bugs in processing. If there is a space token at the beginning of every y_l, and no space token at the beginning of every y_w. Then model may quickly converge without learning meaning signals.
from openrlhf.
Related Issues (20)
- Difference between `DeepSpeedEngine.save_checkpoint()` and `DeepSpeedStrategy.save_model()` HOT 2
- DPO后的模型推理出的结果都是无序符号 HOT 1
- Support training from breakpoint HOT 3
- llama3 70B DPO example script
- where is gradient_accumulation HOT 1
- Support RLOO HOT 1
- 现在Train_PPO_llama_ray 过程中会把Actor Model切分到不同卡上吗 HOT 4
- ConnectionRefusedError: [Errno 111] Connection refused HOT 5
- packing的问题 HOT 2
- "right" padding hardcoded HOT 3
- Error while saving the model under 4bit lora HOT 2
- multinode ppo training extremely slow HOT 15
- 使用ray的时候Request Entity Too Large HOT 3
- dpo 训练显存 OOM HOT 1
- Online DPO 支持 HOT 4
- Feature: add DPO-P
- Zero stage 3 error HOT 1
- Performance of Iterative DPO? HOT 1
- Why multiplying rstd instead of dividing by rstd? HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from openrlhf.