Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature Request] Custom "chat" HF datasets #1088

Open
chimezie opened this issue Nov 3, 2024 · 0 comments
Open

[Feature Request] Custom "chat" HF datasets #1088

chimezie opened this issue Nov 3, 2024 · 0 comments

Comments

@chimezie
Copy link
Contributor

chimezie commented Nov 3, 2024

The LoRa tuners's local datasets support the following data format:

{"messages": [{"role": "system", "content": "You are a helpful assistant."}, 
              {"role": "user", "content": "Hello."}, 
              {"role": "assistant", "content": "How can I assistant you today."}]}

Some HF datasets, such as the UltraFeedback dataset, used for Direct Preference Optimization (see: HF DPO trainer and #513) use a (json) data format such as the following:

[ { "content": "...", "role": "user" }, { "content": "...", "role": "assistant" } ]

To incorporate the use of such HF datasets, it would be helpful to to generalize the use of prompt_feature, text_feature, and completion_feature to include chat_feature, which indicates the HF dataset feature to use for the chat template structure.

chimezie added a commit to chimezie/mlx-examples that referenced this issue Nov 4, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant