Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

k2SSL: A Faster and Better Framework for Self-Supervised Speech Representation Learning #1745

Draft
wants to merge 34 commits into
base: master
Choose a base branch
from

Conversation

yfyeung
Copy link
Collaborator

@yfyeung yfyeung commented Sep 7, 2024

  • Libri-Light data processing script
  • Libri-Light Zipformer multi-node-multi-gpu pre-train recipe
  • LibriSpeech Zipformer bpe-level prund rnn-t fine-tune recipe
  • LibriSpeech Zipformer letter-level ctc fine-tune recipe
  • Release all resource and results for Zipformer Base
  • Release all resource and results for Zipformer Large

@yfyeung
Copy link
Collaborator Author

yfyeung commented Sep 25, 2024

Part of the resources have been released:

Zipformer Base pre-trained with cross entropy loss: Checkpoints, logs, and scripts
Zipformer Base fine-tuned with pruned RNN-T loss: Checkpoints, logs, and scripts
Zipformer Base fine-tuned with letter-level CTC loss: Checkpoints, logs, and scripts
Zipformer Base pre-trained manifests including kmeans labels: Dataset

With these resources, I believe anyone with 8 V100 32G GPUs can easily reproduce our experiments.

@ezerhouni
Copy link
Collaborator

@yfyeung Thank you very much for the PR and sharing the model weights. Do you plan to release a paper also ?

@yfyeung
Copy link
Collaborator Author

yfyeung commented Sep 25, 2024

@yfyeung Thank you very much for the PR and sharing the model weights. Do you plan to release a paper also ?

Maybe, but currently I am not sure whether it is suitable for a technical report or just a normal research paper.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants