We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The current codec model huge, about 14GB, this even not count the LLM size.
Any plan to release a smaller size codec?
The text was updated successfully, but these errors were encountered:
The codec model itself is actually not that large; it only uses about 3GB, with parameter size being less than 1B, https://huggingface.co/HKUST-Audio/xcodec2/blob/main/pytorch_model.bin For inference, you only need this. The larger size is because I also provided checkpoint includes all the state, making it suitable for continued training as well. https://huggingface.co/HKUST-Audio/xcodec2/blob/main/ckpt/epoch%3D4-step%3D1400000.ckpt Hope this clears things up!
Sorry, something went wrong.
thank u, is that possible make 3G even smaller but still maintain the ability?
No branches or pull requests
The current codec model huge, about 14GB, this even not count the LLM size.
Any plan to release a smaller size codec?
The text was updated successfully, but these errors were encountered: