-
-
Notifications
You must be signed in to change notification settings - Fork 4.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
VAE or VQ-VAE is needed #2056
Labels
enhancement
New feature or request
Comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Is your feature request related to a problem? Please describe.
Currently, the timm library lacks implementations for Variational Autoencoder (VAE) and Vector Quantized VAE (VQ-VAE) models. Users looking to utilize these autoencoder architectures may find it inconvenient to implement them from scratch or integrate external implementations into their projects.
Describe the solution you'd like
I would like to request the addition of Variational Autoencoder (VAE) and Vector Quantized VAE (VQ-VAE) models to the timm library. This would involve creating modules for these autoencoder architectures, ensuring they adhere to the existing timm standards for simplicity and compatibility.
Describe alternatives you've considered
Users can currently implement VAE and VQ-VAE models from scratch or use external implementations from other libraries such as diffusener. However, having native support for these models in the timm library would provide a more streamlined and integrated experience for users.
The text was updated successfully, but these errors were encountered: