Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Model parameter file size #2

Closed
thias42 opened this issue May 11, 2020 · 3 comments
Closed

Model parameter file size #2

thias42 opened this issue May 11, 2020 · 3 comments

Comments

@thias42
Copy link

thias42 commented May 11, 2020

Hi ksangeeta2429,

Thank you for making the code and models available. I am experimenting with edgel3 and noticed that the model parameter files are the same size (~35M) as the openl3 models. Shouldn't they be much smaller?

Thank you,
Matthias

@ksangeeta2429
Copy link
Owner

Hello Matthias! If you see the model size, you will see that the edgel3 audio model would have the same size as the openl3 audio model because the pruned weights i.e. weights=0 are saved as float32 values. You can use a CSR/any other sparse representation to save the non-zero weights to see the gain in static memory.

Note: edgel3 is mainly about reducing the static memory of Audio subnetwork. You can check the EdgeL3 paper and use similar techniques to reduce the static memory of the video subnetwork. (I know I should have named it EdgeL3 Audio!)

@thias42
Copy link
Author

thias42 commented May 13, 2020

Hello ksangeeta2429, I am interested in finding a more efficient / less expensive way of calculating the openl3 embeddings for music. Is your compressed network for audio env or music? How would I go about to use your model and benefit from the compressed network? I am currently using tensorflow-serving for inference. How is your work related to TensorFlow Lite?

Thank you for your help.

Matthias

@ksangeeta2429
Copy link
Owner

The edgel3 models (uploaded ones) are based on music. It is easy to convert the models to tflite models but I do not think you will benefit (yet) from using sparse models in tflite. AFAIK embedding out of convolution layer 7 (conv7) had a comparable performance as compared to the ones from conv8 (~50% of the weights come from this layer). So, you could just get rid of conv8 and quantize it to int8 with tflite to get a much smaller model.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants