Conversation
fails at quantization: missing attribute
TODO: copy attributes to the fused node
This saves a buttload of space
Note in the demo i am working on this drops the size of the model file from 25.8 KB to 403*4 B.
|
I think your approach is a good start except that I prefer it's implemented as an optimization pass.
However, I'll work on auto testing code generator first before I dig into this. BTW, I think |
dboyliao
left a comment
There was a problem hiding this comment.
As what I comment, I prefer it as an opt pass.
|
@mbartling what's the run time to go with this? I'm getting a runtime error with this repo: Checked out to this PR and generated the code |
|
@mbartling
works with current runtime dev branch. |
|
@dboyliao Do we already have these changes in the current dev branch? If so we can close this PR |
|
AFAIK, not yet. |
|
I still think this idea should be implemented for the reach, even if it means we close this PR. It will generate the absolute smallest models and is really a way to differentiate between debug/release build. Thoughts? |
|
Sure, I agree with you. |
|
Actually, on second thought it really isn't necessary for the rearch since param names are bound to the operators in the input out name enums. So this is more a debug convenience. I think we can put it off for a while |
|
Feel free to close :) |
Replaces all cstrings (except for those used by sdtensor) with integers.
Massive savings, much wow.
@dboyliao can you suggest a cleaner way to add this functionality to all ops? It's pretty messy at the moment, but I think we can inject a simple tensor_name transformer that decorates each op with a "generate sref func" that gets called on snippet render.