Skip to content

Reducing model size by converting to ORT format #416

Answered by robertknight
ken107 asked this question in Q&A
Discussion options

You must be logged in to vote

Issue 1 is the topic of this thread. I'm wondering if Piper models are already optimized, or if their sizes can be further reduced. As for performing inferencing at the edge, we'd like the download size to be as small as possible

The ORT model format is mostly not about model file size but about being more efficient to load and requiring a smaller binary file size, because it is based on FlatBuffers, a format Google designed to enable efficient loading of resources in mobile games, rather than Protocol Buffers.

The optimizations mentioned in the doc you linked to are mainly about improving execution performance, usually by combining ("fusing") multiple steps of the model into a single s…

Replies: 2 comments

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Answer selected by ken107
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants