Looking to reduce memory usage now that the model is memory resident (~1.9GB macOS)

Memory usage is still a bit large and takes up a bit too much performance on a normal computer, see how it can be tuned down. Is it possible to support remote invocation (e.g. allow us to deploy the model on our own server and get the text via callback). Or provide more lightweight models that are optional.

Please authenticate to join the conversation.

Upvoters
Status

Completed

Board
💡

Feature Request

Date

4 months ago

Author

Teng Fu

Subscribe to post

Get notified by email when there are changes.