Skip to content

Local LLM performance issue #59

Discussion options

You must be logged in to vote

We want to point out that we now use a way more efficient LLM inference backend in SpeziLLM that builds on top of MLX, it should resolve most of the performance constraints we have seen before.

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@PSchmiedmayer
Comment options

Answer selected by PSchmiedmayer
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
2 participants