You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I think it would be best for this limit to be non-existent by default, rather than 4. Or at least something higher like 16. Though client applications are able to detect that TGI is being used and encode the limit, it causes poorer behavior in autocomplete scenarios where >4 stop words are actually necessary. Yes, users of TGI can technically change this value, but many do not know to do this and will have lower-quality first experiences with whatever tools they are using.
All this said, I appreciate that this limit was initially defined by OpenAI and following along with their patterns is something that I in general support.
System Info
text-generation-inference/launcher/src/main.rs
Lines 427 to 428 in 1028996
I think it would be best for this limit to be non-existent by default, rather than 4. Or at least something higher like 16. Though client applications are able to detect that TGI is being used and encode the limit, it causes poorer behavior in autocomplete scenarios where >4 stop words are actually necessary. Yes, users of TGI can technically change this value, but many do not know to do this and will have lower-quality first experiences with whatever tools they are using.
All this said, I appreciate that this limit was initially defined by OpenAI and following along with their patterns is something that I in general support.
For additional background: continuedev/continue#2380
Information
Tasks
Reproduction
Send a request with >4 stop tokens
Expected behavior
Ideally a request with >4 stop tokens does not throw an error, unless the person running TGI has explicitly set a limit
The text was updated successfully, but these errors were encountered: