Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Remove max_stop_sequences by default #2584

Open
3 of 4 tasks
sestinj opened this issue Sep 29, 2024 · 0 comments
Open
3 of 4 tasks

Remove max_stop_sequences by default #2584

sestinj opened this issue Sep 29, 2024 · 0 comments

Comments

@sestinj
Copy link

sestinj commented Sep 29, 2024

System Info

#[clap(default_value = "4", long, env)]
max_stop_sequences: usize,

I think it would be best for this limit to be non-existent by default, rather than 4. Or at least something higher like 16. Though client applications are able to detect that TGI is being used and encode the limit, it causes poorer behavior in autocomplete scenarios where >4 stop words are actually necessary. Yes, users of TGI can technically change this value, but many do not know to do this and will have lower-quality first experiences with whatever tools they are using.

All this said, I appreciate that this limit was initially defined by OpenAI and following along with their patterns is something that I in general support.

For additional background: continuedev/continue#2380

Information

  • Docker
  • The CLI directly

Tasks

  • An officially supported command
  • My own modifications

Reproduction

Send a request with >4 stop tokens

Expected behavior

Ideally a request with >4 stop tokens does not throw an error, unless the person running TGI has explicitly set a limit

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant