Skip to content

The accuracy of using the batch processing function of the client is low #1354

Closed Answered by merrymercy
guleng asked this question in Q&A
Discussion options

You must be logged in to vote
  1. Remove stop="\n" in text_qa to fix the early-stop issue.
  2. For this model, it is better to use chat template (e..g, sgl.user, sgl.assistant)
@sgl.function
def text_qa(s, question):
    s += sgl.user(question)
    s += sgl.assistant(sgl.gen("answer", max_tokens=600, temperature=0))
    return s

Replies: 1 comment 2 replies

Comment options

You must be logged in to vote
2 replies
@guleng
Comment options

@merrymercy
Comment options

Answer selected by merrymercy
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants