-
Notifications
You must be signed in to change notification settings - Fork 31
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support for guidance/structured output with prompt API #35
Comments
In general we're excited about exploring this. Minor API surface nitpicks:
So to summarize: |
Agree with @sushraja-msft that having structured output really helps with ensuring a correct response. For reference, we've implemented a parallel implementation of the Prompt API as an extension based on llama.cpp, available at GitHub. We've exposed a grammer object within the implementation that can be passed into the create function. An example use: sess = await window.aibrow.languageModel.create({
grammar:{
"type": "object",
"properties": {
"first_name": {
"type": "string"
},
"last_name": {
"type": "string"
},
"country": {
"type": "string"
}
}
}
})
const stream = await sess.promptStreaming("Extract data from the following text: 'John Doe is an innovative software developer with a passion for creating intuitive user experiences. Based in the heart of England, John has spent the past decade refining his craft, working with both startups and established tech companies. His deep commitment to quality and creativity is evident in the numerous award-winning apps he has developed, which continue to enrich the digital lives of users worldwide. Beyond his technical skills, John is admired for his collaborative spirit and mentorship, always eager to share his knowledge and inspire the next generation of tech enthusiasts.'");
for await (const chunk of stream) {
console.log(chunk)
} Having experienced quite a few inconsistencies before when trying to "plead with the prompt" to get it to only output JSON (where it often tries to wrap it in markdown), a constraining structured output seems like the best approach. |
To aid programmability, reduce compatibility risk from the API returning different results across browser, avoid challenges in updating a shipping model in the browser (Google Model V1 to Google Model V2), please consider adding techniques like guidance, structured outputs as an integral part of the prompt API.
Problem Illustration
Consider the following web developer scenarios, where a developer is:
Web developers who attempt to parse the response are going to have a hard time writing code that is model/browser agnostic.
Constraining Output
One way to solve this problem is to use guidance or techniques like it. At a high level these techniques work by restricting the next allowed token from the LLM to conform to a grammar. Guidance works on top of a model, is model agnostic and only changes logits from the last layer of a model before sampling. There is an additional implementation detail within guidance in that information about all possible tokens prefixed with the next possible token is required for it to function (explanation).
With guidance (demo) we get better consistency across models and responses that are immediately parseable with JavaScript.
Proposal
The proposal is to add responseJsonSchema to the AIAssistantPromptOptions.
dictionary AIAssistantPromptOptions { AbortSignal signal; DomString? responseJsonSchema; };
JSON schema is familiar to web developers. However, JSON schema is a super set of what techniques like guidance can achieve today. For example, parts of the schema to enforce JSON schema constraints like dependent required cannot be enforced.
Either the API can state that only Property Name, Value Type, Enum, Arrays would be enforced, or Prompt API should validate the response with a JSON schema validator and indicate that the response is non conformant. Slight preference to the first option because of its simplicity.
Other Approaches
The text was updated successfully, but these errors were encountered: