mirror of
https://github.com/x1xhlol/system-prompts-and-models-of-ai-tools.git
synced 2026-04-11 14:21:56 +08:00
Merge 7e277ac278 into db9f7d8b9f
This commit is contained in:
commit
a2b31979a1
@ -568,7 +568,7 @@ const response = await ai.models.generateContent({
|
||||
console.log(response.text);
|
||||
```
|
||||
|
||||
**Incorrect Example for Setting `maxOutputTokens` without `thinkingBudget`**
|
||||
**Example for Setting `maxOutputTokens` with low `thinkingBudget`**
|
||||
```ts
|
||||
import { GoogleGenAI } from "@google/genai";
|
||||
|
||||
@ -577,9 +577,9 @@ const response = await ai.models.generateContent({
|
||||
model: "gemini-2.5-flash",
|
||||
contents: "Tell me a story.",
|
||||
config: {
|
||||
// Problem: The response will be empty since all the tokens are consumed by thinking.
|
||||
// Fix: Add `thinkingConfig: { thinkingBudget: 25 }` to limit thinking usage.
|
||||
// We limit thinking usage to 25 tokens so that we have 25 tokens left for the response (50 - 25).
|
||||
maxOutputTokens: 50,
|
||||
thinkingConfig: { thinkingBudget: 25 },
|
||||
},
|
||||
});
|
||||
console.log(response.text);
|
||||
|
||||
Loading…
Reference in New Issue
Block a user