Skip to content

Commit e86be6f

Browse files
authored
chore: remove logprobs (#5896)
1 parent c855663 commit e86be6f

40 files changed

+182
-1108
lines changed

‎.changeset/itchy-deers-jog.md

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,5 @@
1+
---
2+
'@ai-sdk/provider': major
3+
---
4+
5+
chore: remove logprobs

‎examples/ai-core/src/generate-text/openai-logprobs.ts

Lines changed: 0 additions & 19 deletions
This file was deleted.

‎examples/ai-core/src/stream-object/openai-fullstream.ts

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,6 @@ async function main() {
3636

3737
case 'finish': {
3838
console.log('Finish reason:', part.finishReason);
39-
console.log('Logprobs:', part.logprobs);
4039
console.log('Usage:', part.usage);
4140
break;
4241
}

‎examples/ai-core/src/stream-text/azure-fullstream-logprobs.ts

Lines changed: 0 additions & 31 deletions
This file was deleted.

‎examples/ai-core/src/stream-text/openai-fullstream-logprobs.ts

Lines changed: 0 additions & 33 deletions
This file was deleted.

‎packages/ai/core/generate-object/__snapshots__/stream-object.test.ts.snap

Lines changed: 0 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -92,13 +92,6 @@ exports[`streamObject > output = "object" > result.fullStream > should send full
9292
},
9393
{
9494
"finishReason": "stop",
95-
"logprobs": [
96-
{
97-
"logprob": 1,
98-
"token": "-",
99-
"topLogprobs": [],
100-
},
101-
],
10295
"response": {
10396
"id": "id-0",
10497
"modelId": "mock-model-id",

‎packages/ai/core/generate-object/generate-object-result.ts

Lines changed: 0 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,6 @@ import {
33
FinishReason,
44
LanguageModelRequestMetadata,
55
LanguageModelResponseMetadata,
6-
LogProbs,
76
ProviderMetadata,
87
} from '../types';
98
import { LanguageModelUsage } from '../types/usage';
@@ -47,14 +46,6 @@ Response body (available only for providers that use HTTP requests).
4746
body?: unknown;
4847
};
4948

50-
/**
51-
Logprobs for the completion.
52-
`undefined` if the mode does not support logprobs or if was not enabled.
53-
54-
@deprecated Will become a provider extension in the future.
55-
*/
56-
readonly logprobs: LogProbs | undefined;
57-
5849
/**
5950
Additional provider-specific metadata. They are passed through
6051
from the provider to the AI SDK and enable provider-specific

‎packages/ai/core/generate-object/generate-object.ts

Lines changed: 0 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,6 @@ import {
2323
CallWarning,
2424
FinishReason,
2525
LanguageModel,
26-
LogProbs,
2726
ProviderMetadata,
2827
} from '../types';
2928
import { LanguageModelRequestMetadata } from '../types/language-model-request-metadata';
@@ -264,7 +263,6 @@ export async function generateObject<SCHEMA, RESULT>({
264263
let warnings: CallWarning[] | undefined;
265264
let response: LanguageModelResponseMetadata;
266265
let request: LanguageModelRequestMetadata;
267-
let logprobs: LogProbs | undefined;
268266
let resultProviderMetadata: ProviderMetadata | undefined;
269267

270268
const standardizedPrompt = standardizePrompt({
@@ -377,7 +375,6 @@ export async function generateObject<SCHEMA, RESULT>({
377375
finishReason = generateResult.finishReason;
378376
usage = generateResult.usage;
379377
warnings = generateResult.warnings;
380-
logprobs = generateResult.logprobs;
381378
resultProviderMetadata = generateResult.providerMetadata;
382379
request = generateResult.request ?? {};
383380
response = generateResult.responseData;
@@ -468,7 +465,6 @@ export async function generateObject<SCHEMA, RESULT>({
468465
warnings,
469466
request,
470467
response,
471-
logprobs,
472468
providerMetadata: resultProviderMetadata,
473469
});
474470
},
@@ -480,7 +476,6 @@ class DefaultGenerateObjectResult<T> implements GenerateObjectResult<T> {
480476
readonly finishReason: GenerateObjectResult<T>['finishReason'];
481477
readonly usage: GenerateObjectResult<T>['usage'];
482478
readonly warnings: GenerateObjectResult<T>['warnings'];
483-
readonly logprobs: GenerateObjectResult<T>['logprobs'];
484479
readonly providerMetadata: GenerateObjectResult<T>['providerMetadata'];
485480
readonly response: GenerateObjectResult<T>['response'];
486481
readonly request: GenerateObjectResult<T>['request'];
@@ -490,7 +485,6 @@ class DefaultGenerateObjectResult<T> implements GenerateObjectResult<T> {
490485
finishReason: GenerateObjectResult<T>['finishReason'];
491486
usage: GenerateObjectResult<T>['usage'];
492487
warnings: GenerateObjectResult<T>['warnings'];
493-
logprobs: GenerateObjectResult<T>['logprobs'];
494488
providerMetadata: GenerateObjectResult<T>['providerMetadata'];
495489
response: GenerateObjectResult<T>['response'];
496490
request: GenerateObjectResult<T>['request'];
@@ -502,7 +496,6 @@ class DefaultGenerateObjectResult<T> implements GenerateObjectResult<T> {
502496
this.providerMetadata = options.providerMetadata;
503497
this.response = options.response;
504498
this.request = options.request;
505-
this.logprobs = options.logprobs;
506499
}
507500

508501
toJsonResponse(init?: ResponseInit): Response {

‎packages/ai/core/generate-object/stream-object-result.ts

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,6 @@ import {
44
FinishReason,
55
LanguageModelRequestMetadata,
66
LanguageModelResponseMetadata,
7-
LogProbs,
87
ProviderMetadata,
98
} from '../types';
109
import { LanguageModelUsage } from '../types/usage';
@@ -108,7 +107,6 @@ export type ObjectStreamPart<PARTIAL> =
108107
| {
109108
type: 'finish';
110109
finishReason: FinishReason;
111-
logprobs?: LogProbs;
112110
usage: LanguageModelUsage;
113111
response: LanguageModelResponseMetadata;
114112
providerMetadata?: ProviderMetadata;

‎packages/ai/core/generate-object/stream-object.test.ts

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -223,7 +223,6 @@ describe('streamObject', () => {
223223
type: 'finish',
224224
finishReason: 'stop',
225225
usage: { inputTokens: 2, outputTokens: 10 },
226-
logprobs: [{ token: '-', logprob: 1, topLogprobs: [] }],
227226
},
228227
]),
229228
}),
@@ -440,7 +439,6 @@ describe('streamObject', () => {
440439
{
441440
type: 'finish',
442441
finishReason: 'stop',
443-
logprobs: undefined,
444442
usage: { inputTokens: 3, outputTokens: 10 },
445443
},
446444
]),
@@ -482,7 +480,6 @@ describe('streamObject', () => {
482480
{
483481
type: 'finish',
484482
finishReason: 'stop',
485-
logprobs: undefined,
486483
usage: { inputTokens: 3, outputTokens: 10 },
487484
},
488485
]),

0 commit comments

Comments
 (0)