ai_sdk_openai 1.0.0+1
ai_sdk_openai: ^1.0.0+1 copied to clipboard
OpenAI provider for the AI SDK Dart port.
ai_sdk_openai examples #
Installation #
dart pub add ai_sdk_dart ai_sdk_openai
export OPENAI_API_KEY=sk-...
Text generation #
import 'package:ai_sdk_dart/ai_sdk_dart.dart';
import 'package:ai_sdk_openai/ai_sdk_openai.dart';
final result = await generateText(
model: openai('gpt-4.1-mini'),
prompt: 'Say hello from AI SDK Dart!',
);
print(result.text);
print('tokens used: ${result.usage?.totalTokens}');
Streaming #
import 'dart:io';
final result = await streamText(
model: openai('gpt-4.1-mini'),
prompt: 'Count from 1 to 5.',
);
await for (final chunk in result.textStream) {
stdout.write(chunk);
}
Structured output #
final result = await generateText<Map<String, dynamic>>(
model: openai('gpt-4.1-mini'),
prompt: 'Return the capital and currency of Japan as JSON.',
output: Output.object(
schema: Schema<Map<String, dynamic>>(
jsonSchema: const {
'type': 'object',
'properties': {
'capital': {'type': 'string'},
'currency': {'type': 'string'},
},
},
fromJson: (json) => json,
),
),
);
print(result.output); // {capital: Tokyo, currency: JPY}
Tools (multi-step agent) #
final result = await generateText(
model: openai('gpt-4.1-mini'),
prompt: 'What is the weather in Paris?',
maxSteps: 5,
tools: {
'getWeather': tool<Map<String, dynamic>, String>(
description: 'Get current weather for a city.',
inputSchema: Schema<Map<String, dynamic>>(
jsonSchema: const {
'type': 'object',
'properties': {'city': {'type': 'string'}},
'required': ['city'],
},
fromJson: (json) => json,
),
execute: (input, _) async => 'Sunny, 18°C in ${input['city']}',
),
},
);
print(result.text);
Embeddings #
final a = await embed(
model: openai.embedding('text-embedding-3-small'),
value: 'Hello, world!',
);
final b = await embed(
model: openai.embedding('text-embedding-3-small'),
value: 'Hi there!',
);
print('similarity: ${cosineSimilarity(a.embedding, b.embedding)}');
Image generation (DALL-E 3) #
final result = await generateImage(
model: openai.image('dall-e-3'),
prompt: 'A futuristic city skyline at sunset, digital art.',
);
print(result.images.first.url);
Speech synthesis (TTS) #
final result = await generateSpeech(
model: openai.speech('tts-1'),
text: 'Hello from AI SDK Dart!',
);
// result.audio is a Uint8List of MP3 audio bytes
await File('output.mp3').writeAsBytes(result.audio);
Transcription (Whisper) #
import 'dart:io';
import 'dart:typed_data';
final audioBytes = await File('recording.mp3').readAsBytes();
final result = await transcribe(
model: openai.transcription('whisper-1'),
audio: audioBytes,
mimeType: 'audio/mpeg',
);
print(result.text);
Custom base URL (Azure / compatible endpoint) #
final azureModel = OpenAIProvider(
apiKey: 'your-azure-key',
baseUrl: 'https://my-resource.openai.azure.com/openai/deployments/gpt-4/v1',
).call('gpt-4');
final result = await generateText(
model: azureModel,
prompt: 'Hello from Azure OpenAI!',
);
Runnable example apps #
examples/basic— Dart CLI covering all core APIsexamples/advanced_app— Flutter app with OpenAI image gen, TTS, STT, embeddings