googleai_dart 0.1.0+3 copy "googleai_dart: ^0.1.0+3" to clipboard
googleai_dart: ^0.1.0+3 copied to clipboard

Dart Client for the Google AI API (Gemini Pro, Gemini Pro Vision, embeddings, etc.).

Google AI Dart Client #

tests googleai_dart MIT

Unofficial Dart client for Google AI for Developers (Gemini API v1).

Note: The official google_generative_ai now has feature parity with this package (except for the Model info endpoints). We plan to deprecate this package in the near future.

Features #

  • Generated from the official Google AI OpenAPI specification
  • Fully type-safe, documented and tested
  • All platforms supported
  • Custom base URL, headers and query params support (e.g. HTTP proxies)
  • Custom HTTP client support (e.g. SOCKS5 proxies or advanced use cases)

Supported endpoints:

  • Generate content (with streaming and tuned model support)
  • Embed content (with batch support)
  • Count tokens
  • Models info
  • Operations

Table of contents #

Usage #

Refer to the documentation for more information about the API.

Authentication #

The Google AI API uses API keys for authentication. Visit Google AI Studio dashboard page to retrieve the API key you'll use in your requests.

Remember that your API key is a secret!
Do not share it with others or expose it in any client-side code (browsers, apps). Production requests must be routed through your own backend server where your API key can be securely loaded from an environment variable or key management service.

final googleaiApiKey = Platform.environment['GOOGLEAI_API_KEY'];
final client = GoogleAIClient(apiKey: googleaiApiKey);

Generate content #

Text-only input

Use the generateContent method to generate a response from the model given an input message.

final res = await client.generateContent(
  modelId: 'gemini-pro',
  request: const GenerateContentRequest(
    contents: [
      Content(
        parts: [
          Part(text: 'Write a story about a magic backpack.'),
        ],
      ),
    ],
    generationConfig: GenerationConfig(
      temperature: 0.8,
    ),
  ),
);
print(res.candidates?.first.content?.parts?.first.text);
// In a quaint little town nestled amidst rolling hills, there lived a...

Text-and-image input

If the input contains both text and image, use the gemini-pro-vision model.

final res = await client.generateContent(
  modelId: 'gemini-pro-vision',
  request: GenerateContentRequest(
    contents: [
      Content(
        parts: [
          const Part(text: 'What is this picture?'),
          Part(
            inlineData: Blob(
              mimeType: 'image/png',
              data: base64.encode(
                await File('./test/assets/1.png').readAsBytes(),
              ),
            ),
          ),
        ],
      ),
    ],
  ),
);
print(res.candidates?.first.content?.parts?.first.text);
// The picture shows some scones with blueberries and flowers...

Multi-turn conversations (chat)

Using Gemini, you can build freeform conversations across multiple turns.

The gemini-pro-vision model (for text-and-image input) is not yet optimized for multi-turn conversations. Make sure to use gemini-pro and text-only input for chat use cases.

final res = await client.generateContent(
  modelId: 'gemini-pro',
  request: const GenerateContentRequest(
    contents: [
      Content(
        role: 'user',
        parts: [
          Part(
            text: 'Write the first line of a story about a magic backpack.',
          ),
        ],
      ),
      Content(
        role: 'model',
        parts: [
          Part(
            text:
                'In the bustling city of Meadow brook, lived a young girl named Sophie. '
                'She was a bright and curious soul with an imaginative mind.',
          ),
        ],
      ),
      Content(
        role: 'user',
        parts: [
          Part(
            text: 'Can you set it in a quiet village in 1600s France?',
          ),
        ],
      ),
    ],
  ),
);
print(res.candidates?.first.content?.parts?.first.text);
// In the heart of a tranquil village nestled amidst the rolling hills of 17th century France...

Streaming generated content

By default, generateContent returns a response after completing the entire generation process. You can achieve faster interactions by not waiting for the entire result, and instead use streamGenerateContent to handle partial results as they become available.

final stream = client.streamGenerateContent(
  modelId: 'gemini-pro',
  request: const GenerateContentRequest(
    contents: [
      Content(
        parts: [
          Part(text: 'Write a story about a magic backpack.'),
        ],
      ),
    ],
    generationConfig: GenerationConfig(
      temperature: 0.8,
    ),
  ),
);

stream.listen((final res) P
  print(res.candidates?.first.content?.parts?.first.text);
  // In a quaint little town nestled amidst rolling hills, there lived a...
)

Tuned model

Use the generateContentTunedModel method to generate content using a tuned model:

final res = await client.generateContentTunedModel(
  tunedModelId: 'my-tuned-model',
  request: GenerateContentRequest(
    //...
   ),
);

Count tokens #

When using long prompts, it might be useful to count tokens before sending any content to the model.

final res = await client.countTokens(
  modelId: 'gemini-pro',
  request: const CountTokensRequest(
    contents: [
      Content(
        parts: [
          Part(
            text: 'Write a story about a magic backpack.',
          ),
        ],
      ),
    ],
  ),
);
print(res.totalTokens);
// 8

Embedding #

Use the embedding-001 model with either embedContents or batchEmbedContents:

final res = await client.embedContent(
  modelId: 'embedding-001',
  request: const EmbedContentRequest(
    content: Content(
      parts: [
        Part(text: 'Write a story about a magic backpack.'),
      ],
    ),
  ),
);
print(res.embedding?.values);
// [0.008624583, -0.030451821, -0.042496547, ...]

Model info #

List models

Use the listModels method to list all the models available through the API, including both the Gemini and PaLM family models.

final res = await client.listModels();
print(res.models);
// [Model(name: models/gemini-pro, displayName: Gemini Pro, description: The best model...

Get model

Use the getModel method to return information about that model such as version, display name, input token limit, etc.

final res = await client.getModel(modelId: 'gemini-pro');
print(res);
// Model(name: models/gemini-pro, displayName: Gemini Pro, description: The best model...

Operations #

The following methods are available to manage operations:

  • listOperations()
  • deleteOperation(operationId: operationId)
  • listTunedModelOperations(tunedModelId: tunedModelId)
  • getTunedModelOperation(tunedModelId: tunedModelId, operationId: operationId)
  • cancelTunedModelOperation(tunedModelId: tunedModelId, operationId: operationId)

Advance Usage #

Default HTTP client #

By default, the client uses https://api.mistral.ai/v1 as the baseUrl and the following implementations of http.Client:

Custom HTTP client #

You can always provide your own implementation of http.Client for further customization:

final client = GoogleAIClient(
  apiKey: 'GOOGLEAI_API_KEY',
  client: MyHttpClient(),
);

Using a proxy #

HTTP proxy

You can use your own HTTP proxy by overriding the baseUrl and providing your required headers:

final client = GoogleAIClient(
  baseUrl: 'https://my-proxy.com',
  headers: {
      'x-my-proxy-header': 'value',
  },
);

If you need further customization, you can always provide your own http.Client.

SOCKS5 proxy

To use a SOCKS5 proxy, you can use the socks5_proxy package:

final baseHttpClient = HttpClient();
SocksTCPClient.assignToHttpClient(baseHttpClient, [
  ProxySettings(InternetAddress.loopbackIPv4, 1080),
]);
final httpClient = IOClient(baseClient);

final client = GoogleAIClient(
  client: httpClient,
);

Acknowledgements #

The generation of this client was made possible by the openapi_spec package.

License #

Google AI Dart Client is licensed under the MIT License.

14
likes
160
points
101
downloads

Publisher

verified publisherlangchaindart.dev

Weekly Downloads

Dart Client for the Google AI API (Gemini Pro, Gemini Pro Vision, embeddings, etc.).

Homepage
Repository (GitHub)
View/report issues
Contributing

Topics

#ai #nlp #llms #googleai #gemini

Documentation

Documentation
API reference

License

MIT (license)

Dependencies

fetch_client, freezed_annotation, http, json_annotation, meta

More

Packages that depend on googleai_dart