serverStreamingPredict method
Future<GoogleCloudAiplatformV1StreamingPredictResponse>
serverStreamingPredict(
- GoogleCloudAiplatformV1StreamingPredictRequest request,
- String endpoint, {
- String? $fields,
Perform a server-side streaming online prediction request for Vertex LLM streaming.
request
- The metadata request object.
Request parameters:
endpoint
- Required. The name of the Endpoint requested to serve the
prediction. Format:
projects/{project}/locations/{location}/endpoints/{endpoint}
Value must have pattern
^projects/\[^/\]+/locations/\[^/\]+/publishers/\[^/\]+/models/\[^/\]+$
.
$fields
- Selector specifying which fields to include in a partial
response.
Completes with a GoogleCloudAiplatformV1StreamingPredictResponse.
Completes with a commons.ApiRequestError if the API endpoint returned an error.
If the used http.Client
completes with an error when making a REST call,
this method will complete with the same error.
Implementation
async.Future<GoogleCloudAiplatformV1StreamingPredictResponse>
serverStreamingPredict(
GoogleCloudAiplatformV1StreamingPredictRequest request,
core.String endpoint, {
core.String? $fields,
}) async {
final body_ = convert.json.encode(request);
final queryParams_ = <core.String, core.List<core.String>>{
if ($fields != null) 'fields': [$fields],
};
final url_ =
'v1/' + core.Uri.encodeFull('$endpoint') + ':serverStreamingPredict';
final response_ = await _requester.request(
url_,
'POST',
body: body_,
queryParams: queryParams_,
);
return GoogleCloudAiplatformV1StreamingPredictResponse.fromJson(
response_ as core.Map<core.String, core.dynamic>);
}