generative-ai-python

google.generativeai.protos.GenerateMessageRequest

Request to generate a message response from the model.

`model` `str` Required. The name of the model to use. Format: ``name=models/{model}``.
`prompt` `google.ai.generativelanguage.MessagePrompt` Required. The structured textual input given to the model as a prompt. Given a prompt, the model will return what it predicts is the next message in the discussion.
`temperature` `float` Optional. Controls the randomness of the output. Values can range over ``[0.0,1.0]``, inclusive. A value closer to ``1.0`` will produce responses that are more varied, while a value closer to ``0.0`` will typically result in less surprising responses from the model.
`candidate_count` `int` Optional. The number of generated response messages to return. This value must be between ``[1, 8]``, inclusive. If unset, this will default to ``1``.
`top_p` `float` Optional. The maximum cumulative probability of tokens to consider when sampling. The model uses combined Top-k and nucleus sampling. Nucleus sampling considers the smallest set of tokens whose probability sum is at least ``top_p``.
`top_k` `int` Optional. The maximum number of tokens to consider when sampling. The model uses combined Top-k and nucleus sampling. Top-k sampling considers the set of ``top_k`` most probable tokens.