AWS SDK for C++

AWS SDK for C++ Version 1.11.440

Loading...
Searching...
No Matches
PromptModelInferenceConfiguration.h
1
6#pragma once
7#include <aws/bedrock-agent/BedrockAgent_EXPORTS.h>
8#include <aws/core/utils/memory/stl/AWSVector.h>
9#include <aws/core/utils/memory/stl/AWSString.h>
10#include <utility>
11
12namespace Aws
13{
14namespace Utils
15{
16namespace Json
17{
18 class JsonValue;
19 class JsonView;
20} // namespace Json
21} // namespace Utils
22namespace BedrockAgent
23{
24namespace Model
25{
26
36 {
37 public:
38 AWS_BEDROCKAGENT_API PromptModelInferenceConfiguration();
41 AWS_BEDROCKAGENT_API Aws::Utils::Json::JsonValue Jsonize() const;
42
43
45
48 inline int GetMaxTokens() const{ return m_maxTokens; }
49 inline bool MaxTokensHasBeenSet() const { return m_maxTokensHasBeenSet; }
50 inline void SetMaxTokens(int value) { m_maxTokensHasBeenSet = true; m_maxTokens = value; }
51 inline PromptModelInferenceConfiguration& WithMaxTokens(int value) { SetMaxTokens(value); return *this;}
53
55
59 inline const Aws::Vector<Aws::String>& GetStopSequences() const{ return m_stopSequences; }
60 inline bool StopSequencesHasBeenSet() const { return m_stopSequencesHasBeenSet; }
61 inline void SetStopSequences(const Aws::Vector<Aws::String>& value) { m_stopSequencesHasBeenSet = true; m_stopSequences = value; }
62 inline void SetStopSequences(Aws::Vector<Aws::String>&& value) { m_stopSequencesHasBeenSet = true; m_stopSequences = std::move(value); }
65 inline PromptModelInferenceConfiguration& AddStopSequences(const Aws::String& value) { m_stopSequencesHasBeenSet = true; m_stopSequences.push_back(value); return *this; }
66 inline PromptModelInferenceConfiguration& AddStopSequences(Aws::String&& value) { m_stopSequencesHasBeenSet = true; m_stopSequences.push_back(std::move(value)); return *this; }
67 inline PromptModelInferenceConfiguration& AddStopSequences(const char* value) { m_stopSequencesHasBeenSet = true; m_stopSequences.push_back(value); return *this; }
69
71
75 inline double GetTemperature() const{ return m_temperature; }
76 inline bool TemperatureHasBeenSet() const { return m_temperatureHasBeenSet; }
77 inline void SetTemperature(double value) { m_temperatureHasBeenSet = true; m_temperature = value; }
78 inline PromptModelInferenceConfiguration& WithTemperature(double value) { SetTemperature(value); return *this;}
80
82
86 inline double GetTopP() const{ return m_topP; }
87 inline bool TopPHasBeenSet() const { return m_topPHasBeenSet; }
88 inline void SetTopP(double value) { m_topPHasBeenSet = true; m_topP = value; }
89 inline PromptModelInferenceConfiguration& WithTopP(double value) { SetTopP(value); return *this;}
91 private:
92
93 int m_maxTokens;
94 bool m_maxTokensHasBeenSet = false;
95
96 Aws::Vector<Aws::String> m_stopSequences;
97 bool m_stopSequencesHasBeenSet = false;
98
99 double m_temperature;
100 bool m_temperatureHasBeenSet = false;
101
102 double m_topP;
103 bool m_topPHasBeenSet = false;
104 };
105
106} // namespace Model
107} // namespace BedrockAgent
108} // namespace Aws
PromptModelInferenceConfiguration & AddStopSequences(Aws::String &&value)
PromptModelInferenceConfiguration & WithTemperature(double value)
AWS_BEDROCKAGENT_API PromptModelInferenceConfiguration & operator=(Aws::Utils::Json::JsonView jsonValue)
PromptModelInferenceConfiguration & AddStopSequences(const char *value)
PromptModelInferenceConfiguration & AddStopSequences(const Aws::String &value)
PromptModelInferenceConfiguration & WithStopSequences(Aws::Vector< Aws::String > &&value)
PromptModelInferenceConfiguration & WithStopSequences(const Aws::Vector< Aws::String > &value)
AWS_BEDROCKAGENT_API PromptModelInferenceConfiguration(Aws::Utils::Json::JsonView jsonValue)
AWS_BEDROCKAGENT_API Aws::Utils::Json::JsonValue Jsonize() const
std::basic_string< char, std::char_traits< char >, Aws::Allocator< char > > String
std::vector< T, Aws::Allocator< T > > Vector
Aws::Utils::Json::JsonValue JsonValue