diff --git a/.codegen.json b/.codegen.json index 1abada67..0d6c421f 100644 --- a/.codegen.json +++ b/.codegen.json @@ -1 +1 @@ -{ "engineHash": "11fbad9", "specHash": "abd6037", "version": "1.5.1" } +{ "engineHash": "2efc8ab", "specHash": "90cf4e4", "version": "1.5.1" } diff --git a/box_sdk_gen/schemas/__init__.py b/box_sdk_gen/schemas/__init__.py index 088e2df3..56bb745a 100644 --- a/box_sdk_gen/schemas/__init__.py +++ b/box_sdk_gen/schemas/__init__.py @@ -6,6 +6,8 @@ from box_sdk_gen.schemas.ai_item_base import * +from box_sdk_gen.schemas.ai_llm_endpoint_params_aws import * + from box_sdk_gen.schemas.ai_llm_endpoint_params_google import * from box_sdk_gen.schemas.ai_llm_endpoint_params_open_ai import * diff --git a/box_sdk_gen/schemas/ai_agent_basic_gen_tool.py b/box_sdk_gen/schemas/ai_agent_basic_gen_tool.py index a97194a7..e1bfae68 100644 --- a/box_sdk_gen/schemas/ai_agent_basic_gen_tool.py +++ b/box_sdk_gen/schemas/ai_agent_basic_gen_tool.py @@ -6,6 +6,8 @@ from box_sdk_gen.schemas.ai_llm_endpoint_params_google import AiLlmEndpointParamsGoogle +from box_sdk_gen.schemas.ai_llm_endpoint_params_aws import AiLlmEndpointParamsAws + from box_sdk_gen.schemas.ai_agent_basic_text_tool_base import AiAgentBasicTextToolBase from box_sdk_gen.schemas.ai_agent_basic_text_tool_text_gen import ( @@ -32,7 +34,11 @@ def __init__( model: Optional[str] = None, num_tokens_for_completion: Optional[int] = None, llm_endpoint_params: Optional[ - Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle] + Union[ + AiLlmEndpointParamsOpenAi, + AiLlmEndpointParamsGoogle, + AiLlmEndpointParamsAws, + ] ] = None, **kwargs ): @@ -53,7 +59,7 @@ def __init__( :param num_tokens_for_completion: The number of tokens for completion., defaults to None :type num_tokens_for_completion: Optional[int], optional :param llm_endpoint_params: The parameters for the LLM endpoint specific to OpenAI / Google models., defaults to None - :type llm_endpoint_params: Optional[Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle]], optional + :type llm_endpoint_params: Optional[Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle, AiLlmEndpointParamsAws]], optional """ super().__init__( embeddings=embeddings, diff --git a/box_sdk_gen/schemas/ai_agent_basic_text_tool.py b/box_sdk_gen/schemas/ai_agent_basic_text_tool.py index bd641e16..0dcb792d 100644 --- a/box_sdk_gen/schemas/ai_agent_basic_text_tool.py +++ b/box_sdk_gen/schemas/ai_agent_basic_text_tool.py @@ -6,6 +6,8 @@ from box_sdk_gen.schemas.ai_llm_endpoint_params_google import AiLlmEndpointParamsGoogle +from box_sdk_gen.schemas.ai_llm_endpoint_params_aws import AiLlmEndpointParamsAws + from box_sdk_gen.schemas.ai_agent_basic_text_tool_base import AiAgentBasicTextToolBase @@ -18,7 +20,11 @@ def __init__( model: Optional[str] = None, num_tokens_for_completion: Optional[int] = None, llm_endpoint_params: Optional[ - Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle] + Union[ + AiLlmEndpointParamsOpenAi, + AiLlmEndpointParamsGoogle, + AiLlmEndpointParamsAws, + ] ] = None, **kwargs ): @@ -34,7 +40,7 @@ def __init__( :param num_tokens_for_completion: The number of tokens for completion., defaults to None :type num_tokens_for_completion: Optional[int], optional :param llm_endpoint_params: The parameters for the LLM endpoint specific to OpenAI / Google models., defaults to None - :type llm_endpoint_params: Optional[Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle]], optional + :type llm_endpoint_params: Optional[Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle, AiLlmEndpointParamsAws]], optional """ super().__init__( model=model, diff --git a/box_sdk_gen/schemas/ai_agent_basic_text_tool_base.py b/box_sdk_gen/schemas/ai_agent_basic_text_tool_base.py index cef0f155..df7a9cea 100644 --- a/box_sdk_gen/schemas/ai_agent_basic_text_tool_base.py +++ b/box_sdk_gen/schemas/ai_agent_basic_text_tool_base.py @@ -8,6 +8,8 @@ from box_sdk_gen.schemas.ai_llm_endpoint_params_google import AiLlmEndpointParamsGoogle +from box_sdk_gen.schemas.ai_llm_endpoint_params_aws import AiLlmEndpointParamsAws + class AiAgentBasicTextToolBase(BaseObject): def __init__( @@ -16,7 +18,11 @@ def __init__( model: Optional[str] = None, num_tokens_for_completion: Optional[int] = None, llm_endpoint_params: Optional[ - Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle] + Union[ + AiLlmEndpointParamsOpenAi, + AiLlmEndpointParamsGoogle, + AiLlmEndpointParamsAws, + ] ] = None, **kwargs ): @@ -26,7 +32,7 @@ def __init__( :param num_tokens_for_completion: The number of tokens for completion., defaults to None :type num_tokens_for_completion: Optional[int], optional :param llm_endpoint_params: The parameters for the LLM endpoint specific to OpenAI / Google models., defaults to None - :type llm_endpoint_params: Optional[Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle]], optional + :type llm_endpoint_params: Optional[Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle, AiLlmEndpointParamsAws]], optional """ super().__init__(**kwargs) self.model = model diff --git a/box_sdk_gen/schemas/ai_agent_basic_text_tool_text_gen.py b/box_sdk_gen/schemas/ai_agent_basic_text_tool_text_gen.py index abab0282..fc1d1870 100644 --- a/box_sdk_gen/schemas/ai_agent_basic_text_tool_text_gen.py +++ b/box_sdk_gen/schemas/ai_agent_basic_text_tool_text_gen.py @@ -6,6 +6,8 @@ from box_sdk_gen.schemas.ai_llm_endpoint_params_google import AiLlmEndpointParamsGoogle +from box_sdk_gen.schemas.ai_llm_endpoint_params_aws import AiLlmEndpointParamsAws + from box_sdk_gen.schemas.ai_agent_basic_text_tool_base import AiAgentBasicTextToolBase @@ -18,7 +20,11 @@ def __init__( model: Optional[str] = None, num_tokens_for_completion: Optional[int] = None, llm_endpoint_params: Optional[ - Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle] + Union[ + AiLlmEndpointParamsOpenAi, + AiLlmEndpointParamsGoogle, + AiLlmEndpointParamsAws, + ] ] = None, **kwargs ): @@ -36,7 +42,7 @@ def __init__( :param num_tokens_for_completion: The number of tokens for completion., defaults to None :type num_tokens_for_completion: Optional[int], optional :param llm_endpoint_params: The parameters for the LLM endpoint specific to OpenAI / Google models., defaults to None - :type llm_endpoint_params: Optional[Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle]], optional + :type llm_endpoint_params: Optional[Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle, AiLlmEndpointParamsAws]], optional """ super().__init__( model=model, diff --git a/box_sdk_gen/schemas/ai_agent_long_text_tool.py b/box_sdk_gen/schemas/ai_agent_long_text_tool.py index ce7b3ec4..54865b67 100644 --- a/box_sdk_gen/schemas/ai_agent_long_text_tool.py +++ b/box_sdk_gen/schemas/ai_agent_long_text_tool.py @@ -8,6 +8,8 @@ from box_sdk_gen.schemas.ai_llm_endpoint_params_google import AiLlmEndpointParamsGoogle +from box_sdk_gen.schemas.ai_llm_endpoint_params_aws import AiLlmEndpointParamsAws + from box_sdk_gen.schemas.ai_agent_basic_text_tool_base import AiAgentBasicTextToolBase from box_sdk_gen.schemas.ai_agent_basic_text_tool import AiAgentBasicTextTool @@ -59,7 +61,11 @@ def __init__( model: Optional[str] = None, num_tokens_for_completion: Optional[int] = None, llm_endpoint_params: Optional[ - Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle] + Union[ + AiLlmEndpointParamsOpenAi, + AiLlmEndpointParamsGoogle, + AiLlmEndpointParamsAws, + ] ] = None, **kwargs ): @@ -75,7 +81,7 @@ def __init__( :param num_tokens_for_completion: The number of tokens for completion., defaults to None :type num_tokens_for_completion: Optional[int], optional :param llm_endpoint_params: The parameters for the LLM endpoint specific to OpenAI / Google models., defaults to None - :type llm_endpoint_params: Optional[Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle]], optional + :type llm_endpoint_params: Optional[Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle, AiLlmEndpointParamsAws]], optional """ super().__init__( system_message=system_message, diff --git a/box_sdk_gen/schemas/ai_agent_long_text_tool_text_gen.py b/box_sdk_gen/schemas/ai_agent_long_text_tool_text_gen.py index da04fb75..e2c6e191 100644 --- a/box_sdk_gen/schemas/ai_agent_long_text_tool_text_gen.py +++ b/box_sdk_gen/schemas/ai_agent_long_text_tool_text_gen.py @@ -8,6 +8,8 @@ from box_sdk_gen.schemas.ai_llm_endpoint_params_google import AiLlmEndpointParamsGoogle +from box_sdk_gen.schemas.ai_llm_endpoint_params_aws import AiLlmEndpointParamsAws + from box_sdk_gen.schemas.ai_agent_basic_text_tool_base import AiAgentBasicTextToolBase from box_sdk_gen.schemas.ai_agent_basic_text_tool_text_gen import ( @@ -61,7 +63,11 @@ def __init__( model: Optional[str] = None, num_tokens_for_completion: Optional[int] = None, llm_endpoint_params: Optional[ - Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle] + Union[ + AiLlmEndpointParamsOpenAi, + AiLlmEndpointParamsGoogle, + AiLlmEndpointParamsAws, + ] ] = None, **kwargs ): @@ -79,7 +85,7 @@ def __init__( :param num_tokens_for_completion: The number of tokens for completion., defaults to None :type num_tokens_for_completion: Optional[int], optional :param llm_endpoint_params: The parameters for the LLM endpoint specific to OpenAI / Google models., defaults to None - :type llm_endpoint_params: Optional[Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle]], optional + :type llm_endpoint_params: Optional[Union[AiLlmEndpointParamsOpenAi, AiLlmEndpointParamsGoogle, AiLlmEndpointParamsAws]], optional """ super().__init__( system_message=system_message, diff --git a/box_sdk_gen/schemas/ai_llm_endpoint_params_aws.py b/box_sdk_gen/schemas/ai_llm_endpoint_params_aws.py new file mode 100644 index 00000000..ede7d4f6 --- /dev/null +++ b/box_sdk_gen/schemas/ai_llm_endpoint_params_aws.py @@ -0,0 +1,39 @@ +from enum import Enum + +from typing import Optional + +from box_sdk_gen.internal.base_object import BaseObject + + +class AiLlmEndpointParamsAwsTypeField(str, Enum): + AWS_PARAMS = 'aws_params' + + +class AiLlmEndpointParamsAws(BaseObject): + _discriminator = 'type', {'aws_params'} + + def __init__( + self, + *, + type: AiLlmEndpointParamsAwsTypeField = AiLlmEndpointParamsAwsTypeField.AWS_PARAMS.value, + temperature: Optional[float] = None, + top_p: Optional[float] = None, + **kwargs + ): + """ + :param type: The type of the AI LLM endpoint params object for AWS. + This parameter is **required**., defaults to AiLlmEndpointParamsAwsTypeField.AWS_PARAMS.value + :type type: AiLlmEndpointParamsAwsTypeField, optional + :param temperature: What sampling temperature to use, between 0 and 1. Higher values like 0.8 will make the output more random, + while lower values like 0.2 will make it more focused and deterministic. + We generally recommend altering this or `top_p` but not both., defaults to None + :type temperature: Optional[float], optional + :param top_p: An alternative to sampling with temperature, called nucleus sampling, where the model considers the results + of the tokens with `top_p` probability mass. So 0.1 means only the tokens comprising the top 10% probability + mass are considered. We generally recommend altering this or temperature but not both., defaults to None + :type top_p: Optional[float], optional + """ + super().__init__(**kwargs) + self.type = type + self.temperature = temperature + self.top_p = top_p diff --git a/box_sdk_gen/schemas/client_error.py b/box_sdk_gen/schemas/client_error.py index 7c03a84b..d347756b 100644 --- a/box_sdk_gen/schemas/client_error.py +++ b/box_sdk_gen/schemas/client_error.py @@ -1,9 +1,11 @@ from enum import Enum -from box_sdk_gen.internal.base_object import BaseObject - from typing import Optional +from typing import Dict + +from box_sdk_gen.internal.base_object import BaseObject + class ClientErrorTypeField(str, Enum): ERROR = 'error' @@ -29,11 +31,6 @@ class ClientErrorCodeField(str, Enum): INSUFFICIENT_SCOPE = 'insufficient_scope' -class ClientErrorContextInfoField(BaseObject): - def __init__(self, **kwargs): - super().__init__(**kwargs) - - class ClientError(BaseObject): _discriminator = 'type', {'error'} @@ -44,7 +41,7 @@ def __init__( status: Optional[int] = None, code: Optional[ClientErrorCodeField] = None, message: Optional[str] = None, - context_info: Optional[ClientErrorContextInfoField] = None, + context_info: Optional[Dict] = None, help_url: Optional[str] = None, request_id: Optional[str] = None, **kwargs @@ -61,7 +58,7 @@ def __init__( :param context_info: A free-form object that contains additional context about the error. The possible fields are defined on a per-endpoint basis. `message` is only one example., defaults to None - :type context_info: Optional[ClientErrorContextInfoField], optional + :type context_info: Optional[Dict], optional :param help_url: A URL that links to more information about why this error occurred., defaults to None :type help_url: Optional[str], optional :param request_id: A unique identifier for this response, which can be used diff --git a/box_sdk_gen/schemas/conflict_error.py b/box_sdk_gen/schemas/conflict_error.py index 602bd055..4c1ff4b5 100644 --- a/box_sdk_gen/schemas/conflict_error.py +++ b/box_sdk_gen/schemas/conflict_error.py @@ -4,12 +4,12 @@ from box_sdk_gen.internal.base_object import BaseObject +from typing import Dict + from box_sdk_gen.schemas.client_error import ClientErrorTypeField from box_sdk_gen.schemas.client_error import ClientErrorCodeField -from box_sdk_gen.schemas.client_error import ClientErrorContextInfoField - from box_sdk_gen.schemas.client_error import ClientError from box_sdk_gen.schemas.file_conflict import FileConflict @@ -33,7 +33,7 @@ def __init__( status: Optional[int] = None, code: Optional[ClientErrorCodeField] = None, message: Optional[str] = None, - context_info: Optional[ClientErrorContextInfoField] = None, + context_info: Optional[Dict] = None, help_url: Optional[str] = None, request_id: Optional[str] = None, **kwargs @@ -50,7 +50,7 @@ def __init__( :param context_info: A free-form object that contains additional context about the error. The possible fields are defined on a per-endpoint basis. `message` is only one example., defaults to None - :type context_info: Optional[ClientErrorContextInfoField], optional + :type context_info: Optional[Dict], optional :param help_url: A URL that links to more information about why this error occurred., defaults to None :type help_url: Optional[str], optional :param request_id: A unique identifier for this response, which can be used