@@ -34,6 +34,7 @@ class GetLLMModelEndpointV1Response(schemas.DictSchema):
3434"name" ,
3535"id" ,
3636"source" ,
37+ "status" ,
3738 }
3839
3940class properties :
@@ -47,6 +48,9 @@ class GetLLMModelEndpointV1Response(schemas.DictSchema):
4748@staticmethod
4849def source () -> typing .Type ["LLMSource" ]:
4950return LLMSource
51+ @staticmethod
52+ def status () -> typing .Type ["ModelEndpointStatus" ]:
53+ return ModelEndpointStatus
5054inference_framework_image_tag = schemas .StrSchema
5155model_name = schemas .StrSchema
5256num_shards = schemas .IntSchema
@@ -62,6 +66,7 @@ class GetLLMModelEndpointV1Response(schemas.DictSchema):
6266"inference_framework" : inference_framework ,
6367"name" : name ,
6468"source" : source ,
69+ "status" : status ,
6570"inference_framework_image_tag" : inference_framework_image_tag ,
6671"model_name" : model_name ,
6772"num_shards" : num_shards ,
@@ -72,6 +77,7 @@ class GetLLMModelEndpointV1Response(schemas.DictSchema):
7277name : MetaOapg .properties .name
7378id : MetaOapg .properties .id
7479source : "LLMSource"
80+ status : "ModelEndpointStatus"
7581
7682@typing .overload
7783def __getitem__ (self , name : typing_extensions .Literal ["id" ]) -> MetaOapg .properties .id : ...
@@ -82,6 +88,8 @@ class GetLLMModelEndpointV1Response(schemas.DictSchema):
8288@typing .overload
8389def __getitem__ (self , name : typing_extensions .Literal ["source" ]) -> "LLMSource" : ...
8490@typing .overload
91+ def __getitem__ (self , name : typing_extensions .Literal ["status" ]) -> "ModelEndpointStatus" : ...
92+ @typing .overload
8593def __getitem__ (
8694self , name : typing_extensions .Literal ["inference_framework_image_tag" ]
8795 ) -> MetaOapg .properties .inference_framework_image_tag : ...
@@ -103,6 +111,7 @@ class GetLLMModelEndpointV1Response(schemas.DictSchema):
103111"inference_framework" ,
104112"name" ,
105113"source" ,
114+ "status" ,
106115"inference_framework_image_tag" ,
107116"model_name" ,
108117"num_shards" ,
@@ -123,6 +132,8 @@ class GetLLMModelEndpointV1Response(schemas.DictSchema):
123132@typing .overload
124133def get_item_oapg (self , name : typing_extensions .Literal ["source" ]) -> "LLMSource" : ...
125134@typing .overload
135+ def get_item_oapg (self , name : typing_extensions .Literal ["status" ]) -> "ModelEndpointStatus" : ...
136+ @typing .overload
126137def get_item_oapg (
127138self , name : typing_extensions .Literal ["inference_framework_image_tag" ]
128139 ) -> typing .Union [MetaOapg .properties .inference_framework_image_tag , schemas .Unset ]: ...
@@ -152,6 +163,7 @@ class GetLLMModelEndpointV1Response(schemas.DictSchema):
152163"inference_framework" ,
153164"name" ,
154165"source" ,
166+ "status" ,
155167"inference_framework_image_tag" ,
156168"model_name" ,
157169"num_shards" ,
@@ -178,6 +190,7 @@ class GetLLMModelEndpointV1Response(schemas.DictSchema):
178190str ,
179191 ],
180192source : "LLMSource" ,
193+ status : "ModelEndpointStatus" ,
181194inference_framework_image_tag : typing .Union [
182195MetaOapg .properties .inference_framework_image_tag , str , schemas .Unset
183196 ] = schemas .unset ,
@@ -210,6 +223,7 @@ class GetLLMModelEndpointV1Response(schemas.DictSchema):
210223name = name ,
211224id = id ,
212225source = source ,
226+ status = status ,
213227inference_framework_image_tag = inference_framework_image_tag ,
214228model_name = model_name ,
215229num_shards = num_shards ,
@@ -224,4 +238,5 @@ from launch_client.model.get_model_endpoint_v1_response import (
224238)
225239from launch_client .model .llm_inference_framework import LLMInferenceFramework
226240from launch_client .model .llm_source import LLMSource
241+ from launch_client .model .model_endpoint_status import ModelEndpointStatus
227242from launch_client .model .quantization import Quantization
0 commit comments