44from mindee .client_mixin import ClientMixin
55from mindee .error .mindee_error import MindeeError
66from mindee .error .mindee_http_error_v2 import handle_error_v2
7- from mindee .input .inference_predict_options import InferencePredictOptions
7+ from mindee .input .inference_parameters import InferenceParameters
88from mindee .input .local_response import LocalResponse
99from mindee .input .polling_options import PollingOptions
1010from mindee .input .sources .local_input_source import LocalInputSource
@@ -38,28 +38,28 @@ def __init__(self, api_key: Optional[str] = None) -> None:
3838 self .mindee_api = MindeeApiV2 (api_key )
3939
4040 def enqueue (
41- self , input_source : LocalInputSource , options : InferencePredictOptions
41+ self , input_source : LocalInputSource , params : InferenceParameters
4242 ) -> JobResponse :
4343 """
4444 Enqueues a document to a given model.
4545
4646 :param input_source: The document/source file to use.
4747 Has to be created beforehand.
4848
49- :param options: Options for the prediction .
49+ :param params: Parameters to set when sending a file .
5050 :return: A valid inference response.
5151 """
52- logger .debug ("Enqueuing document to '%s'" , options .model_id )
52+ logger .debug ("Enqueuing document to '%s'" , params .model_id )
5353
54- if options .page_options and input_source .is_pdf ():
54+ if params .page_options and input_source .is_pdf ():
5555 input_source .process_pdf (
56- options .page_options .operation ,
57- options .page_options .on_min_pages ,
58- options .page_options .page_indexes ,
56+ params .page_options .operation ,
57+ params .page_options .on_min_pages ,
58+ params .page_options .page_indexes ,
5959 )
6060
6161 response = self .mindee_api .predict_async_req_post (
62- input_source = input_source , options = options
62+ input_source = input_source , options = params
6363 )
6464 dict_response = response .json ()
6565
@@ -89,35 +89,35 @@ def parse_queued(
8989 return InferenceResponse (dict_response )
9090
9191 def enqueue_and_parse (
92- self , input_source : LocalInputSource , options : InferencePredictOptions
92+ self , input_source : LocalInputSource , params : InferenceParameters
9393 ) -> InferenceResponse :
9494 """
9595 Enqueues to an asynchronous endpoint and automatically polls for a response.
9696
9797 :param input_source: The document/source file to use.
9898 Has to be created beforehand.
9999
100- :param options: Options for the prediction .
100+ :param params: Parameters to set when sending a file .
101101
102102 :return: A valid inference response.
103103 """
104- if not options .polling_options :
105- options .polling_options = PollingOptions ()
104+ if not params .polling_options :
105+ params .polling_options = PollingOptions ()
106106 self ._validate_async_params (
107- options .polling_options .initial_delay_sec ,
108- options .polling_options .delay_sec ,
109- options .polling_options .max_retries ,
107+ params .polling_options .initial_delay_sec ,
108+ params .polling_options .delay_sec ,
109+ params .polling_options .max_retries ,
110110 )
111- queue_result = self .enqueue (input_source , options )
111+ queue_result = self .enqueue (input_source , params )
112112 logger .debug (
113113 "Successfully enqueued document with job id: %s" , queue_result .job .id
114114 )
115- sleep (options .polling_options .initial_delay_sec )
115+ sleep (params .polling_options .initial_delay_sec )
116116 retry_counter = 1
117117 poll_results = self .parse_queued (
118118 queue_result .job .id ,
119119 )
120- while retry_counter < options .polling_options .max_retries :
120+ while retry_counter < params .polling_options .max_retries :
121121 if not isinstance (poll_results , JobResponse ):
122122 break
123123 if poll_results .job .status == "Failed" :
@@ -133,7 +133,7 @@ def enqueue_and_parse(
133133 queue_result .job .id ,
134134 )
135135 retry_counter += 1
136- sleep (options .polling_options .delay_sec )
136+ sleep (params .polling_options .delay_sec )
137137 poll_results = self .parse_queued (queue_result .job .id )
138138
139139 if not isinstance (poll_results , InferenceResponse ):
0 commit comments