-
POST request - submits a prediction task for a specific
useCase
andmodelId
. The API responds with the following information:predictionId
. A unique UUID for the submitted prediction task that can be used later to retrieve the results.status
. The current state of the prediction task.
-
GET request - uses the
predictionId
you submit from a previously-submitted POST request and returns the results associated with that previous request.
For detailed API specifications in Swagger/OpenAPI format, see Platform APIs.
Prerequisites
To use this API, you need:- The unique
APPLICATION_ID
for your Lucidworks AI application. For more information, see credentials to use APIs. - A bearer token generated with a scope value of
machinelearning.predict
. For more information, see Authentication API. - The
USE_CASE
andMODEL_ID
fields in the/async-prediction
for the POST request. The path is/ai/async-prediction/USE_CASE/MODEL_ID
. A list of supported modes is returned in the Lucidworks AI Use Case API. For more information about supported models, see Generative AI models.
Common POST request parameters and fields
Some parameters in the/ai/async-prediction/USE_CASE/MODEL_ID
POST request are common to all of the generative AI (GenAI) use cases, such as the modelConfig
parameter.
Also referred to as hyperparameters, these fields set certain controls on the response.
Refer to the API spec for more information.
Unique values for the pass-through use case
Some parameter values available in thepass-through
use case are unique to this use case, including values for the useCaseConfig
parameter.
Refer to the API spec for more information.
Use System Prompt
This parameter controls whether the LLM input is automatically wrapped with a system prompt or passed directly.Use this parameter if custom prompts are needed or if the prompt response format needs to be manipulated. Be aware that including a system prompt may increase response time.
- The format for the
mistral-7b-instruct
model must be specific to Mistral:
https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2 - The format for the
llama-3-8b-instruct
model must be specific to Llama:
https://huggingface.co/blog/llama3#how-to-prompt-llama-3 - The text input for OpenAI models must be valid JSON to match the OpenAI API specification:
https://platform.openai.com/docs/api-reference/chat/create - The format for the Google Vertex AI models must adhere to the guidelines at:
https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/gemini
useSystemPrompt
POST example does not include modelConfig
parameters, but you can submit requests that include parameters described in Common POST request parameters and fields.
Data Type
This optional parameter enables model-specific handling in the Async Prediction API to help improve model accuracy. Use the most applicable value based on available data and the data type that best aligns with the text sent to the API.
"dataType": "json_prompt"`` example does not include
modelConfig` parameters, but you can submit requests that include parameters described in Common parameters and fields.