Product Selector

Fusion 5.9
    Fusion 5.9

    Use Case APILucidworks AI

    The Lucidworks AI Use Case API returns a list of each use case and its associated models.

    You can enter the values returned in the response of the Use Case API in the:

    Prerequisites

    To use this API, you need:

    • The unique APPLICATION_ID for your Lucidworks AI application. For more information, see credentials to use APIs.

    • A bearer token generated with a scope value of machinelearning.predict. For more information, see Authentication API.

    Request

    curl --request GET \
      --url https://APPLICATION_ID.applications.lucidworks.com/ai/usecase\
      --header 'Authorization: Bearer ACCESS_TOKEN'

    Prediction API Response

    The following example response lists some of the supported models. For a complete list of supported models, see Generative AI models.

    [
    	{
    		"useCase": "embedding",
    		"models": [
    
    			{
    				"id": "e5-small-v2",
    				"name": "e5-small-v2",
    				"modelType": "shared",
    				"vectorSize": 384
    			},
    			{
    				"id": "text-encoder",
    				"name": "text-encoder",
    				"modelType": "shared",
    				"vectorSize": 768
    			}
    
    			{
    				"id": "multi-qa-distilbert-cos-v1",
    				"name": "multi-qa-distilbert-cos-v1",
    				"modelType": "shared",
    				"vectorSize": 768
    			},
    			{
    				"id": "multilingual-e5-base",
    				"name": "multilingual-e5-base",
    				"modelType": "shared",
    				"vectorSize": 768
    			},
    			{
    				"id": "6dc3462d-ae64-4247-ac5e-e1bd95db67d8",
    				"name": "model-6dc3462d-ae64-4247-ac5e-e1bd95db67d8",
    				"modelType": "ecommerce-rnn"
    			}
    		]
    	},
    	{
    		"useCase": "rag",
    		"models": [
    			{
    				"id": "llama-3-8b-instruct",
    				"name": "llama-3-8b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-3.5-turbo",
    				"name": "gpt-3.5-turbo",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-4-turbo",
    				"name": "gpt-4-turbo",
    				"modelType": "shared"
    			}
    		]
    	},
    	{
    		"useCase": "passthrough",
    		"models": [
    			{
    				"id": "mistral-7b-instruct",
    				"name": "mistral-7b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "llama-3-8b-instruct",
    				"name": "llama-3-8b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-3.5-turbo",
    				"name": "gpt-3.5-turbo",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-4-turbo",
    				"name": "gpt-4-turbo",
    				"modelType": "shared"
    			}
    		]
    	},
    	{
    		"useCase": "summarization",
    		"models": [
    			{
    				"id": "mistral-7b-instruct",
    				"name": "mistral-7b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "llama-3-8b-instruct",
    				"name": "llama-3-8b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-3.5-turbo",
    				"name": "gpt-3.5-turbo",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-4-turbo",
    				"name": "gpt-4-turbo",
    				"modelType": "shared"
    			}
    		]
    	},
    	{
    		"useCase": "keyword_extraction",
    		"models": [
    			{
    				"id": "mistral-7b-instruct",
    				"name": "mistral-7b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "llama-3-8b-instruct",
    				"name": "llama-3-8b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-3.5-turbo",
    				"name": "gpt-3.5-turbo",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-4-turbo",
    				"name": "gpt-4-turbo",
    				"modelType": "shared"
    			}
    		]
    	},
    	{
    		"useCase": "standalone_query_rewriter",
    		"models": [
    			{
    				"id": "llama-3-8b-instruct",
    				"name": "llama-3-8b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-3.5-turbo",
    				"name": "gpt-3.5-turbo",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-4-turbo",
    				"name": "gpt-4-turbo",
    				"modelType": "shared"
    			},
    		]
    	},
    	{
    		"useCase": "ner",
    		"models": [
    			{
    				"id": "mistral-7b-instruct",
    				"name": "mistral-7b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "llama-3-8b-instruct",
    				"name": "llama-3-8b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-3.5-turbo",
    				"name": "gpt-3.5-turbo",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-4-turbo",
    				"name": "gpt-4-turbo",
    				"modelType": "shared"
    			}
    		]
    	}
    ]

    Async Prediction API Response

    The following example response lists some of the supported models. For a complete list of supported models, see Generative AI models.

    [
    	{
    		"useCase": "rag",
    		"models": [
    			{
    				"id": "llama-3-8b-instruct",
    				"name": "llama-3-8b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-3.5-turbo",
    				"name": "gpt-3.5-turbo",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-4-turbo",
    				"name": "gpt-4-turbo",
    				"modelType": "shared"
    			}
    		]
    	},
    	{
    		"useCase": "passthrough",
    		"models": [
    			{
    				"id": "mistral-7b-instruct",
    				"name": "mistral-7b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "llama-3-8b-instruct",
    				"name": "llama-3-8b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-3.5-turbo",
    				"name": "gpt-3.5-turbo",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-4-turbo",
    				"name": "gpt-4-turbo",
    				"modelType": "shared"
    			}
    		]
    	},
    	{
    		"useCase": "summarization",
    		"models": [
    			{
    				"id": "mistral-7b-instruct",
    				"name": "mistral-7b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "llama-3-8b-instruct",
    				"name": "llama-3-8b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-3.5-turbo",
    				"name": "gpt-3.5-turbo",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-4-turbo",
    				"name": "gpt-4-turbo",
    				"modelType": "shared"
    			}
    		]
    	},
    	{
    		"useCase": "keyword_extraction",
    		"models": [
    			{
    				"id": "mistral-7b-instruct",
    				"name": "mistral-7b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "llama-3-8b-instruct",
    				"name": "llama-3-8b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-3.5-turbo",
    				"name": "gpt-3.5-turbo",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-4-turbo",
    				"name": "gpt-4-turbo",
    				"modelType": "shared"
    			}
    		]
    	},
    	{
    		"useCase": "standalone_query_rewriter",
    		"models": [
    			{
    				"id": "llama-3-8b-instruct",
    				"name": "llama-3-8b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-3.5-turbo",
    				"name": "gpt-3.5-turbo",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-4-turbo",
    				"name": "gpt-4-turbo",
    				"modelType": "shared"
    			},
    		]
    	},
    	{
    		"useCase": "ner",
    		"models": [
    			{
    				"id": "mistral-7b-instruct",
    				"name": "mistral-7b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "llama-3-8b-instruct",
    				"name": "llama-3-8b-instruct",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-3.5-turbo",
    				"name": "gpt-3.5-turbo",
    				"modelType": "shared"
    			},
    			{
    				"id": "gpt-4-turbo",
    				"name": "gpt-4-turbo",
    				"modelType": "shared"
    			}
    		]
    	}
    ]