|
@@ -54,6 +54,7 @@ import org.elasticsearch.xpack.inference.services.elastic.authorization.ElasticI
|
|
import org.elasticsearch.xpack.inference.services.elastic.completion.ElasticInferenceServiceCompletionModel;
|
|
import org.elasticsearch.xpack.inference.services.elastic.completion.ElasticInferenceServiceCompletionModel;
|
|
import org.elasticsearch.xpack.inference.services.elastic.completion.ElasticInferenceServiceCompletionServiceSettings;
|
|
import org.elasticsearch.xpack.inference.services.elastic.completion.ElasticInferenceServiceCompletionServiceSettings;
|
|
import org.elasticsearch.xpack.inference.services.settings.RateLimitSettings;
|
|
import org.elasticsearch.xpack.inference.services.settings.RateLimitSettings;
|
|
|
|
+import org.elasticsearch.xpack.inference.services.validation.ModelValidatorBuilder;
|
|
import org.elasticsearch.xpack.inference.telemetry.TraceContext;
|
|
import org.elasticsearch.xpack.inference.telemetry.TraceContext;
|
|
|
|
|
|
import java.util.ArrayList;
|
|
import java.util.ArrayList;
|
|
@@ -557,11 +558,8 @@ public class ElasticInferenceService extends SenderService {
|
|
|
|
|
|
@Override
|
|
@Override
|
|
public void checkModelConfig(Model model, ActionListener<Model> listener) {
|
|
public void checkModelConfig(Model model, ActionListener<Model> listener) {
|
|
- if (model instanceof ElasticInferenceServiceSparseEmbeddingsModel embeddingsModel) {
|
|
|
|
- listener.onResponse(updateModelWithEmbeddingDetails(embeddingsModel));
|
|
|
|
- } else {
|
|
|
|
- listener.onResponse(model);
|
|
|
|
- }
|
|
|
|
|
|
+ // TODO: Remove this function once all services have been updated to use the new model validators
|
|
|
|
+ ModelValidatorBuilder.buildModelValidator(model.getTaskType()).validate(this, model, listener);
|
|
}
|
|
}
|
|
|
|
|
|
private static List<ChunkedInference> translateToChunkedResults(InferenceInputs inputs, InferenceServiceResults inferenceResults) {
|
|
private static List<ChunkedInference> translateToChunkedResults(InferenceInputs inputs, InferenceServiceResults inferenceResults) {
|
|
@@ -576,18 +574,6 @@ public class ElasticInferenceService extends SenderService {
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
- private ElasticInferenceServiceSparseEmbeddingsModel updateModelWithEmbeddingDetails(
|
|
|
|
- ElasticInferenceServiceSparseEmbeddingsModel model
|
|
|
|
- ) {
|
|
|
|
- ElasticInferenceServiceSparseEmbeddingsServiceSettings serviceSettings = new ElasticInferenceServiceSparseEmbeddingsServiceSettings(
|
|
|
|
- model.getServiceSettings().modelId(),
|
|
|
|
- model.getServiceSettings().maxInputTokens(),
|
|
|
|
- model.getServiceSettings().rateLimitSettings()
|
|
|
|
- );
|
|
|
|
-
|
|
|
|
- return new ElasticInferenceServiceSparseEmbeddingsModel(model, serviceSettings);
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
private TraceContext getCurrentTraceInfo() {
|
|
private TraceContext getCurrentTraceInfo() {
|
|
var threadPool = getServiceComponents().threadPool();
|
|
var threadPool = getServiceComponents().threadPool();
|
|
|
|
|