| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213 | from typing import IO, Generator, List, Optional, Union, castfrom core.entities.provider_configuration import ProviderModelBundlefrom core.errors.error import ProviderTokenNotInitErrorfrom core.model_runtime.callbacks.base_callback import Callbackfrom core.model_runtime.entities.llm_entities import LLMResultfrom core.model_runtime.entities.message_entities import PromptMessage, PromptMessageToolfrom core.model_runtime.entities.model_entities import ModelTypefrom core.model_runtime.entities.rerank_entities import RerankResultfrom core.model_runtime.entities.text_embedding_entities import TextEmbeddingResultfrom core.model_runtime.model_providers.__base.large_language_model import LargeLanguageModelfrom core.model_runtime.model_providers.__base.moderation_model import ModerationModelfrom core.model_runtime.model_providers.__base.rerank_model import RerankModelfrom core.model_runtime.model_providers.__base.speech2text_model import Speech2TextModelfrom core.model_runtime.model_providers.__base.text_embedding_model import TextEmbeddingModelfrom core.provider_manager import ProviderManagerclass ModelInstance:    """    Model instance class    """    def __init__(self, provider_model_bundle: ProviderModelBundle, model: str) -> None:        self._provider_model_bundle = provider_model_bundle        self.model = model        self.provider = provider_model_bundle.configuration.provider.provider        self.credentials = self._fetch_credentials_from_bundle(provider_model_bundle, model)        self.model_type_instance = self._provider_model_bundle.model_type_instance    def _fetch_credentials_from_bundle(self, provider_model_bundle: ProviderModelBundle, model: str) -> dict:        """        Fetch credentials from provider model bundle        :param provider_model_bundle: provider model bundle        :param model: model name        :return:        """        credentials = provider_model_bundle.configuration.get_current_credentials(            model_type=provider_model_bundle.model_type_instance.model_type,            model=model        )        if credentials is None:            raise ProviderTokenNotInitError(f"Model {model} credentials is not initialized.")        return credentials    def invoke_llm(self, prompt_messages: list[PromptMessage], model_parameters: Optional[dict] = None,                   tools: Optional[list[PromptMessageTool]] = None, stop: Optional[List[str]] = None,                   stream: bool = True, user: Optional[str] = None, callbacks: list[Callback] = None) \            -> Union[LLMResult, Generator]:        """        Invoke large language model        :param prompt_messages: prompt messages        :param model_parameters: model parameters        :param tools: tools for tool calling        :param stop: stop words        :param stream: is stream response        :param user: unique user id        :param callbacks: callbacks        :return: full response or stream response chunk generator result        """        if not isinstance(self.model_type_instance, LargeLanguageModel):            raise Exception(f"Model type instance is not LargeLanguageModel")        self.model_type_instance = cast(LargeLanguageModel, self.model_type_instance)        return self.model_type_instance.invoke(            model=self.model,            credentials=self.credentials,            prompt_messages=prompt_messages,            model_parameters=model_parameters,            tools=tools,            stop=stop,            stream=stream,            user=user,            callbacks=callbacks        )    def invoke_text_embedding(self, texts: list[str], user: Optional[str] = None) \            -> TextEmbeddingResult:        """        Invoke large language model        :param texts: texts to embed        :param user: unique user id        :return: embeddings result        """        if not isinstance(self.model_type_instance, TextEmbeddingModel):            raise Exception(f"Model type instance is not TextEmbeddingModel")        self.model_type_instance = cast(TextEmbeddingModel, self.model_type_instance)        return self.model_type_instance.invoke(            model=self.model,            credentials=self.credentials,            texts=texts,            user=user        )    def invoke_rerank(self, query: str, docs: list[str], score_threshold: Optional[float] = None, top_n: Optional[int] = None,                      user: Optional[str] = None) \            -> RerankResult:        """        Invoke rerank model        :param query: search query        :param docs: docs for reranking        :param score_threshold: score threshold        :param top_n: top n        :param user: unique user id        :return: rerank result        """        if not isinstance(self.model_type_instance, RerankModel):            raise Exception(f"Model type instance is not RerankModel")        self.model_type_instance = cast(RerankModel, self.model_type_instance)        return self.model_type_instance.invoke(            model=self.model,            credentials=self.credentials,            query=query,            docs=docs,            score_threshold=score_threshold,            top_n=top_n,            user=user        )    def invoke_moderation(self, text: str, user: Optional[str] = None) \            -> bool:        """        Invoke moderation model        :param text: text to moderate        :param user: unique user id        :return: false if text is safe, true otherwise        """        if not isinstance(self.model_type_instance, ModerationModel):            raise Exception(f"Model type instance is not ModerationModel")        self.model_type_instance = cast(ModerationModel, self.model_type_instance)        return self.model_type_instance.invoke(            model=self.model,            credentials=self.credentials,            text=text,            user=user        )    def invoke_speech2text(self, file: IO[bytes], user: Optional[str] = None, **params) \            -> str:        """        Invoke large language model        :param file: audio file        :param user: unique user id        :return: text for given audio file        """        if not isinstance(self.model_type_instance, Speech2TextModel):            raise Exception(f"Model type instance is not Speech2TextModel")        self.model_type_instance = cast(Speech2TextModel, self.model_type_instance)        return self.model_type_instance.invoke(            model=self.model,            credentials=self.credentials,            file=file,            user=user,            **params        )class ModelManager:    def __init__(self) -> None:        self._provider_manager = ProviderManager()    def get_model_instance(self, tenant_id: str, provider: str, model_type: ModelType, model: str) -> ModelInstance:        """        Get model instance        :param tenant_id: tenant id        :param provider: provider name        :param model_type: model type        :param model: model name        :return:        """        if not provider:            return self.get_default_model_instance(tenant_id, model_type)        provider_model_bundle = self._provider_manager.get_provider_model_bundle(            tenant_id=tenant_id,            provider=provider,            model_type=model_type        )        return ModelInstance(provider_model_bundle, model)    def get_default_model_instance(self, tenant_id: str, model_type: ModelType) -> ModelInstance:        """        Get default model instance        :param tenant_id: tenant id        :param model_type: model type        :return:        """        default_model_entity = self._provider_manager.get_default_model(            tenant_id=tenant_id,            model_type=model_type        )        if not default_model_entity:            raise ProviderTokenNotInitError(f"Default model not found for {model_type}")        return self.get_model_instance(            tenant_id=tenant_id,            provider=default_model_entity.provider.provider,            model_type=model_type,            model=default_model_entity.model        )
 |