| 1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950 | 
							- import logging
 
- from core.app.app_config.entities import AppConfig
 
- from core.moderation.base import ModerationAction, ModerationException
 
- from core.moderation.factory import ModerationFactory
 
- logger = logging.getLogger(__name__)
 
- class InputModeration:
 
-     def check(self, app_id: str,
 
-               tenant_id: str,
 
-               app_config: AppConfig,
 
-               inputs: dict,
 
-               query: str) -> tuple[bool, dict, str]:
 
-         """
 
-         Process sensitive_word_avoidance.
 
-         :param app_id: app id
 
-         :param tenant_id: tenant id
 
-         :param app_config: app config
 
-         :param inputs: inputs
 
-         :param query: query
 
-         :return:
 
-         """
 
-         if not app_config.sensitive_word_avoidance:
 
-             return False, inputs, query
 
-         sensitive_word_avoidance_config = app_config.sensitive_word_avoidance
 
-         moderation_type = sensitive_word_avoidance_config.type
 
-         moderation_factory = ModerationFactory(
 
-             name=moderation_type,
 
-             app_id=app_id,
 
-             tenant_id=tenant_id,
 
-             config=sensitive_word_avoidance_config.config
 
-         )
 
-         moderation_result = moderation_factory.moderation_for_inputs(inputs, query)
 
-         if not moderation_result.flagged:
 
-             return False, inputs, query
 
-         if moderation_result.action == ModerationAction.DIRECT_OUTPUT:
 
-             raise ModerationException(moderation_result.preset_response)
 
-         elif moderation_result.action == ModerationAction.OVERRIDED:
 
-             inputs = moderation_result.inputs
 
-             query = moderation_result.query
 
-         return True, inputs, query
 
 
  |