llm_and_rule_mix.py•1.12 kB
import os
from dingo.config import InputArgs
from dingo.exec import Executor
if __name__ == '__main__':
OPENAI_MODEL = 'deepseek-chat'
OPENAI_URL = 'https://api.deepseek.com/v1'
OPENAI_KEY = os.getenv("OPENAI_KEY")
input_data = {
"input_path": "../../test/data/test_local_jsonl.jsonl",
"dataset": {
"source": "local",
"format": "jsonl",
"field": {
"content": "content"
}
},
"executor": {
"rule_list": ["RuleColonEnd"],
"prompt_list": ["PromptRepeat"],
"result_save": {
"bad": True,
"good": True
}
},
"evaluator": {
"llm_config": {
"LLMTextQualityPromptBase": {
"model": OPENAI_MODEL,
"key": OPENAI_KEY,
"api_url": OPENAI_URL,
}
}
}
}
input_args = InputArgs(**input_data)
executor = Executor.exec_map["local"](input_args)
result = executor.execute()
print(result)