import requests from ollama import Client, ResponseError import tiktoken class AICore: modelMaxTokens = 128000 # 初始化 DeepSeek 使用的 Tokenizer (cl100k_base) encoder = tiktoken.get_encoding("cl100k_base") def __init__(self, model): #初始化ollama客户端 ollamaClient = Client(host='http://localhost:11434', headers={'x-some-header': 'some-value'}) response = ollamaClient.show(model) modelMaxTokens = response.modelinfo['qwen2.context_length'] def getPromptToken(self, prompt)-> int: tokens = self.encoder.encode(prompt) return len(tokens)