21 lines
653 B
Python
21 lines
653 B
Python
import requests
|
|
from ollama import Client, ResponseError
|
|
import tiktoken
|
|
|
|
class AICore:
|
|
modelMaxTokens = 128000
|
|
# 初始化 DeepSeek 使用的 Tokenizer (cl100k_base)
|
|
encoder = tiktoken.get_encoding("cl100k_base")
|
|
|
|
|
|
|
|
def __init__(self, model):
|
|
#初始化ollama客户端
|
|
self.ollamaClient = Client(host='http://localhost:11434', headers={'x-some-header': 'some-value'})
|
|
response = self.ollamaClient.show(model)
|
|
modelMaxTokens = response.modelinfo['qwen2.context_length']
|
|
|
|
def getPromptToken(self, prompt)-> int:
|
|
tokens = self.encoder.encode(prompt)
|
|
return len(tokens)
|
|
|