Project02/AIGC/AICore.py

21 lines
643 B
Python

import requests
from ollama import Client, ResponseError
import tiktoken
class AICore:
modelMaxTokens = 128000
# 初始化 DeepSeek 使用的 Tokenizer (cl100k_base)
encoder = tiktoken.get_encoding("cl100k_base")
def __init__(self, model):
#初始化ollama客户端
ollamaClient = Client(host='http://localhost:11434', headers={'x-some-header': 'some-value'})
response = ollamaClient.show(model)
modelMaxTokens = response.modelinfo['qwen2.context_length']
def getPromptToken(self, prompt)-> int:
tokens = self.encoder.encode(prompt)
return len(tokens)