add: 添加ollama调用AI chat逻辑
This commit is contained in:
@@ -27,7 +27,9 @@ class Ollama(Platform):
|
|||||||
headers = {}
|
headers = {}
|
||||||
if self.api_key is not None:
|
if self.api_key is not None:
|
||||||
headers['Authorization'] = self.api_key
|
headers['Authorization'] = self.api_key
|
||||||
|
plugin.log.debug(f"{json.dumps(req_body)}")
|
||||||
async with self.http.post(endpoint, headers=headers, data=json.dumps(req_body)) as response:
|
async with self.http.post(endpoint, headers=headers, data=json.dumps(req_body)) as response:
|
||||||
|
plugin.log.debug(f"响应内容:{response.status}, {response.json()}")
|
||||||
if response.status != 200:
|
if response.status != 200:
|
||||||
return ChatCompletion(
|
return ChatCompletion(
|
||||||
message={},
|
message={},
|
||||||
|
|||||||
@@ -131,6 +131,7 @@ async def generate_context_messages(plugin: Plugin, platform: Platform, evt: Mes
|
|||||||
else:
|
else:
|
||||||
event_context = await plugin.client.get_event_context(room_id=evt.room_id, event_id=evt.event_id,
|
event_context = await plugin.client.get_event_context(room_id=evt.room_id, event_id=evt.event_id,
|
||||||
limit=platform.max_context_messages * 2)
|
limit=platform.max_context_messages * 2)
|
||||||
|
plugin.log.debug(f"event_context: {event_context}")
|
||||||
previous_messages = iter(event_context.events_before)
|
previous_messages = iter(event_context.events_before)
|
||||||
for evt in previous_messages:
|
for evt in previous_messages:
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user