Add huggingface back-end

defaults to mistral 7b instruct
This commit is contained in:
Jeff Epler 2023-09-29 10:13:12 -05:00
parent 2c04964b93
commit b6fa44f53e
No known key found for this signature in database
GPG key ID: D5BF15AB975AB4DE

View file

@ -0,0 +1,117 @@
# SPDX-FileCopyrightText: 2023 Jeff Epler <jepler@gmail.com>
#
# SPDX-License-Identifier: MIT
import asyncio
import json
from dataclasses import dataclass
import httpx
from ..key import get_key
from ..session import Assistant, User
class HuggingFace:
@dataclass
class Parameters:
url: str = "https://api-inference.huggingface.co"
model: str = "mistralai/Mistral-7B-Instruct-v0.1"
max_new_tokens: int = 250
start_prompt: str = """<s>[INST] <<SYS>>\n"""
after_system: str = "\n<</SYS>>\n\n"
after_user: str = """ [/INST] """
after_assistant: str = """ </s><s>[INST] """
stop_token_id = 2
def __init__(self):
self.parameters = self.Parameters()
system_message = """\
A dialog, where USER interacts with AI. AI is helpful, kind, obedient, honest, and knows its own limits.
"""
def make_full_query(self, messages, max_query_size):
del messages[1:-max_query_size]
result = [self.parameters.start_prompt]
for m in messages:
content = (m.content or "").strip()
if not content:
continue
result.append(content)
if m.role == "system":
result.append(self.parameters.after_system)
elif m.role == "assistant":
result.append(self.parameters.after_assistant)
elif m.role == "user":
result.append(self.parameters.after_user)
full_query = "".join(result)
return full_query
async def chained_query(self, inputs, timeout):
async with httpx.AsyncClient(timeout=timeout) as client:
while inputs:
params = {
"inputs": inputs,
"stream": True,
}
inputs = None
async with client.stream(
"POST",
f"{self.parameters.url}/models/{self.parameters.model}",
json=params,
headers={
"Authorization": f"Bearer {self.get_key()}",
},
) as response:
if response.status_code == 200:
async for line in response.aiter_lines():
if line.startswith("data:"):
data = line.removeprefix("data:").strip()
j = json.loads(data)
token = j.get("token", {})
inputs = j.get("generated_text", inputs)
if token.get("id") == self.parameters.stop_token_id:
return
yield token.get("text", "")
else:
yield f"\nFailed with {response=!r}"
return
async def aask(
self, session, query, *, max_query_size=5, timeout=180
): # pylint: disable=unused-argument,too-many-locals,too-many-branches
new_content = []
inputs = self.make_full_query(session.session + [User(query)], max_query_size)
try:
async for content in self.chained_query(inputs, timeout=timeout):
if not new_content:
content = content.lstrip()
if content:
if not new_content:
content = content.lstrip()
if content:
new_content.append(content)
yield content
except httpx.HTTPError as e:
content = f"\nException: {e!r}"
new_content.append(content)
yield content
session.session.extend([User(query), Assistant("".join(new_content))])
def ask(self, session, query, *, max_query_size=5, timeout=60):
asyncio.run(
self.aask(session, query, max_query_size=max_query_size, timeout=timeout)
)
return session.session[-1].message
@classmethod
def get_key(cls):
return get_key("huggingface_api_token")
def factory():
"""Uses the huggingface text-generation-interface web API"""
return HuggingFace()