bugfix
This commit is contained in:
parent
0a42aa88f0
commit
d4ff002c16
@ -1,4 +1,5 @@
|
||||
from llmage.llmclient import (
|
||||
b64media.
|
||||
get_llm,
|
||||
inference,
|
||||
get_llmcatelogs,
|
||||
@ -16,6 +17,7 @@ from ahserver.serverenv import ServerEnv
|
||||
def load_llmage():
|
||||
env = ServerEnv()
|
||||
env.get_llm = get_llm
|
||||
env.b64media = b64media
|
||||
env.inference = inference
|
||||
env.get_llms_by_catelog = get_llms_by_catelog
|
||||
env.get_llmcatelogs = get_llmcatelogs
|
||||
|
||||
@ -83,6 +83,15 @@ async def uapi_request(request, sor, caller_orgid, callerid, uapi, llm, params):
|
||||
yield l
|
||||
debug(f'{d=}, {txt=}')
|
||||
|
||||
def b64media(meidafile):
|
||||
if meidafile.startswith('data:'):
|
||||
return meidafile
|
||||
fs = FileStorage()
|
||||
fn = fs.realPath(meidafile)
|
||||
with open(fn, 'rb') as f:
|
||||
b = f.read()
|
||||
return base64.b64encode(b).decode('iso-8859-1')
|
||||
|
||||
async def inference(request, *args, **kw):
|
||||
env = request._run_ns
|
||||
caller_orgid = await env.get_userorgid()
|
||||
|
||||
@ -28,7 +28,7 @@
|
||||
{
|
||||
"widgettype":"Text",
|
||||
"options":{
|
||||
"otype": "没找到模型",
|
||||
"otext": "没找到模型",
|
||||
"i18n":true
|
||||
}
|
||||
}
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user