mirror of
https://github.com/Jermolene/TiddlyWiki5
synced 2025-02-02 12:19:11 +00:00
Refactor completion servers so that they handle their own response
This commit is contained in:
parent
fb641d340c
commit
370ff3057e
@ -14,35 +14,11 @@ statusTitle - Optional title of a tiddler to which the status of the request wil
|
||||
completionServer - Optional URL of server
|
||||
-->
|
||||
\procedure get-llm-completion(conversationTitle,resultTitlePrefix,resultTags,statusTitle,completionServer)
|
||||
<!--
|
||||
Callback for the HTTP response from the LLM
|
||||
-->
|
||||
\procedure get-llm-completion-callback()
|
||||
<%if [<status>compare:number:gteq[200]compare:number:lteq[299]] %>
|
||||
<!-- Success -->
|
||||
<$action-createtiddler
|
||||
$basetitle=<<resultTitlePrefix>>
|
||||
tags=<<resultTags>>
|
||||
type="text/markdown"
|
||||
role={{{ [<data>jsonget[choices],[0],[message],[role]] }}}
|
||||
text={{{ [<data>jsonget[choices],[0],[message],[content]] }}}
|
||||
/>
|
||||
<%else%>
|
||||
<!-- Error -->
|
||||
<$action-createtiddler
|
||||
$basetitle=<<resultTitlePrefix>>
|
||||
tags=<<resultTags>>
|
||||
type="text/markdown"
|
||||
role="error"
|
||||
text={{{ [[Error:]] [<statusText>] [<data>jsonget[error],[message]] +[join[]] }}}
|
||||
/>
|
||||
<%endif%>
|
||||
\end get-llm-completion-callback
|
||||
|
||||
<$let
|
||||
completionServer={{{ [<completionServer>!is[blank]else<default-llm-completion-server>] }}}
|
||||
>
|
||||
<$wikify name="json" text={{{ [<completionServer>get[text]] }}}>
|
||||
<$importvariables filter="[<completionServer>]">
|
||||
<$wikify name="json" text=<<json-prompt>>>
|
||||
<$action-log message="get-llm-completion"/>
|
||||
<$action-log/>
|
||||
<$action-sendmessage
|
||||
@ -52,11 +28,12 @@ completionServer - Optional URL of server
|
||||
header-content-type="application/json"
|
||||
bearer-auth-token-from-store="openai-secret-key"
|
||||
method="POST"
|
||||
oncompletion=<<get-llm-completion-callback>>
|
||||
oncompletion=<<completion-callback>>
|
||||
bind-status=<<statusTitle>>
|
||||
var-resultTitlePrefix=<<resultTitlePrefix>>
|
||||
var-resultTags=<<resultTags>>
|
||||
/>
|
||||
</$wikify>
|
||||
</$importvariables>
|
||||
</$let>
|
||||
\end get-llm-completion
|
||||
|
@ -3,6 +3,10 @@ tags: $:/tags/AI/CompletionServer
|
||||
url: http://127.0.0.1:8080/v1/chat/completions
|
||||
caption: Locally running Llamafile server
|
||||
|
||||
<!--
|
||||
Wikified JSON text to be sent to server
|
||||
-->
|
||||
\procedure json-prompt()
|
||||
\rules only filteredtranscludeinline transcludeinline macrodef macrocallinline html conditional commentblock commentinline
|
||||
{
|
||||
"model": "gpt-4o",
|
||||
@ -22,3 +26,29 @@ caption: Locally running Llamafile server
|
||||
</$list>
|
||||
]
|
||||
}
|
||||
\end json-prompt
|
||||
|
||||
<!--
|
||||
Callback for the HTTP response from the LLM
|
||||
-->
|
||||
\procedure completion-callback()
|
||||
<%if [<status>compare:number:gteq[200]compare:number:lteq[299]] %>
|
||||
<!-- Success -->
|
||||
<$action-createtiddler
|
||||
$basetitle=<<resultTitlePrefix>>
|
||||
tags=<<resultTags>>
|
||||
type="text/markdown"
|
||||
role={{{ [<data>jsonget[choices],[0],[message],[role]] }}}
|
||||
text={{{ [<data>jsonget[choices],[0],[message],[content]] }}}
|
||||
/>
|
||||
<%else%>
|
||||
<!-- Error -->
|
||||
<$action-createtiddler
|
||||
$basetitle=<<resultTitlePrefix>>
|
||||
tags=<<resultTags>>
|
||||
type="text/markdown"
|
||||
role="error"
|
||||
text={{{ [[Error:]] [<statusText>] [<data>jsonget[error],[message]] +[join[]] }}}
|
||||
/>
|
||||
<%endif%>
|
||||
\end completion-callback
|
||||
|
@ -3,6 +3,10 @@ tags: $:/tags/AI/CompletionServer
|
||||
url: https://api.openai.com/v1/chat/completions
|
||||
caption: OpenAI Service
|
||||
|
||||
<!--
|
||||
Wikified JSON text to be sent to server
|
||||
-->
|
||||
\procedure json-prompt()
|
||||
\rules only filteredtranscludeinline transcludeinline macrodef macrocallinline html conditional commentblock commentinline
|
||||
{
|
||||
"model": "gpt-4o",
|
||||
@ -37,3 +41,29 @@ caption: OpenAI Service
|
||||
</$list>
|
||||
]
|
||||
}
|
||||
\end json-prompt
|
||||
|
||||
<!--
|
||||
Callback for the HTTP response from the LLM
|
||||
-->
|
||||
\procedure completion-callback()
|
||||
<%if [<status>compare:number:gteq[200]compare:number:lteq[299]] %>
|
||||
<!-- Success -->
|
||||
<$action-createtiddler
|
||||
$basetitle=<<resultTitlePrefix>>
|
||||
tags=<<resultTags>>
|
||||
type="text/markdown"
|
||||
role={{{ [<data>jsonget[choices],[0],[message],[role]] }}}
|
||||
text={{{ [<data>jsonget[choices],[0],[message],[content]] }}}
|
||||
/>
|
||||
<%else%>
|
||||
<!-- Error -->
|
||||
<$action-createtiddler
|
||||
$basetitle=<<resultTitlePrefix>>
|
||||
tags=<<resultTags>>
|
||||
type="text/markdown"
|
||||
role="error"
|
||||
text={{{ [[Error:]] [<statusText>] [<data>jsonget[error],[message]] +[join[]] }}}
|
||||
/>
|
||||
<%endif%>
|
||||
\end completion-callback
|
||||
|
Loading…
Reference in New Issue
Block a user