Compare commits
50 Commits
feat/kokor
...
feat/cli
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
326a1a4d09 | ||
|
|
0b7f621a75 | ||
|
|
46dbb7b981 | ||
|
|
ef0940daa8 | ||
|
|
3d44686a51 | ||
|
|
df04d8c21c | ||
|
|
7c8697f48e | ||
|
|
7f8bbefb05 | ||
|
|
2c9c36e2c6 | ||
|
|
e476575334 | ||
|
|
7e346b5e19 | ||
|
|
1396b3eb05 | ||
|
|
619b19cb46 | ||
|
|
26377702d3 | ||
|
|
fdcaa6c5e2 | ||
|
|
77c365959d | ||
|
|
f4fcb85570 | ||
|
|
92acfb7ed4 | ||
|
|
2901208c80 | ||
|
|
4cfe2fe37f | ||
|
|
648035b194 | ||
|
|
13773bcc97 | ||
|
|
d9b820c9c4 | ||
|
|
adc4dea644 | ||
|
|
6456cb0922 | ||
|
|
e77dbb3160 | ||
|
|
3013f21a75 | ||
|
|
bb57be95c6 | ||
|
|
528d4210fc | ||
|
|
54b4cccf92 | ||
|
|
c2c90f6d2b | ||
|
|
94769225cf | ||
|
|
0e42a6f069 | ||
|
|
a1b5f9cdc5 | ||
|
|
e74ff8c03f | ||
|
|
b6e802c12e | ||
|
|
c0d5db29a5 | ||
|
|
6ed96c9bd3 | ||
|
|
b5f0eabeea | ||
|
|
e0201886f8 | ||
|
|
5b175c12a6 | ||
|
|
c200c9328c | ||
|
|
23cb8f2578 | ||
|
|
4f0bce50c5 | ||
|
|
bf655a1087 | ||
|
|
c8f00198d6 | ||
|
|
c5a24b2a3f | ||
|
|
0f0c43f327 | ||
|
|
0e55e44f62 | ||
|
|
2687f38d00 |
2
.gitignore
vendored
2
.gitignore
vendored
@@ -1,4 +1,3 @@
|
|||||||
*.txt
|
|
||||||
*.json
|
*.json
|
||||||
testlog
|
testlog
|
||||||
history/
|
history/
|
||||||
@@ -18,3 +17,4 @@ chat_exports/*.json
|
|||||||
ragimport
|
ragimport
|
||||||
.env
|
.env
|
||||||
onnx/
|
onnx/
|
||||||
|
*.log
|
||||||
|
|||||||
37
Makefile
37
Makefile
@@ -1,4 +1,4 @@
|
|||||||
.PHONY: setconfig run lint lintall install-linters setup-whisper build-whisper download-whisper-model docker-up docker-down docker-logs noextra-run installdelve checkdelve fetch-onnx install-onnx-deps fetch-kokoro-voices install-espeak
|
.PHONY: setconfig run lint lintall install-linters setup-whisper build-whisper download-whisper-model docker-up docker-down docker-logs noextra-run installdelve checkdelve fetch-onnx install-onnx-deps
|
||||||
|
|
||||||
run: setconfig
|
run: setconfig
|
||||||
go build -tags extra -o gf-lt && ./gf-lt
|
go build -tags extra -o gf-lt && ./gf-lt
|
||||||
@@ -33,9 +33,6 @@ lintall: lint
|
|||||||
fetch-onnx:
|
fetch-onnx:
|
||||||
mkdir -p onnx/embedgemma && curl -o onnx/embedgemma/config.json -L https://huggingface.co/onnx-community/embeddinggemma-300m-ONNX/resolve/main/config.json && curl -o onnx/embedgemma/tokenizer.json -L https://huggingface.co/onnx-community/embeddinggemma-300m-ONNX/resolve/main/tokenizer.json && curl -o onnx/embedgemma/model_q4.onnx -L https://huggingface.co/onnx-community/embeddinggemma-300m-ONNX/resolve/main/onnx/model_q4.onnx && curl -o onnx/embedgemma/model_q4.onnx_data -L https://huggingface.co/onnx-community/embeddinggemma-300m-ONNX/resolve/main/onnx/model_q4.onnx_data?download=true
|
mkdir -p onnx/embedgemma && curl -o onnx/embedgemma/config.json -L https://huggingface.co/onnx-community/embeddinggemma-300m-ONNX/resolve/main/config.json && curl -o onnx/embedgemma/tokenizer.json -L https://huggingface.co/onnx-community/embeddinggemma-300m-ONNX/resolve/main/tokenizer.json && curl -o onnx/embedgemma/model_q4.onnx -L https://huggingface.co/onnx-community/embeddinggemma-300m-ONNX/resolve/main/onnx/model_q4.onnx && curl -o onnx/embedgemma/model_q4.onnx_data -L https://huggingface.co/onnx-community/embeddinggemma-300m-ONNX/resolve/main/onnx/model_q4.onnx_data?download=true
|
||||||
|
|
||||||
fetch-kokoro-onnx:
|
|
||||||
mkdir -p onnx/kokoro && curl -o onnx/kokoro/config.json -L https://huggingface.co/onnx-community/Kokoro-82M-v1.0-ONNX/resolve/main/config.json && curl -o onnx/kokoro/tokenizer.json -L https://huggingface.co/onnx-community/Kokoro-82M-v1.0-ONNX/resolve/main/tokenizer.json && curl -o onnx/kokoro/model_quantized.onnx -L https://huggingface.co/onnx-community/Kokoro-82M-v1.0-ONNX/resolve/main/onnx/model_quantized.onnx && curl -o onnx/kokoro/voices.bin -L https://github.com/thewh1teagle/kokoro-onnx/releases/download/model-files-v1.0/voices-v1.0.bin
|
|
||||||
|
|
||||||
install-onnx-deps: ## Install ONNX Runtime with CUDA support (or CPU fallback)
|
install-onnx-deps: ## Install ONNX Runtime with CUDA support (or CPU fallback)
|
||||||
@echo "=== ONNX Runtime Installer ===" && \
|
@echo "=== ONNX Runtime Installer ===" && \
|
||||||
echo "" && \
|
echo "" && \
|
||||||
@@ -137,8 +134,9 @@ setup-whisper: build-whisper download-whisper-model
|
|||||||
|
|
||||||
build-whisper: ## Build whisper.cpp from source in batteries directory
|
build-whisper: ## Build whisper.cpp from source in batteries directory
|
||||||
@echo "Building whisper.cpp from source in batteries directory..."
|
@echo "Building whisper.cpp from source in batteries directory..."
|
||||||
@if [ ! -d "batteries/whisper.cpp" ]; then \
|
@if [ ! -f "batteries/whisper.cpp/CMakeLists.txt" ]; then \
|
||||||
echo "Cloning whisper.cpp repository to batteries directory..."; \
|
echo "Cloning whisper.cpp repository to batteries directory..."; \
|
||||||
|
rm -rf batteries/whisper.cpp; \
|
||||||
git clone https://github.com/ggml-org/whisper.cpp.git batteries/whisper.cpp; \
|
git clone https://github.com/ggml-org/whisper.cpp.git batteries/whisper.cpp; \
|
||||||
fi
|
fi
|
||||||
cd batteries/whisper.cpp && cmake -B build -DGGML_CUDA=ON -DWHISPER_SDL2=ON; cmake --build build --config Release -j 8
|
cd batteries/whisper.cpp && cmake -B build -DGGML_CUDA=ON -DWHISPER_SDL2=ON; cmake --build build --config Release -j 8
|
||||||
@@ -146,11 +144,10 @@ build-whisper: ## Build whisper.cpp from source in batteries directory
|
|||||||
|
|
||||||
download-whisper-model: ## Download Whisper model for STT in batteries directory
|
download-whisper-model: ## Download Whisper model for STT in batteries directory
|
||||||
@echo "Downloading Whisper model for STT..."
|
@echo "Downloading Whisper model for STT..."
|
||||||
@if [ ! -d "batteries/whisper.cpp" ]; then \
|
@if [ ! -d "batteries/whisper.cpp/models" ]; then \
|
||||||
echo "Please run 'make setup-whisper' first to clone the repository."; \
|
mkdir -p "batteries/whisper.cpp/models"; \
|
||||||
exit 1; \
|
|
||||||
fi
|
fi
|
||||||
@cd batteries/whisper.cpp && bash ./models/download-ggml-model.sh large-v3-turbo-q5_0
|
curl -o batteries/whisper.cpp/models/ggml-large-v3-turbo-q5_0.bin -L "https://huggingface.co/ggerganov/whisper.cpp/resolve/main/ggml-large-v3-turbo-q5_0.bin?download=true"
|
||||||
@echo "Whisper model downloaded successfully!"
|
@echo "Whisper model downloaded successfully!"
|
||||||
|
|
||||||
# Docker targets for STT/TTS services (in batteries directory)
|
# Docker targets for STT/TTS services (in batteries directory)
|
||||||
@@ -197,25 +194,3 @@ docker-logs-whisper: ## View logs from Whisper STT service only
|
|||||||
docker-logs-kokoro: ## View logs from Kokoro TTS service only
|
docker-logs-kokoro: ## View logs from Kokoro TTS service only
|
||||||
@echo "Displaying logs from Kokoro TTS service..."
|
@echo "Displaying logs from Kokoro TTS service..."
|
||||||
docker-compose -f batteries/docker-compose.yml logs -f kokoro-tts
|
docker-compose -f batteries/docker-compose.yml logs -f kokoro-tts
|
||||||
|
|
||||||
# Kokoro ONNX TTS Setup
|
|
||||||
install-espeak: ## Install espeak-ng for phoneme tokenization
|
|
||||||
@echo "=== Installing espeak-ng ===" && \
|
|
||||||
if command -v espeak-ng >/dev/null 2>&1; then \
|
|
||||||
echo "espeak-ng is already installed:" && \
|
|
||||||
espeak-ng --version && \
|
|
||||||
exit 0; \
|
|
||||||
fi && \
|
|
||||||
echo "Installing espeak-ng..." && \
|
|
||||||
sudo apt-get update && \
|
|
||||||
sudo apt-get install -y espeak-ng espeak && \
|
|
||||||
echo "espeak-ng installed successfully!" && \
|
|
||||||
espeak-ng --version
|
|
||||||
|
|
||||||
fetch-kokoro-voices: ## Download Kokoro voice files (PyTorch format)
|
|
||||||
@echo "=== Downloading Kokoro voices ===" && \
|
|
||||||
mkdir -p onnx/kokoro/voices && \
|
|
||||||
echo "Downloading af_bella voice..." && \
|
|
||||||
curl -L -o onnx/kokoro/voices/af_bella.pt https://raw.githubusercontent.com/hexgrad/kokoro/main/kokoro/voices/af_heart.pt && \
|
|
||||||
echo "Voice file downloaded to onnx/kokoro/voices/" && \
|
|
||||||
ls -lh onnx/kokoro/voices/
|
|
||||||
|
|||||||
13
README.md
13
README.md
@@ -10,9 +10,19 @@ made with use of [tview](https://github.com/rivo/tview)
|
|||||||
- function calls (function calls are implemented natively, to avoid calling outside sources);
|
- function calls (function calls are implemented natively, to avoid calling outside sources);
|
||||||
- [character specific context (unique feature)](docs/char-specific-context.md)
|
- [character specific context (unique feature)](docs/char-specific-context.md)
|
||||||
|
|
||||||
|
|
||||||
|
#### showcase on youtube
|
||||||
|
[](https://youtu.be/WCS4Xc902F8 "gf-lt showcase")
|
||||||
|
|
||||||
#### how it looks
|
#### how it looks
|
||||||

|

|
||||||
|
|
||||||
|
|
||||||
|
#### dependencies
|
||||||
|
- make
|
||||||
|
- go
|
||||||
|
- ffmpeg (extra)
|
||||||
|
|
||||||
#### how to install
|
#### how to install
|
||||||
(requires golang)
|
(requires golang)
|
||||||
clone the project
|
clone the project
|
||||||
@@ -28,7 +38,8 @@ make noextra-run
|
|||||||
```
|
```
|
||||||
|
|
||||||
#### keybinds
|
#### keybinds
|
||||||
while running you can press f12 for list of keys;
|
- use `insert` button to paste text from the clipboard to the text area, instead of shift+insert (might freeze the program);
|
||||||
|
- press f12 for list of keys;
|
||||||

|

|
||||||
|
|
||||||
#### setting up config
|
#### setting up config
|
||||||
|
|||||||
@@ -4,11 +4,12 @@ package agent
|
|||||||
// ones who do their own tools calls
|
// ones who do their own tools calls
|
||||||
// ones that works only with the output
|
// ones that works only with the output
|
||||||
|
|
||||||
// A: main chat -> agent (handles everything: tool + processing)
|
// A: main chat -> agent (handles everything: tool + processing), supports tool chaining
|
||||||
// B: main chat -> tool -> agent (process tool output)
|
// B: main chat -> tool -> agent (process tool output)
|
||||||
|
|
||||||
// AgenterA gets a task "find out weather in london"
|
// AgenterA gets a task like "go to the webpage, login and take a screenshot (tell me what you see)"
|
||||||
// proceeds to make tool calls on its own
|
// proceeds to make a plan and executes it.
|
||||||
|
// returns with final result or an error
|
||||||
type AgenterA interface {
|
type AgenterA interface {
|
||||||
ProcessTask(task string) []byte
|
ProcessTask(task string) []byte
|
||||||
}
|
}
|
||||||
@@ -38,8 +39,3 @@ func RegisterA(toolNames []string, a AgenterA) {
|
|||||||
func Get(toolName string) AgenterB {
|
func Get(toolName string) AgenterB {
|
||||||
return RegistryB[toolName]
|
return RegistryB[toolName]
|
||||||
}
|
}
|
||||||
|
|
||||||
// Register is a convenience wrapper for RegisterB.
|
|
||||||
func Register(toolName string, a AgenterB) {
|
|
||||||
RegisterB(toolName, a)
|
|
||||||
}
|
|
||||||
|
|||||||
119
agent/pw_agent.go
Normal file
119
agent/pw_agent.go
Normal file
@@ -0,0 +1,119 @@
|
|||||||
|
package agent
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/json"
|
||||||
|
"gf-lt/models"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
// PWAgent: is AgenterA type agent (enclosed with tool chaining)
|
||||||
|
// sysprompt explain tools and how to plan for execution
|
||||||
|
type PWAgent struct {
|
||||||
|
*AgentClient
|
||||||
|
sysprompt string
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewPWAgent creates a PWAgent with the given client and system prompt
|
||||||
|
func NewPWAgent(client *AgentClient, sysprompt string) *PWAgent {
|
||||||
|
return &PWAgent{AgentClient: client, sysprompt: sysprompt}
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetTools sets the tools available to the agent
|
||||||
|
func (a *PWAgent) SetTools(tools []models.Tool) {
|
||||||
|
a.tools = tools
|
||||||
|
}
|
||||||
|
|
||||||
|
func (a *PWAgent) ProcessTask(task string) []byte {
|
||||||
|
req, err := a.FormFirstMsg(a.sysprompt, task)
|
||||||
|
if err != nil {
|
||||||
|
a.Log().Error("PWAgent failed to process the request", "error", err)
|
||||||
|
return []byte("PWAgent failed to process the request; err: " + err.Error())
|
||||||
|
}
|
||||||
|
toolCallLimit := 10
|
||||||
|
for i := 0; i < toolCallLimit; i++ {
|
||||||
|
resp, err := a.LLMRequest(req)
|
||||||
|
if err != nil {
|
||||||
|
a.Log().Error("failed to process the request", "error", err)
|
||||||
|
return []byte("failed to process the request; err: " + err.Error())
|
||||||
|
}
|
||||||
|
execTool, toolCallID, hasToolCall := findToolCall(resp)
|
||||||
|
if !hasToolCall {
|
||||||
|
return resp
|
||||||
|
}
|
||||||
|
|
||||||
|
a.setToolCallOnLastMessage(resp, toolCallID)
|
||||||
|
|
||||||
|
toolResp := string(execTool())
|
||||||
|
req, err = a.FormMsgWithToolCallID(toolResp, toolCallID)
|
||||||
|
if err != nil {
|
||||||
|
a.Log().Error("failed to form next message", "error", err)
|
||||||
|
return []byte("failed to form next message; err: " + err.Error())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (a *PWAgent) setToolCallOnLastMessage(resp []byte, toolCallID string) {
|
||||||
|
if toolCallID == "" {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
var genericResp map[string]interface{}
|
||||||
|
if err := json.Unmarshal(resp, &genericResp); err != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
var name string
|
||||||
|
var args map[string]string
|
||||||
|
if choices, ok := genericResp["choices"].([]interface{}); ok && len(choices) > 0 {
|
||||||
|
if firstChoice, ok := choices[0].(map[string]interface{}); ok {
|
||||||
|
if message, ok := firstChoice["message"].(map[string]interface{}); ok {
|
||||||
|
if toolCalls, ok := message["tool_calls"].([]interface{}); ok && len(toolCalls) > 0 {
|
||||||
|
if tc, ok := toolCalls[0].(map[string]interface{}); ok {
|
||||||
|
if fn, ok := tc["function"].(map[string]interface{}); ok {
|
||||||
|
name, _ = fn["name"].(string)
|
||||||
|
argsStr, _ := fn["arguments"].(string)
|
||||||
|
_ = json.Unmarshal([]byte(argsStr), &args)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if name == "" {
|
||||||
|
content, _ := genericResp["content"].(string)
|
||||||
|
name = extractToolNameFromText(content)
|
||||||
|
}
|
||||||
|
lastIdx := len(a.chatBody.Messages) - 1
|
||||||
|
if lastIdx >= 0 {
|
||||||
|
a.chatBody.Messages[lastIdx].ToolCallID = toolCallID
|
||||||
|
if name != "" {
|
||||||
|
argsJSON, _ := json.Marshal(args)
|
||||||
|
a.chatBody.Messages[lastIdx].ToolCall = &models.ToolCall{
|
||||||
|
ID: toolCallID,
|
||||||
|
Name: name,
|
||||||
|
Args: string(argsJSON),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func extractToolNameFromText(text string) string {
|
||||||
|
jsStr := toolCallRE.FindString(text)
|
||||||
|
if jsStr == "" {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
jsStr = strings.TrimSpace(jsStr)
|
||||||
|
jsStr = strings.TrimPrefix(jsStr, "__tool_call__")
|
||||||
|
jsStr = strings.TrimSuffix(jsStr, "__tool_call__")
|
||||||
|
jsStr = strings.TrimSpace(jsStr)
|
||||||
|
start := strings.Index(jsStr, "{")
|
||||||
|
end := strings.LastIndex(jsStr, "}")
|
||||||
|
if start == -1 || end == -1 || end <= start {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
jsStr = jsStr[start : end+1]
|
||||||
|
var fc models.FuncCall
|
||||||
|
if err := json.Unmarshal([]byte(jsStr), &fc); err != nil {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
return fc.Name
|
||||||
|
}
|
||||||
338
agent/pw_tools.go
Normal file
338
agent/pw_tools.go
Normal file
@@ -0,0 +1,338 @@
|
|||||||
|
package agent
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
"regexp"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"gf-lt/models"
|
||||||
|
)
|
||||||
|
|
||||||
|
type ToolFunc func(map[string]string) []byte
|
||||||
|
|
||||||
|
var pwToolMap = make(map[string]ToolFunc)
|
||||||
|
|
||||||
|
func RegisterPWTool(name string, fn ToolFunc) {
|
||||||
|
pwToolMap[name] = fn
|
||||||
|
}
|
||||||
|
|
||||||
|
func GetPWTools() []models.Tool {
|
||||||
|
return pwTools
|
||||||
|
}
|
||||||
|
|
||||||
|
var pwTools = []models.Tool{
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_start",
|
||||||
|
Description: "Start a Playwright browser instance. Must be called first before any other browser automation. Uses headless mode by default.",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{},
|
||||||
|
Properties: map[string]models.ToolArgProps{},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_stop",
|
||||||
|
Description: "Stop the Playwright browser instance. Call when done with browser automation.",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{},
|
||||||
|
Properties: map[string]models.ToolArgProps{},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_is_running",
|
||||||
|
Description: "Check if Playwright browser is currently running.",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{},
|
||||||
|
Properties: map[string]models.ToolArgProps{},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_navigate",
|
||||||
|
Description: "Navigate to a URL in the browser.",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{"url"},
|
||||||
|
Properties: map[string]models.ToolArgProps{
|
||||||
|
"url": {Type: "string", Description: "URL to navigate to"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_click",
|
||||||
|
Description: "Click on an element on the current webpage. Use 'index' for multiple matches (default 0).",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{"selector"},
|
||||||
|
Properties: map[string]models.ToolArgProps{
|
||||||
|
"selector": {Type: "string", Description: "CSS selector for the element"},
|
||||||
|
"index": {Type: "integer", Description: "Index for multiple matches (default 0)"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_fill",
|
||||||
|
Description: "Type text into an input field. Use 'index' for multiple matches (default 0).",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{"selector", "text"},
|
||||||
|
Properties: map[string]models.ToolArgProps{
|
||||||
|
"selector": {Type: "string", Description: "CSS selector for the input element"},
|
||||||
|
"text": {Type: "string", Description: "Text to type into the field"},
|
||||||
|
"index": {Type: "integer", Description: "Index for multiple matches (default 0)"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_extract_text",
|
||||||
|
Description: "Extract text content from the page or specific elements. Use selector 'body' for all page text.",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{},
|
||||||
|
Properties: map[string]models.ToolArgProps{
|
||||||
|
"selector": {Type: "string", Description: "CSS selector (default 'body' for all page text)"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_screenshot",
|
||||||
|
Description: "Take a screenshot of the page or a specific element. Returns a file path to the image.",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{},
|
||||||
|
Properties: map[string]models.ToolArgProps{
|
||||||
|
"selector": {Type: "string", Description: "CSS selector for element to screenshot"},
|
||||||
|
"full_page": {Type: "boolean", Description: "Capture full page (default false)"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_screenshot_and_view",
|
||||||
|
Description: "Take a screenshot and return the image for viewing. Use to visually verify page state.",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{},
|
||||||
|
Properties: map[string]models.ToolArgProps{
|
||||||
|
"selector": {Type: "string", Description: "CSS selector for element to screenshot"},
|
||||||
|
"full_page": {Type: "boolean", Description: "Capture full page (default false)"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_wait_for_selector",
|
||||||
|
Description: "Wait for an element to appear on the page before proceeding.",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{"selector"},
|
||||||
|
Properties: map[string]models.ToolArgProps{
|
||||||
|
"selector": {Type: "string", Description: "CSS selector to wait for"},
|
||||||
|
"timeout": {Type: "integer", Description: "Timeout in milliseconds (default 30000)"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_drag",
|
||||||
|
Description: "Drag the mouse from point (x1,y1) to (x2,y2).",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{"x1", "y1", "x2", "y2"},
|
||||||
|
Properties: map[string]models.ToolArgProps{
|
||||||
|
"x1": {Type: "number", Description: "Starting X coordinate"},
|
||||||
|
"y1": {Type: "number", Description: "Starting Y coordinate"},
|
||||||
|
"x2": {Type: "number", Description: "Ending X coordinate"},
|
||||||
|
"y2": {Type: "number", Description: "Ending Y coordinate"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_click_at",
|
||||||
|
Description: "Click at specific X,Y coordinates on the page.",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{"x", "y"},
|
||||||
|
Properties: map[string]models.ToolArgProps{
|
||||||
|
"x": {Type: "number", Description: "X coordinate"},
|
||||||
|
"y": {Type: "number", Description: "Y coordinate"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_get_html",
|
||||||
|
Description: "Get the HTML content of the page or a specific element.",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{},
|
||||||
|
Properties: map[string]models.ToolArgProps{
|
||||||
|
"selector": {Type: "string", Description: "CSS selector (default 'body')"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_get_dom",
|
||||||
|
Description: "Get a structured DOM representation with tag, attributes, text, and children.",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{},
|
||||||
|
Properties: map[string]models.ToolArgProps{
|
||||||
|
"selector": {Type: "string", Description: "CSS selector (default 'body')"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "pw_search_elements",
|
||||||
|
Description: "Search for elements by text content or CSS selector.",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{},
|
||||||
|
Properties: map[string]models.ToolArgProps{
|
||||||
|
"text": {Type: "string", Description: "Text content to search for"},
|
||||||
|
"selector": {Type: "string", Description: "CSS selector to search for"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
var toolCallRE = regexp.MustCompile(`__tool_call__(.+?)__tool_call__`)
|
||||||
|
|
||||||
|
type ParsedToolCall struct {
|
||||||
|
ID string
|
||||||
|
Name string
|
||||||
|
Args map[string]string
|
||||||
|
}
|
||||||
|
|
||||||
|
func findToolCall(resp []byte) (func() []byte, string, bool) {
|
||||||
|
var genericResp map[string]interface{}
|
||||||
|
if err := json.Unmarshal(resp, &genericResp); err != nil {
|
||||||
|
return findToolCallFromText(string(resp))
|
||||||
|
}
|
||||||
|
if choices, ok := genericResp["choices"].([]interface{}); ok && len(choices) > 0 {
|
||||||
|
if firstChoice, ok := choices[0].(map[string]interface{}); ok {
|
||||||
|
if message, ok := firstChoice["message"].(map[string]interface{}); ok {
|
||||||
|
if toolCalls, ok := message["tool_calls"].([]interface{}); ok && len(toolCalls) > 0 {
|
||||||
|
return parseOpenAIToolCall(toolCalls)
|
||||||
|
}
|
||||||
|
if content, ok := message["content"].(string); ok {
|
||||||
|
return findToolCallFromText(content)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if text, ok := firstChoice["text"].(string); ok {
|
||||||
|
return findToolCallFromText(text)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if content, ok := genericResp["content"].(string); ok {
|
||||||
|
return findToolCallFromText(content)
|
||||||
|
}
|
||||||
|
return findToolCallFromText(string(resp))
|
||||||
|
}
|
||||||
|
|
||||||
|
func parseOpenAIToolCall(toolCalls []interface{}) (func() []byte, string, bool) {
|
||||||
|
if len(toolCalls) == 0 {
|
||||||
|
return nil, "", false
|
||||||
|
}
|
||||||
|
tc := toolCalls[0].(map[string]interface{})
|
||||||
|
id, _ := tc["id"].(string)
|
||||||
|
function, _ := tc["function"].(map[string]interface{})
|
||||||
|
name, _ := function["name"].(string)
|
||||||
|
argsStr, _ := function["arguments"].(string)
|
||||||
|
var args map[string]string
|
||||||
|
if err := json.Unmarshal([]byte(argsStr), &args); err != nil {
|
||||||
|
return func() []byte {
|
||||||
|
return []byte(fmt.Sprintf(`{"error": "failed to parse arguments: %v"}`, err))
|
||||||
|
}, id, true
|
||||||
|
}
|
||||||
|
return func() []byte {
|
||||||
|
fn, ok := pwToolMap[name]
|
||||||
|
if !ok {
|
||||||
|
return []byte(fmt.Sprintf(`{"error": "tool %s not found"}`, name))
|
||||||
|
}
|
||||||
|
return fn(args)
|
||||||
|
}, id, true
|
||||||
|
}
|
||||||
|
|
||||||
|
func findToolCallFromText(text string) (func() []byte, string, bool) {
|
||||||
|
jsStr := toolCallRE.FindString(text)
|
||||||
|
if jsStr == "" {
|
||||||
|
return nil, "", false
|
||||||
|
}
|
||||||
|
jsStr = strings.TrimSpace(jsStr)
|
||||||
|
jsStr = strings.TrimPrefix(jsStr, "__tool_call__")
|
||||||
|
jsStr = strings.TrimSuffix(jsStr, "__tool_call__")
|
||||||
|
jsStr = strings.TrimSpace(jsStr)
|
||||||
|
start := strings.Index(jsStr, "{")
|
||||||
|
end := strings.LastIndex(jsStr, "}")
|
||||||
|
if start == -1 || end == -1 || end <= start {
|
||||||
|
return func() []byte {
|
||||||
|
return []byte(`{"error": "no valid JSON found in tool call"}`)
|
||||||
|
}, "", true
|
||||||
|
}
|
||||||
|
jsStr = jsStr[start : end+1]
|
||||||
|
var fc models.FuncCall
|
||||||
|
if err := json.Unmarshal([]byte(jsStr), &fc); err != nil {
|
||||||
|
return func() []byte {
|
||||||
|
return []byte(fmt.Sprintf(`{"error": "failed to parse tool call: %v}`, err))
|
||||||
|
}, "", true
|
||||||
|
}
|
||||||
|
if fc.ID == "" {
|
||||||
|
fc.ID = "call_" + generateToolCallID()
|
||||||
|
}
|
||||||
|
return func() []byte {
|
||||||
|
fn, ok := pwToolMap[fc.Name]
|
||||||
|
if !ok {
|
||||||
|
return []byte(fmt.Sprintf(`{"error": "tool %s not found"}`, fc.Name))
|
||||||
|
}
|
||||||
|
return fn(fc.Args)
|
||||||
|
}, fc.ID, true
|
||||||
|
}
|
||||||
|
|
||||||
|
func generateToolCallID() string {
|
||||||
|
return strconv.Itoa(len(pwToolMap) % 10000)
|
||||||
|
}
|
||||||
123
agent/request.go
123
agent/request.go
@@ -32,10 +32,14 @@ func detectAPI(api string) (isCompletion, isChat, isDeepSeek, isOpenRouter bool)
|
|||||||
type AgentClient struct {
|
type AgentClient struct {
|
||||||
cfg *config.Config
|
cfg *config.Config
|
||||||
getToken func() string
|
getToken func() string
|
||||||
log slog.Logger
|
log *slog.Logger
|
||||||
|
chatBody *models.ChatBody
|
||||||
|
sysprompt string
|
||||||
|
// lastToolCallID string
|
||||||
|
tools []models.Tool
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewAgentClient(cfg *config.Config, log slog.Logger, gt func() string) *AgentClient {
|
func NewAgentClient(cfg *config.Config, log *slog.Logger, gt func() string) *AgentClient {
|
||||||
return &AgentClient{
|
return &AgentClient{
|
||||||
cfg: cfg,
|
cfg: cfg,
|
||||||
getToken: gt,
|
getToken: gt,
|
||||||
@@ -44,11 +48,46 @@ func NewAgentClient(cfg *config.Config, log slog.Logger, gt func() string) *Agen
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (ag *AgentClient) Log() *slog.Logger {
|
func (ag *AgentClient) Log() *slog.Logger {
|
||||||
return &ag.log
|
return ag.log
|
||||||
}
|
}
|
||||||
|
|
||||||
func (ag *AgentClient) FormMsg(sysprompt, msg string) (io.Reader, error) {
|
func (ag *AgentClient) FormFirstMsg(sysprompt, msg string) (io.Reader, error) {
|
||||||
b, err := ag.buildRequest(sysprompt, msg)
|
ag.sysprompt = sysprompt
|
||||||
|
ag.chatBody = &models.ChatBody{
|
||||||
|
Messages: []models.RoleMsg{
|
||||||
|
{Role: "system", Content: ag.sysprompt},
|
||||||
|
{Role: "user", Content: msg},
|
||||||
|
},
|
||||||
|
Stream: false,
|
||||||
|
Model: ag.cfg.CurrentModel,
|
||||||
|
}
|
||||||
|
b, err := ag.buildRequest()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return bytes.NewReader(b), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (ag *AgentClient) FormMsg(msg string) (io.Reader, error) {
|
||||||
|
m := models.RoleMsg{
|
||||||
|
Role: "tool", Content: msg,
|
||||||
|
}
|
||||||
|
ag.chatBody.Messages = append(ag.chatBody.Messages, m)
|
||||||
|
b, err := ag.buildRequest()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return bytes.NewReader(b), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (ag *AgentClient) FormMsgWithToolCallID(msg, toolCallID string) (io.Reader, error) {
|
||||||
|
m := models.RoleMsg{
|
||||||
|
Role: "tool",
|
||||||
|
Content: msg,
|
||||||
|
ToolCallID: toolCallID,
|
||||||
|
}
|
||||||
|
ag.chatBody.Messages = append(ag.chatBody.Messages, m)
|
||||||
|
b, err := ag.buildRequest()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@@ -56,81 +95,52 @@ func (ag *AgentClient) FormMsg(sysprompt, msg string) (io.Reader, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// buildRequest creates the appropriate LLM request based on the current API endpoint.
|
// buildRequest creates the appropriate LLM request based on the current API endpoint.
|
||||||
func (ag *AgentClient) buildRequest(sysprompt, msg string) ([]byte, error) {
|
func (ag *AgentClient) buildRequest() ([]byte, error) {
|
||||||
api := ag.cfg.CurrentAPI
|
isCompletion, isChat, isDeepSeek, isOpenRouter := detectAPI(ag.cfg.CurrentAPI)
|
||||||
model := ag.cfg.CurrentModel
|
ag.log.Debug("agent building request", "api", ag.cfg.CurrentAPI, "isCompletion", isCompletion, "isChat", isChat, "isDeepSeek", isDeepSeek, "isOpenRouter", isOpenRouter)
|
||||||
messages := []models.RoleMsg{
|
|
||||||
{Role: "system", Content: sysprompt},
|
|
||||||
{Role: "user", Content: msg},
|
|
||||||
}
|
|
||||||
|
|
||||||
// Determine API type
|
|
||||||
isCompletion, isChat, isDeepSeek, isOpenRouter := detectAPI(api)
|
|
||||||
ag.log.Debug("agent building request", "api", api, "isCompletion", isCompletion, "isChat", isChat, "isDeepSeek", isDeepSeek, "isOpenRouter", isOpenRouter)
|
|
||||||
|
|
||||||
// Build prompt for completion endpoints
|
// Build prompt for completion endpoints
|
||||||
if isCompletion {
|
if isCompletion {
|
||||||
var sb strings.Builder
|
var sb strings.Builder
|
||||||
for i := range messages {
|
for i := range ag.chatBody.Messages {
|
||||||
sb.WriteString(messages[i].ToPrompt())
|
sb.WriteString(ag.chatBody.Messages[i].ToPrompt())
|
||||||
sb.WriteString("\n")
|
sb.WriteString("\n")
|
||||||
}
|
}
|
||||||
prompt := strings.TrimSpace(sb.String())
|
prompt := strings.TrimSpace(sb.String())
|
||||||
|
|
||||||
switch {
|
switch {
|
||||||
case isDeepSeek:
|
case isDeepSeek:
|
||||||
// DeepSeek completion
|
// DeepSeek completion
|
||||||
req := models.NewDSCompletionReq(prompt, model, defaultProps["temperature"], []string{})
|
req := models.NewDSCompletionReq(prompt, ag.chatBody.Model, defaultProps["temperature"], []string{})
|
||||||
req.Stream = false // Agents don't need streaming
|
req.Stream = false // Agents don't need streaming
|
||||||
return json.Marshal(req)
|
return json.Marshal(req)
|
||||||
case isOpenRouter:
|
case isOpenRouter:
|
||||||
// OpenRouter completion
|
// OpenRouter completion
|
||||||
req := models.NewOpenRouterCompletionReq(model, prompt, defaultProps, []string{})
|
req := models.NewOpenRouterCompletionReq(ag.chatBody.Model, prompt, defaultProps, []string{})
|
||||||
req.Stream = false // Agents don't need streaming
|
req.Stream = false // Agents don't need streaming
|
||||||
return json.Marshal(req)
|
return json.Marshal(req)
|
||||||
default:
|
default:
|
||||||
// Assume llama.cpp completion
|
// Assume llama.cpp completion
|
||||||
req := models.NewLCPReq(prompt, model, nil, defaultProps, []string{})
|
req := models.NewLCPReq(prompt, ag.chatBody.Model, nil, defaultProps, []string{})
|
||||||
req.Stream = false // Agents don't need streaming
|
req.Stream = false // Agents don't need streaming
|
||||||
return json.Marshal(req)
|
return json.Marshal(req)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Chat completions endpoints
|
|
||||||
if isChat || !isCompletion {
|
|
||||||
chatBody := &models.ChatBody{
|
|
||||||
Model: model,
|
|
||||||
Stream: false, // Agents don't need streaming
|
|
||||||
Messages: messages,
|
|
||||||
}
|
|
||||||
|
|
||||||
switch {
|
switch {
|
||||||
case isDeepSeek:
|
case isDeepSeek:
|
||||||
// DeepSeek chat
|
// DeepSeek chat
|
||||||
req := models.NewDSChatReq(*chatBody)
|
req := models.NewDSChatReq(*ag.chatBody)
|
||||||
return json.Marshal(req)
|
return json.Marshal(req)
|
||||||
case isOpenRouter:
|
case isOpenRouter:
|
||||||
// OpenRouter chat - agents don't use reasoning by default
|
// OpenRouter chat - agents don't use reasoning by default
|
||||||
req := models.NewOpenRouterChatReq(*chatBody, defaultProps, "")
|
req := models.NewOpenRouterChatReq(*ag.chatBody, defaultProps, ag.cfg.ReasoningEffort)
|
||||||
return json.Marshal(req)
|
return json.Marshal(req)
|
||||||
default:
|
default:
|
||||||
// Assume llama.cpp chat (OpenAI format)
|
// Assume llama.cpp chat (OpenAI format)
|
||||||
req := models.OpenAIReq{
|
req := models.OpenAIReq{
|
||||||
ChatBody: chatBody,
|
ChatBody: ag.chatBody,
|
||||||
Tools: nil,
|
Tools: ag.tools,
|
||||||
}
|
}
|
||||||
return json.Marshal(req)
|
return json.Marshal(req)
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
// Fallback (should not reach here)
|
|
||||||
ag.log.Warn("unknown API, using default chat completions format", "api", api)
|
|
||||||
chatBody := &models.ChatBody{
|
|
||||||
Model: model,
|
|
||||||
Stream: false, // Agents don't need streaming
|
|
||||||
Messages: messages,
|
|
||||||
}
|
|
||||||
return json.Marshal(chatBody)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (ag *AgentClient) LLMRequest(body io.Reader) ([]byte, error) {
|
func (ag *AgentClient) LLMRequest(body io.Reader) ([]byte, error) {
|
||||||
@@ -165,7 +175,6 @@ func (ag *AgentClient) LLMRequest(body io.Reader) ([]byte, error) {
|
|||||||
ag.log.Error("agent LLM request failed", "status", resp.StatusCode, "response", string(responseBytes[:min(len(responseBytes), 1000)]))
|
ag.log.Error("agent LLM request failed", "status", resp.StatusCode, "response", string(responseBytes[:min(len(responseBytes), 1000)]))
|
||||||
return responseBytes, fmt.Errorf("HTTP %d: %s", resp.StatusCode, string(responseBytes[:min(len(responseBytes), 200)]))
|
return responseBytes, fmt.Errorf("HTTP %d: %s", resp.StatusCode, string(responseBytes[:min(len(responseBytes), 200)]))
|
||||||
}
|
}
|
||||||
|
|
||||||
// Parse response and extract text content
|
// Parse response and extract text content
|
||||||
text, err := extractTextFromResponse(responseBytes)
|
text, err := extractTextFromResponse(responseBytes)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -179,17 +188,16 @@ func (ag *AgentClient) LLMRequest(body io.Reader) ([]byte, error) {
|
|||||||
// extractTextFromResponse parses common LLM response formats and extracts the text content.
|
// extractTextFromResponse parses common LLM response formats and extracts the text content.
|
||||||
func extractTextFromResponse(data []byte) (string, error) {
|
func extractTextFromResponse(data []byte) (string, error) {
|
||||||
// Try to parse as generic JSON first
|
// Try to parse as generic JSON first
|
||||||
var genericResp map[string]interface{}
|
var genericResp map[string]any
|
||||||
if err := json.Unmarshal(data, &genericResp); err != nil {
|
if err := json.Unmarshal(data, &genericResp); err != nil {
|
||||||
// Not JSON, return as string
|
// Not JSON, return as string
|
||||||
return string(data), nil
|
return string(data), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check for OpenAI chat completion format
|
// Check for OpenAI chat completion format
|
||||||
if choices, ok := genericResp["choices"].([]interface{}); ok && len(choices) > 0 {
|
if choices, ok := genericResp["choices"].([]any); ok && len(choices) > 0 {
|
||||||
if firstChoice, ok := choices[0].(map[string]interface{}); ok {
|
if firstChoice, ok := choices[0].(map[string]any); ok {
|
||||||
// Chat completion: choices[0].message.content
|
// Chat completion: choices[0].message.content
|
||||||
if message, ok := firstChoice["message"].(map[string]interface{}); ok {
|
if message, ok := firstChoice["message"].(map[string]any); ok {
|
||||||
if content, ok := message["content"].(string); ok {
|
if content, ok := message["content"].(string); ok {
|
||||||
return content, nil
|
return content, nil
|
||||||
}
|
}
|
||||||
@@ -199,19 +207,17 @@ func extractTextFromResponse(data []byte) (string, error) {
|
|||||||
return text, nil
|
return text, nil
|
||||||
}
|
}
|
||||||
// Delta format for streaming (should not happen with stream: false)
|
// Delta format for streaming (should not happen with stream: false)
|
||||||
if delta, ok := firstChoice["delta"].(map[string]interface{}); ok {
|
if delta, ok := firstChoice["delta"].(map[string]any); ok {
|
||||||
if content, ok := delta["content"].(string); ok {
|
if content, ok := delta["content"].(string); ok {
|
||||||
return content, nil
|
return content, nil
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check for llama.cpp completion format
|
// Check for llama.cpp completion format
|
||||||
if content, ok := genericResp["content"].(string); ok {
|
if content, ok := genericResp["content"].(string); ok {
|
||||||
return content, nil
|
return content, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Unknown format, return pretty-printed JSON
|
// Unknown format, return pretty-printed JSON
|
||||||
prettyJSON, err := json.MarshalIndent(genericResp, "", " ")
|
prettyJSON, err := json.MarshalIndent(genericResp, "", " ")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -219,10 +225,3 @@ func extractTextFromResponse(data []byte) (string, error) {
|
|||||||
}
|
}
|
||||||
return string(prettyJSON), nil
|
return string(prettyJSON), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func min(a, b int) int {
|
|
||||||
if a < b {
|
|
||||||
return a
|
|
||||||
}
|
|
||||||
return b
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -17,7 +17,8 @@ func NewWebAgentB(client *AgentClient, sysprompt string) *WebAgentB {
|
|||||||
|
|
||||||
// Process applies the formatting function to raw output
|
// Process applies the formatting function to raw output
|
||||||
func (a *WebAgentB) Process(args map[string]string, rawOutput []byte) []byte {
|
func (a *WebAgentB) Process(args map[string]string, rawOutput []byte) []byte {
|
||||||
msg, err := a.FormMsg(a.sysprompt,
|
msg, err := a.FormFirstMsg(
|
||||||
|
a.sysprompt,
|
||||||
fmt.Sprintf("request:\n%+v\ntool response:\n%v", args, string(rawOutput)))
|
fmt.Sprintf("request:\n%+v\ntool response:\n%v", args, string(rawOutput)))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
a.Log().Error("failed to process the request", "error", err)
|
a.Log().Error("failed to process the request", "error", err)
|
||||||
|
|||||||
BIN
assets/yt_thumb.jpg
Normal file
BIN
assets/yt_thumb.jpg
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 42 KiB |
@@ -6,19 +6,27 @@ services:
|
|||||||
ports:
|
ports:
|
||||||
- "8081:8081"
|
- "8081:8081"
|
||||||
volumes:
|
volumes:
|
||||||
- whisper_models:/app/models
|
- ./whisper.cpp/models/ggml-large-v3-turbo-q5_0.bin:/app/models/ggml-large-v3-turbo-q5_0.bin
|
||||||
working_dir: /app
|
working_dir: /app
|
||||||
entrypoint: ""
|
entrypoint: ""
|
||||||
command: >
|
command: >
|
||||||
sh -c "
|
sh -c "
|
||||||
if [ ! -f /app/models/ggml-large-v3-turbo.bin ]; then
|
if [ ! -f /app/models/ggml-large-v3-turbo-q5_0.bin ]; then
|
||||||
echo 'Downloading ggml-large-v3-turbo model...'
|
echo 'Downloading ggml-large-v3-turboq5_0 model...'
|
||||||
./download-ggml-model.sh large-v3-turbo /app/models
|
curl -o /app/models/ggml-large-v3-turbo-q5_0.bin -L "https://huggingface.co/ggerganov/whisper.cpp/resolve/main/ggml-large-v3-turbo-q5_0.bin?download=true"
|
||||||
fi &&
|
fi &&
|
||||||
./build/bin/whisper-server -m /app/models/ggml-large-v3-turbo.bin -t 4 -p 1 --port 8081 --host 0.0.0.0
|
./build/bin/whisper-server -m /app/models/ggml-large-v3-turbo-q5_0.bin -t 4 -p 1 --port 8081 --host 0.0.0.0
|
||||||
"
|
"
|
||||||
environment:
|
environment:
|
||||||
- WHISPER_LOG_LEVEL=3
|
- WHISPER_LOG_LEVEL=3
|
||||||
|
# For GPU support, uncomment the following lines:
|
||||||
|
deploy:
|
||||||
|
resources:
|
||||||
|
reservations:
|
||||||
|
devices:
|
||||||
|
- driver: nvidia
|
||||||
|
count: 1
|
||||||
|
capabilities: [gpu]
|
||||||
# Restart policy in case the service fails
|
# Restart policy in case the service fails
|
||||||
restart: unless-stopped
|
restart: unless-stopped
|
||||||
|
|
||||||
@@ -45,7 +53,5 @@ services:
|
|||||||
volumes:
|
volumes:
|
||||||
models:
|
models:
|
||||||
driver: local
|
driver: local
|
||||||
audio:
|
|
||||||
driver: local
|
|
||||||
whisper_models:
|
whisper_models:
|
||||||
driver: local
|
driver: local
|
||||||
|
|||||||
390
bot.go
390
bot.go
@@ -11,19 +11,22 @@ import (
|
|||||||
"gf-lt/models"
|
"gf-lt/models"
|
||||||
"gf-lt/rag"
|
"gf-lt/rag"
|
||||||
"gf-lt/storage"
|
"gf-lt/storage"
|
||||||
|
"gf-lt/tools"
|
||||||
"html"
|
"html"
|
||||||
"io"
|
"io"
|
||||||
"log/slog"
|
"log/slog"
|
||||||
"net"
|
"net"
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/url"
|
|
||||||
"os"
|
"os"
|
||||||
"regexp"
|
"regexp"
|
||||||
"slices"
|
"slices"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/rivo/tview"
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
@@ -39,14 +42,82 @@ var (
|
|||||||
streamDone = make(chan bool, 1)
|
streamDone = make(chan bool, 1)
|
||||||
chatBody *models.ChatBody
|
chatBody *models.ChatBody
|
||||||
store storage.FullRepo
|
store storage.FullRepo
|
||||||
defaultFirstMsg = "Hello! What can I do for you?"
|
|
||||||
defaultStarter = []models.RoleMsg{}
|
defaultStarter = []models.RoleMsg{}
|
||||||
interruptResp = false
|
interruptResp atomic.Bool
|
||||||
ragger *rag.RAG
|
ragger *rag.RAG
|
||||||
chunkParser ChunkParser
|
chunkParser ChunkParser
|
||||||
lastToolCall *models.FuncCall
|
lastToolCall *models.FuncCall
|
||||||
lastRespStats *models.ResponseStats
|
lastRespStats *models.ResponseStats
|
||||||
//nolint:unused // TTS_ENABLED conditionally uses this
|
|
||||||
|
outputHandler OutputHandler
|
||||||
|
cliPrevOutput string
|
||||||
|
cliRespDone chan bool
|
||||||
|
)
|
||||||
|
|
||||||
|
type OutputHandler interface {
|
||||||
|
Write(p string)
|
||||||
|
Writef(format string, args ...interface{})
|
||||||
|
ScrollToEnd()
|
||||||
|
}
|
||||||
|
|
||||||
|
type TUIOutputHandler struct {
|
||||||
|
tv *tview.TextView
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *TUIOutputHandler) Write(p string) {
|
||||||
|
if h.tv != nil {
|
||||||
|
fmt.Fprint(h.tv, p)
|
||||||
|
}
|
||||||
|
if cfg != nil && cfg.CLIMode {
|
||||||
|
fmt.Print(p)
|
||||||
|
cliPrevOutput = p
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *TUIOutputHandler) Writef(format string, args ...interface{}) {
|
||||||
|
s := fmt.Sprintf(format, args...)
|
||||||
|
if h.tv != nil {
|
||||||
|
fmt.Fprint(h.tv, s)
|
||||||
|
}
|
||||||
|
if cfg != nil && cfg.CLIMode {
|
||||||
|
fmt.Print(s)
|
||||||
|
cliPrevOutput = s
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *TUIOutputHandler) ScrollToEnd() {
|
||||||
|
if h.tv != nil {
|
||||||
|
h.tv.ScrollToEnd()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type CLIOutputHandler struct{}
|
||||||
|
|
||||||
|
func (h *CLIOutputHandler) Write(p string) {
|
||||||
|
fmt.Print(p)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *CLIOutputHandler) Writef(format string, args ...interface{}) {
|
||||||
|
fmt.Printf(format, args...)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (h *CLIOutputHandler) ScrollToEnd() {
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
basicCard = &models.CharCard{
|
||||||
|
ID: models.ComputeCardID("assistant", "basic_sys"),
|
||||||
|
SysPrompt: models.BasicSysMsg,
|
||||||
|
FirstMsg: models.DefaultFirstMsg,
|
||||||
|
Role: "assistant",
|
||||||
|
FilePath: "basic_sys",
|
||||||
|
}
|
||||||
|
sysMap = map[string]*models.CharCard{}
|
||||||
|
roleToID = map[string]string{}
|
||||||
|
modelHasVision bool
|
||||||
|
windowToolsAvailable bool
|
||||||
|
// tooler *tools.Tools
|
||||||
|
//
|
||||||
orator Orator
|
orator Orator
|
||||||
asr STT
|
asr STT
|
||||||
localModelsMu sync.RWMutex
|
localModelsMu sync.RWMutex
|
||||||
@@ -253,12 +324,7 @@ func createClient(connectTimeout time.Duration) *http.Client {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func warmUpModel() {
|
func warmUpModel() {
|
||||||
u, err := url.Parse(cfg.CurrentAPI)
|
if !isLocalLlamacpp() {
|
||||||
if err != nil {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
host := u.Hostname()
|
|
||||||
if host != "localhost" && host != "127.0.0.1" && host != "::1" {
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
// Check if model is already loaded
|
// Check if model is already loaded
|
||||||
@@ -418,9 +484,7 @@ func fetchLCPModelsWithStatus() (*models.LCPModels, error) {
|
|||||||
if err := json.NewDecoder(resp.Body).Decode(data); err != nil {
|
if err := json.NewDecoder(resp.Body).Decode(data); err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
localModelsMu.Lock()
|
|
||||||
localModelsData = data
|
localModelsData = data
|
||||||
localModelsMu.Unlock()
|
|
||||||
return data, nil
|
return data, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -465,6 +529,29 @@ func ModelHasVision(api, modelID string) bool {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func UpdateToolCapabilities() {
|
||||||
|
if !cfg.ToolUse {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
modelHasVision = false
|
||||||
|
if cfg == nil || cfg.CurrentAPI == "" {
|
||||||
|
logger.Warn("cannot determine model capabilities: cfg or CurrentAPI is nil")
|
||||||
|
// tooler.RegisterWindowTools(modelHasVision)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
prevHasVision := modelHasVision
|
||||||
|
modelHasVision = ModelHasVision(cfg.CurrentAPI, cfg.CurrentModel)
|
||||||
|
if modelHasVision {
|
||||||
|
logger.Info("model has vision support", "model", cfg.CurrentModel, "api", cfg.CurrentAPI)
|
||||||
|
} else {
|
||||||
|
logger.Info("model does not have vision support", "model", cfg.CurrentModel, "api", cfg.CurrentAPI)
|
||||||
|
if windowToolsAvailable && !prevHasVision && !modelHasVision {
|
||||||
|
showToast("window tools", "Window capture-and-view unavailable: model lacks vision support")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// tooler.RegisterWindowTools(modelHasVision)
|
||||||
|
}
|
||||||
|
|
||||||
// monitorModelLoad starts a goroutine that periodically checks if the specified model is loaded.
|
// monitorModelLoad starts a goroutine that periodically checks if the specified model is loaded.
|
||||||
func monitorModelLoad(modelID string) {
|
func monitorModelLoad(modelID string) {
|
||||||
go func() {
|
go func() {
|
||||||
@@ -651,7 +738,7 @@ func sendMsgToLLM(body io.Reader) {
|
|||||||
// continue
|
// continue
|
||||||
}
|
}
|
||||||
if len(line) <= 1 {
|
if len(line) <= 1 {
|
||||||
if interruptResp {
|
if interruptResp.Load() {
|
||||||
goto interrupt // get unstuck from bad connection
|
goto interrupt // get unstuck from bad connection
|
||||||
}
|
}
|
||||||
continue // skip \n
|
continue // skip \n
|
||||||
@@ -744,8 +831,7 @@ func sendMsgToLLM(body io.Reader) {
|
|||||||
lastToolCall.ID = chunk.ToolID
|
lastToolCall.ID = chunk.ToolID
|
||||||
}
|
}
|
||||||
interrupt:
|
interrupt:
|
||||||
if interruptResp { // read bytes, so it would not get into beginning of the next req
|
if interruptResp.Load() { // read bytes, so it would not get into beginning of the next req
|
||||||
// interruptResp = false
|
|
||||||
logger.Info("interrupted bot response", "chunk_counter", counter)
|
logger.Info("interrupted bot response", "chunk_counter", counter)
|
||||||
streamDone <- true
|
streamDone <- true
|
||||||
break
|
break
|
||||||
@@ -772,20 +858,24 @@ func chatWatcher(ctx context.Context) {
|
|||||||
|
|
||||||
// inpired by https://github.com/rivo/tview/issues/225
|
// inpired by https://github.com/rivo/tview/issues/225
|
||||||
func showSpinner() {
|
func showSpinner() {
|
||||||
|
if cfg.CLIMode {
|
||||||
|
showSpinnerCLI()
|
||||||
|
return
|
||||||
|
}
|
||||||
spinners := []string{"⠋", "⠙", "⠹", "⠸", "⠼", "⠴", "⠦", "⠧", "⠇", "⠏"}
|
spinners := []string{"⠋", "⠙", "⠹", "⠸", "⠼", "⠴", "⠦", "⠧", "⠇", "⠏"}
|
||||||
var i int
|
var i int
|
||||||
botPersona := cfg.AssistantRole
|
botPersona := cfg.AssistantRole
|
||||||
if cfg.WriteNextMsgAsCompletionAgent != "" {
|
if cfg.WriteNextMsgAsCompletionAgent != "" {
|
||||||
botPersona = cfg.WriteNextMsgAsCompletionAgent
|
botPersona = cfg.WriteNextMsgAsCompletionAgent
|
||||||
}
|
}
|
||||||
for botRespMode || toolRunningMode {
|
for botRespMode.Load() || toolRunningMode.Load() {
|
||||||
time.Sleep(400 * time.Millisecond)
|
time.Sleep(400 * time.Millisecond)
|
||||||
spin := i % len(spinners)
|
spin := i % len(spinners)
|
||||||
app.QueueUpdateDraw(func() {
|
app.QueueUpdateDraw(func() {
|
||||||
switch {
|
switch {
|
||||||
case toolRunningMode:
|
case toolRunningMode.Load():
|
||||||
textArea.SetTitle(spinners[spin] + " tool")
|
textArea.SetTitle(spinners[spin] + " tool")
|
||||||
case botRespMode:
|
case botRespMode.Load():
|
||||||
textArea.SetTitle(spinners[spin] + " " + botPersona + " (F6 to interrupt)")
|
textArea.SetTitle(spinners[spin] + " " + botPersona + " (F6 to interrupt)")
|
||||||
default:
|
default:
|
||||||
textArea.SetTitle(spinners[spin] + " input")
|
textArea.SetTitle(spinners[spin] + " input")
|
||||||
@@ -798,9 +888,15 @@ func showSpinner() {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func showSpinnerCLI() {
|
||||||
|
for botRespMode.Load() || toolRunningMode.Load() {
|
||||||
|
time.Sleep(400 * time.Millisecond)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func chatRound(r *models.ChatRoundReq) error {
|
func chatRound(r *models.ChatRoundReq) error {
|
||||||
interruptResp = false
|
interruptResp.Store(false)
|
||||||
botRespMode = true
|
botRespMode.Store(true)
|
||||||
go showSpinner()
|
go showSpinner()
|
||||||
updateStatusLine()
|
updateStatusLine()
|
||||||
botPersona := cfg.AssistantRole
|
botPersona := cfg.AssistantRole
|
||||||
@@ -808,7 +904,7 @@ func chatRound(r *models.ChatRoundReq) error {
|
|||||||
botPersona = cfg.WriteNextMsgAsCompletionAgent
|
botPersona = cfg.WriteNextMsgAsCompletionAgent
|
||||||
}
|
}
|
||||||
defer func() {
|
defer func() {
|
||||||
botRespMode = false
|
botRespMode.Store(false)
|
||||||
ClearImageAttachment()
|
ClearImageAttachment()
|
||||||
}()
|
}()
|
||||||
// check that there is a model set to use if is not local
|
// check that there is a model set to use if is not local
|
||||||
@@ -830,13 +926,22 @@ func chatRound(r *models.ChatRoundReq) error {
|
|||||||
Role: botPersona, Content: "",
|
Role: botPersona, Content: "",
|
||||||
})
|
})
|
||||||
nl := "\n\n"
|
nl := "\n\n"
|
||||||
prevText := textView.GetText(true)
|
prevText := cliPrevOutput
|
||||||
|
if cfg.CLIMode {
|
||||||
if strings.HasSuffix(prevText, nl) {
|
if strings.HasSuffix(prevText, nl) {
|
||||||
nl = ""
|
nl = ""
|
||||||
} else if strings.HasSuffix(prevText, "\n") {
|
} else if strings.HasSuffix(prevText, "\n") {
|
||||||
nl = "\n"
|
nl = "\n"
|
||||||
}
|
}
|
||||||
fmt.Fprintf(textView, "%s[-:-:b](%d) %s[-:-:-]\n", nl, msgIdx, roleToIcon(botPersona))
|
} else {
|
||||||
|
prevText = textView.GetText(true)
|
||||||
|
if strings.HasSuffix(prevText, nl) {
|
||||||
|
nl = ""
|
||||||
|
} else if strings.HasSuffix(prevText, "\n") {
|
||||||
|
nl = "\n"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
outputHandler.Writef("%s[-:-:b](%d) %s[-:-:-]\n", nl, msgIdx, roleToIcon(botPersona))
|
||||||
} else {
|
} else {
|
||||||
msgIdx = len(chatBody.Messages) - 1
|
msgIdx = len(chatBody.Messages) - 1
|
||||||
}
|
}
|
||||||
@@ -858,9 +963,9 @@ out:
|
|||||||
thinkingBuffer.WriteString(chunk)
|
thinkingBuffer.WriteString(chunk)
|
||||||
if thinkingCollapsed {
|
if thinkingCollapsed {
|
||||||
// Show placeholder immediately when thinking starts in collapsed mode
|
// Show placeholder immediately when thinking starts in collapsed mode
|
||||||
fmt.Fprint(textView, "[yellow::i][thinking... (press Alt+T to expand)][-:-:-]")
|
outputHandler.Write("[yellow::i][thinking... (press Alt+T to expand)][-:-:-]")
|
||||||
if scrollToEndEnabled {
|
if cfg.AutoScrollEnabled {
|
||||||
textView.ScrollToEnd()
|
outputHandler.ScrollToEnd()
|
||||||
}
|
}
|
||||||
respText.WriteString(chunk)
|
respText.WriteString(chunk)
|
||||||
continue
|
continue
|
||||||
@@ -874,8 +979,8 @@ out:
|
|||||||
// Thinking already displayed as placeholder, just update respText
|
// Thinking already displayed as placeholder, just update respText
|
||||||
respText.WriteString(chunk)
|
respText.WriteString(chunk)
|
||||||
justExitedThinkingCollapsed = true
|
justExitedThinkingCollapsed = true
|
||||||
if scrollToEndEnabled {
|
if cfg.AutoScrollEnabled {
|
||||||
textView.ScrollToEnd()
|
outputHandler.ScrollToEnd()
|
||||||
}
|
}
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
@@ -892,30 +997,32 @@ out:
|
|||||||
chunk = "\n\n" + chunk
|
chunk = "\n\n" + chunk
|
||||||
justExitedThinkingCollapsed = false
|
justExitedThinkingCollapsed = false
|
||||||
}
|
}
|
||||||
fmt.Fprint(textView, chunk)
|
outputHandler.Write(chunk)
|
||||||
respText.WriteString(chunk)
|
respText.WriteString(chunk)
|
||||||
// Update the message in chatBody.Messages so it persists during Alt+T
|
// Update the message in chatBody.Messages so it persists during Alt+T
|
||||||
chatBody.Messages[msgIdx].Content = respText.String()
|
if !r.Resume {
|
||||||
if scrollToEndEnabled {
|
chatBody.Messages[msgIdx].Content += respText.String()
|
||||||
textView.ScrollToEnd()
|
}
|
||||||
|
if cfg.AutoScrollEnabled {
|
||||||
|
outputHandler.ScrollToEnd()
|
||||||
}
|
}
|
||||||
// Send chunk to audio stream handler
|
// Send chunk to audio stream handler
|
||||||
if cfg.TTS_ENABLED {
|
if cfg.TTS_ENABLED {
|
||||||
TTSTextChan <- chunk
|
TTSTextChan <- chunk
|
||||||
}
|
}
|
||||||
case toolChunk := <-openAIToolChan:
|
case toolChunk := <-openAIToolChan:
|
||||||
fmt.Fprint(textView, toolChunk)
|
outputHandler.Write(toolChunk)
|
||||||
toolResp.WriteString(toolChunk)
|
toolResp.WriteString(toolChunk)
|
||||||
if scrollToEndEnabled {
|
if cfg.AutoScrollEnabled {
|
||||||
textView.ScrollToEnd()
|
outputHandler.ScrollToEnd()
|
||||||
}
|
}
|
||||||
case <-streamDone:
|
case <-streamDone:
|
||||||
for len(chunkChan) > 0 {
|
for len(chunkChan) > 0 {
|
||||||
chunk := <-chunkChan
|
chunk := <-chunkChan
|
||||||
fmt.Fprint(textView, chunk)
|
outputHandler.Write(chunk)
|
||||||
respText.WriteString(chunk)
|
respText.WriteString(chunk)
|
||||||
if scrollToEndEnabled {
|
if cfg.AutoScrollEnabled {
|
||||||
textView.ScrollToEnd()
|
outputHandler.ScrollToEnd()
|
||||||
}
|
}
|
||||||
if cfg.TTS_ENABLED {
|
if cfg.TTS_ENABLED {
|
||||||
TTSTextChan <- chunk
|
TTSTextChan <- chunk
|
||||||
@@ -936,7 +1043,7 @@ out:
|
|||||||
}
|
}
|
||||||
lastRespStats = nil
|
lastRespStats = nil
|
||||||
}
|
}
|
||||||
botRespMode = false
|
botRespMode.Store(false)
|
||||||
if r.Resume {
|
if r.Resume {
|
||||||
chatBody.Messages[len(chatBody.Messages)-1].Content += respText.String()
|
chatBody.Messages[len(chatBody.Messages)-1].Content += respText.String()
|
||||||
updatedMsg := chatBody.Messages[len(chatBody.Messages)-1]
|
updatedMsg := chatBody.Messages[len(chatBody.Messages)-1]
|
||||||
@@ -957,20 +1064,27 @@ out:
|
|||||||
cleanChatBody()
|
cleanChatBody()
|
||||||
refreshChatDisplay()
|
refreshChatDisplay()
|
||||||
updateStatusLine()
|
updateStatusLine()
|
||||||
// bot msg is done;
|
// bot msg is done; now check it for func call
|
||||||
// now check it for func call
|
|
||||||
// logChat(activeChatName, chatBody.Messages)
|
// logChat(activeChatName, chatBody.Messages)
|
||||||
if err := updateStorageChat(activeChatName, chatBody.Messages); err != nil {
|
if err := updateStorageChat(activeChatName, chatBody.Messages); err != nil {
|
||||||
logger.Warn("failed to update storage", "error", err, "name", activeChatName)
|
logger.Warn("failed to update storage", "error", err, "name", activeChatName)
|
||||||
}
|
}
|
||||||
// Strip think blocks before parsing for tool calls
|
// Strip think blocks before parsing for tool calls
|
||||||
respTextNoThink := thinkBlockRE.ReplaceAllString(respText.String(), "")
|
respTextNoThink := thinkBlockRE.ReplaceAllString(respText.String(), "")
|
||||||
if interruptResp {
|
if interruptResp.Load() {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if findCall(respTextNoThink, toolResp.String()) {
|
if findCall(respTextNoThink, toolResp.String()) {
|
||||||
|
// Tool was found and executed, subsequent chatRound will signal cliRespDone when complete
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
// No tool call - signal completion now
|
||||||
|
if cfg.CLIMode && cliRespDone != nil {
|
||||||
|
select {
|
||||||
|
case cliRespDone <- true:
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
}
|
||||||
// Check if this message was sent privately to specific characters
|
// Check if this message was sent privately to specific characters
|
||||||
// If so, trigger those characters to respond if that char is not controlled by user
|
// If so, trigger those characters to respond if that char is not controlled by user
|
||||||
// perhaps we should have narrator role to determine which char is next to act
|
// perhaps we should have narrator role to determine which char is next to act
|
||||||
@@ -1108,7 +1222,7 @@ func findCall(msg, toolCall string) bool {
|
|||||||
// The ID should come from the streaming response (chunk.ToolID) set earlier.
|
// The ID should come from the streaming response (chunk.ToolID) set earlier.
|
||||||
// Some tools like todo_create have "id" in their arguments which is NOT the tool call ID.
|
// Some tools like todo_create have "id" in their arguments which is NOT the tool call ID.
|
||||||
} else {
|
} else {
|
||||||
jsStr := toolCallRE.FindString(msg)
|
jsStr := models.ToolCallRE.FindString(msg)
|
||||||
if jsStr == "" { // no tool call case
|
if jsStr == "" { // no tool call case
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
@@ -1176,17 +1290,42 @@ func findCall(msg, toolCall string) bool {
|
|||||||
Args: mapToString(lastToolCall.Args),
|
Args: mapToString(lastToolCall.Args),
|
||||||
}
|
}
|
||||||
// call a func
|
// call a func
|
||||||
_, ok := fnMap[fc.Name]
|
// _, ok := tools.FnMap[fc.Name]
|
||||||
if !ok {
|
// if !ok {
|
||||||
m := fc.Name + " is not implemented"
|
// m := fc.Name + " is not implemented"
|
||||||
|
// // Create tool response message with the proper tool_call_id
|
||||||
|
// toolResponseMsg := models.RoleMsg{
|
||||||
|
// Role: cfg.ToolRole,
|
||||||
|
// Content: m,
|
||||||
|
// ToolCallID: lastToolCall.ID, // Use the stored tool call ID
|
||||||
|
// }
|
||||||
|
// chatBody.Messages = append(chatBody.Messages, toolResponseMsg)
|
||||||
|
// logger.Debug("findCall: added tool not implemented response", "role", toolResponseMsg.Role, "content_len", len(toolResponseMsg.Content), "tool_call_id", toolResponseMsg.ToolCallID, "message_count_after_add", len(chatBody.Messages))
|
||||||
|
// // Clear the stored tool call ID after using it
|
||||||
|
// lastToolCall.ID = ""
|
||||||
|
// // Trigger the assistant to continue processing with the new tool response
|
||||||
|
// // by calling chatRound with empty content to continue the assistant's response
|
||||||
|
// crr := &models.ChatRoundReq{
|
||||||
|
// Role: cfg.AssistantRole,
|
||||||
|
// }
|
||||||
|
// // failed to find tool
|
||||||
|
// chatRoundChan <- crr
|
||||||
|
// return true
|
||||||
|
// }
|
||||||
|
// Show tool call progress indicator before execution
|
||||||
|
outputHandler.Writef("\n[yellow::i][tool: %s...][-:-:-]", fc.Name)
|
||||||
|
toolRunningMode.Store(true)
|
||||||
|
resp, okT := tools.CallToolWithAgent(fc.Name, fc.Args)
|
||||||
|
if !okT {
|
||||||
// Create tool response message with the proper tool_call_id
|
// Create tool response message with the proper tool_call_id
|
||||||
toolResponseMsg := models.RoleMsg{
|
toolResponseMsg := models.RoleMsg{
|
||||||
Role: cfg.ToolRole,
|
Role: cfg.ToolRole,
|
||||||
Content: m,
|
Content: string(resp),
|
||||||
ToolCallID: lastToolCall.ID, // Use the stored tool call ID
|
ToolCallID: lastToolCall.ID, // Use the stored tool call ID
|
||||||
}
|
}
|
||||||
chatBody.Messages = append(chatBody.Messages, toolResponseMsg)
|
chatBody.Messages = append(chatBody.Messages, toolResponseMsg)
|
||||||
logger.Debug("findCall: added tool not implemented response", "role", toolResponseMsg.Role, "content_len", len(toolResponseMsg.Content), "tool_call_id", toolResponseMsg.ToolCallID, "message_count_after_add", len(chatBody.Messages))
|
logger.Debug("findCall: added tool not implemented response", "role", toolResponseMsg.Role,
|
||||||
|
"content_len", len(toolResponseMsg.Content), "tool_call_id", toolResponseMsg.ToolCallID)
|
||||||
// Clear the stored tool call ID after using it
|
// Clear the stored tool call ID after using it
|
||||||
lastToolCall.ID = ""
|
lastToolCall.ID = ""
|
||||||
// Trigger the assistant to continue processing with the new tool response
|
// Trigger the assistant to continue processing with the new tool response
|
||||||
@@ -1198,11 +1337,7 @@ func findCall(msg, toolCall string) bool {
|
|||||||
chatRoundChan <- crr
|
chatRoundChan <- crr
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
// Show tool call progress indicator before execution
|
toolRunningMode.Store(false)
|
||||||
fmt.Fprintf(textView, "\n[yellow::i][tool: %s...][-:-:-]", fc.Name)
|
|
||||||
toolRunningMode = true
|
|
||||||
resp := callToolWithAgent(fc.Name, fc.Args)
|
|
||||||
toolRunningMode = false
|
|
||||||
toolMsg := string(resp)
|
toolMsg := string(resp)
|
||||||
logger.Info("llm used a tool call", "tool_name", fc.Name, "too_args", fc.Args, "id", fc.ID, "tool_resp", toolMsg)
|
logger.Info("llm used a tool call", "tool_name", fc.Name, "too_args", fc.Args, "id", fc.ID, "tool_resp", toolMsg)
|
||||||
// Create tool response message with the proper tool_call_id
|
// Create tool response message with the proper tool_call_id
|
||||||
@@ -1256,10 +1391,9 @@ func findCall(msg, toolCall string) bool {
|
|||||||
IsShellCommand: isShellCommand,
|
IsShellCommand: isShellCommand,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
fmt.Fprintf(textView, "%s[-:-:b](%d) <%s>: [-:-:-]\n%s\n",
|
outputHandler.Writef("%s[-:-:b](%d) <%s>: [-:-:-]\n%s\n",
|
||||||
"\n\n", len(chatBody.Messages), cfg.ToolRole, toolResponseMsg.GetText())
|
"\n\n", len(chatBody.Messages), cfg.ToolRole, toolResponseMsg.GetText())
|
||||||
chatBody.Messages = append(chatBody.Messages, toolResponseMsg)
|
chatBody.Messages = append(chatBody.Messages, toolResponseMsg)
|
||||||
logger.Debug("findCall: added actual tool response", "role", toolResponseMsg.Role, "content_len", len(toolResponseMsg.Content), "tool_call_id", toolResponseMsg.ToolCallID, "message_count_after_add", len(chatBody.Messages))
|
|
||||||
// Clear the stored tool call ID after using it
|
// Clear the stored tool call ID after using it
|
||||||
lastToolCall.ID = ""
|
lastToolCall.ID = ""
|
||||||
// Trigger the assistant to continue processing with the new tool response
|
// Trigger the assistant to continue processing with the new tool response
|
||||||
@@ -1280,11 +1414,19 @@ func chatToTextSlice(messages []models.RoleMsg, showSys bool) []string {
|
|||||||
// This is a tool call indicator - show collapsed
|
// This is a tool call indicator - show collapsed
|
||||||
if toolCollapsed {
|
if toolCollapsed {
|
||||||
toolName := messages[i].ToolCall.Name
|
toolName := messages[i].ToolCall.Name
|
||||||
resp[i] = strings.ReplaceAll(fmt.Sprintf("%s\n%s\n[yellow::i][tool call: %s (press Ctrl+T to expand)][-:-:-]\n", icon, messages[i].GetText(), toolName), "\n\n", "\n")
|
resp[i] = strings.ReplaceAll(
|
||||||
|
fmt.Sprintf(
|
||||||
|
"%s\n%s\n[yellow::i][tool call: %s (press Ctrl+T to expand)][-:-:-]\n",
|
||||||
|
icon, messages[i].GetText(), toolName),
|
||||||
|
"\n\n", "\n")
|
||||||
} else {
|
} else {
|
||||||
// Show full tool call info
|
// Show full tool call info
|
||||||
toolName := messages[i].ToolCall.Name
|
toolName := messages[i].ToolCall.Name
|
||||||
resp[i] = strings.ReplaceAll(fmt.Sprintf("%s\n%s\n[yellow::i][tool call: %s][-:-:-]\nargs: %s\nid: %s\n", icon, messages[i].GetText(), toolName, messages[i].ToolCall.Args, messages[i].ToolCall.ID), "\n\n", "\n")
|
resp[i] = strings.ReplaceAll(
|
||||||
|
fmt.Sprintf(
|
||||||
|
"%s\n%s\n[yellow::i][tool call: %s][-:-:-]\nargs: %s\nid: %s\n",
|
||||||
|
icon, messages[i].GetText(), toolName, messages[i].ToolCall.Args, messages[i].ToolCall.ID),
|
||||||
|
"\n\n", "\n")
|
||||||
}
|
}
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
@@ -1318,7 +1460,7 @@ func chatToText(messages []models.RoleMsg, showSys bool) string {
|
|||||||
text := strings.Join(s, "\n")
|
text := strings.Join(s, "\n")
|
||||||
// Collapse thinking blocks if enabled
|
// Collapse thinking blocks if enabled
|
||||||
if thinkingCollapsed {
|
if thinkingCollapsed {
|
||||||
text = thinkRE.ReplaceAllStringFunc(text, func(match string) string {
|
text = models.ThinkRE.ReplaceAllStringFunc(text, func(match string) string {
|
||||||
// Extract content between <think> and </think>
|
// Extract content between <think> and </think>
|
||||||
start := len("<think>")
|
start := len("<think>")
|
||||||
end := len(match) - len("</think>")
|
end := len(match) - len("</think>")
|
||||||
@@ -1334,7 +1476,9 @@ func chatToText(messages []models.RoleMsg, showSys bool) string {
|
|||||||
startIdx := strings.Index(text, "<think>")
|
startIdx := strings.Index(text, "<think>")
|
||||||
if startIdx != -1 {
|
if startIdx != -1 {
|
||||||
content := text[startIdx+len("<think>"):]
|
content := text[startIdx+len("<think>"):]
|
||||||
placeholder := fmt.Sprintf("[yellow::i][thinking... (%d chars) (press Alt+T to expand)][-:-:-]", len(content))
|
placeholder := fmt.Sprintf(
|
||||||
|
"[yellow::i][thinking... (%d chars) (press Alt+T to expand)][-:-:-]",
|
||||||
|
len(content))
|
||||||
text = text[:startIdx] + placeholder
|
text = text[:startIdx] + placeholder
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -1402,11 +1546,10 @@ func updateModelLists() {
|
|||||||
localModelsMu.Lock()
|
localModelsMu.Lock()
|
||||||
LocalModels = ml
|
LocalModels = ml
|
||||||
localModelsMu.Unlock()
|
localModelsMu.Unlock()
|
||||||
for statusLineWidget == nil {
|
|
||||||
time.Sleep(time.Millisecond * 100)
|
|
||||||
}
|
|
||||||
// set already loaded model in llama.cpp
|
// set already loaded model in llama.cpp
|
||||||
if strings.Contains(cfg.CurrentAPI, "localhost") || strings.Contains(cfg.CurrentAPI, "127.0.0.1") {
|
if !isLocalLlamacpp() {
|
||||||
|
return
|
||||||
|
}
|
||||||
localModelsMu.Lock()
|
localModelsMu.Lock()
|
||||||
defer localModelsMu.Unlock()
|
defer localModelsMu.Unlock()
|
||||||
for i := range LocalModels {
|
for i := range LocalModels {
|
||||||
@@ -1414,14 +1557,13 @@ func updateModelLists() {
|
|||||||
m := strings.TrimPrefix(LocalModels[i], models.LoadedMark)
|
m := strings.TrimPrefix(LocalModels[i], models.LoadedMark)
|
||||||
cfg.CurrentModel = m
|
cfg.CurrentModel = m
|
||||||
chatBody.Model = m
|
chatBody.Model = m
|
||||||
cachedModelColor = "green"
|
cachedModelColor.Store("green")
|
||||||
updateStatusLine()
|
updateStatusLine()
|
||||||
updateToolCapabilities()
|
UpdateToolCapabilities()
|
||||||
app.Draw()
|
app.Draw()
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func refreshLocalModelsIfEmpty() {
|
func refreshLocalModelsIfEmpty() {
|
||||||
@@ -1442,14 +1584,42 @@ func refreshLocalModelsIfEmpty() {
|
|||||||
localModelsMu.Unlock()
|
localModelsMu.Unlock()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func startNewCLIChat() []models.RoleMsg {
|
||||||
|
id, err := store.ChatGetMaxID()
|
||||||
|
if err != nil {
|
||||||
|
logger.Error("failed to get chat id", "error", err)
|
||||||
|
}
|
||||||
|
id++
|
||||||
|
charToStart(cfg.AssistantRole, false)
|
||||||
|
newChat := &models.Chat{
|
||||||
|
ID: id,
|
||||||
|
Name: fmt.Sprintf("%d_%s", id, cfg.AssistantRole),
|
||||||
|
CreatedAt: time.Now(),
|
||||||
|
UpdatedAt: time.Now(),
|
||||||
|
Msgs: "",
|
||||||
|
Agent: cfg.AssistantRole,
|
||||||
|
}
|
||||||
|
activeChatName = newChat.Name
|
||||||
|
chatMap[newChat.Name] = newChat
|
||||||
|
cliPrevOutput = ""
|
||||||
|
return chatBody.Messages
|
||||||
|
}
|
||||||
|
|
||||||
|
func startNewCLIErrors() []models.RoleMsg {
|
||||||
|
return startNewCLIChat()
|
||||||
|
}
|
||||||
|
|
||||||
func summarizeAndStartNewChat() {
|
func summarizeAndStartNewChat() {
|
||||||
if len(chatBody.Messages) == 0 {
|
if len(chatBody.Messages) == 0 {
|
||||||
showToast("info", "No chat history to summarize")
|
showToast("info", "No chat history to summarize")
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
showToast("info", "Summarizing chat history...")
|
showToast("info", "Summarizing chat history...")
|
||||||
|
arg := map[string]string{
|
||||||
|
"chat": chatToText(chatBody.Messages, false),
|
||||||
|
}
|
||||||
// Call the summarize_chat tool via agent
|
// Call the summarize_chat tool via agent
|
||||||
summaryBytes := callToolWithAgent("summarize_chat", map[string]string{})
|
summaryBytes, _ := tools.CallToolWithAgent("summarize_chat", arg)
|
||||||
summary := string(summaryBytes)
|
summary := string(summaryBytes)
|
||||||
if summary == "" {
|
if summary == "" {
|
||||||
showToast("error", "Failed to generate summary")
|
showToast("error", "Failed to generate summary")
|
||||||
@@ -1465,8 +1635,10 @@ func summarizeAndStartNewChat() {
|
|||||||
}
|
}
|
||||||
chatBody.Messages = append(chatBody.Messages, toolMsg)
|
chatBody.Messages = append(chatBody.Messages, toolMsg)
|
||||||
// Update UI
|
// Update UI
|
||||||
|
if !cfg.CLIMode {
|
||||||
textView.SetText(chatToText(chatBody.Messages, cfg.ShowSys))
|
textView.SetText(chatToText(chatBody.Messages, cfg.ShowSys))
|
||||||
colorText()
|
colorText()
|
||||||
|
}
|
||||||
// Update storage
|
// Update storage
|
||||||
if err := updateStorageChat(activeChatName, chatBody.Messages); err != nil {
|
if err := updateStorageChat(activeChatName, chatBody.Messages); err != nil {
|
||||||
logger.Warn("failed to update storage after injecting summary", "error", err)
|
logger.Warn("failed to update storage after injecting summary", "error", err)
|
||||||
@@ -1485,8 +1657,8 @@ func init() {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
defaultStarter = []models.RoleMsg{
|
defaultStarter = []models.RoleMsg{
|
||||||
{Role: "system", Content: basicSysMsg},
|
{Role: "system", Content: models.BasicSysMsg},
|
||||||
{Role: cfg.AssistantRole, Content: defaultFirstMsg},
|
{Role: cfg.AssistantRole, Content: models.DefaultFirstMsg},
|
||||||
}
|
}
|
||||||
logfile, err := os.OpenFile(cfg.LogFile,
|
logfile, err := os.OpenFile(cfg.LogFile,
|
||||||
os.O_APPEND|os.O_CREATE|os.O_WRONLY, 0644)
|
os.O_APPEND|os.O_CREATE|os.O_WRONLY, 0644)
|
||||||
@@ -1497,9 +1669,11 @@ func init() {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
// load cards
|
// load cards
|
||||||
|
sysMap[basicCard.ID] = basicCard
|
||||||
|
roleToID["assistant"] = basicCard.ID
|
||||||
basicCard.Role = cfg.AssistantRole
|
basicCard.Role = cfg.AssistantRole
|
||||||
logLevel.Set(slog.LevelInfo)
|
logLevel.Set(slog.LevelInfo)
|
||||||
logger = slog.New(slog.NewTextHandler(logfile, &slog.HandlerOptions{Level: logLevel, AddSource: true}))
|
logger = slog.New(slog.NewTextHandler(logfile, &slog.HandlerOptions{Level: logLevel}))
|
||||||
store = storage.NewProviderSQL(cfg.DBPATH, logger)
|
store = storage.NewProviderSQL(cfg.DBPATH, logger)
|
||||||
if store == nil {
|
if store == nil {
|
||||||
cancel()
|
cancel()
|
||||||
@@ -1522,7 +1696,12 @@ func init() {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
lastToolCall = &models.FuncCall{}
|
lastToolCall = &models.FuncCall{}
|
||||||
lastChat := loadOldChatOrGetNew()
|
var lastChat []models.RoleMsg
|
||||||
|
if cfg.CLIMode {
|
||||||
|
lastChat = startNewCLIErrors()
|
||||||
|
} else {
|
||||||
|
lastChat = loadOldChatOrGetNew()
|
||||||
|
}
|
||||||
chatBody = &models.ChatBody{
|
chatBody = &models.ChatBody{
|
||||||
Model: "modelname",
|
Model: "modelname",
|
||||||
Stream: true,
|
Stream: true,
|
||||||
@@ -1537,73 +1716,30 @@ func init() {
|
|||||||
asr = NewSTT(logger, cfg)
|
asr = NewSTT(logger, cfg)
|
||||||
}
|
}
|
||||||
if cfg.PlaywrightEnabled {
|
if cfg.PlaywrightEnabled {
|
||||||
if err := checkPlaywright(); err != nil {
|
go func() {
|
||||||
// slow, need a faster check if playwright install
|
if err := tools.CheckPlaywright(); err != nil {
|
||||||
if err := installPW(); err != nil {
|
if err := tools.InstallPW(); err != nil {
|
||||||
logger.Error("failed to install playwright", "error", err)
|
logger.Error("failed to install playwright", "error", err)
|
||||||
cancel()
|
cancel()
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
if err := checkPlaywright(); err != nil {
|
if err := tools.CheckPlaywright(); err != nil {
|
||||||
logger.Error("failed to run playwright", "error", err)
|
logger.Error("failed to run playwright", "error", err)
|
||||||
cancel()
|
cancel()
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
}()
|
||||||
}
|
}
|
||||||
// Initialize scrollToEndEnabled based on config
|
// atomic default values
|
||||||
scrollToEndEnabled = cfg.AutoScrollEnabled
|
cachedModelColor.Store("orange")
|
||||||
go updateModelLists()
|
|
||||||
go chatWatcher(ctx)
|
go chatWatcher(ctx)
|
||||||
}
|
if !cfg.CLIMode {
|
||||||
|
initTUI()
|
||||||
func getValidKnowToRecipient(msg *models.RoleMsg) (string, bool) {
|
}
|
||||||
if cfg == nil || !cfg.CharSpecificContextEnabled {
|
tools.InitTools(cfg, logger, store)
|
||||||
return "", false
|
// tooler = tools.InitTools(cfg, logger, store)
|
||||||
}
|
// tooler.RegisterWindowTools(modelHasVision)
|
||||||
// case where all roles are in the tag => public message
|
|
||||||
cr := listChatRoles()
|
|
||||||
slices.Sort(cr)
|
|
||||||
slices.Sort(msg.KnownTo)
|
|
||||||
if slices.Equal(cr, msg.KnownTo) {
|
|
||||||
logger.Info("got msg with tag mentioning every role")
|
|
||||||
return "", false
|
|
||||||
}
|
|
||||||
// Check each character in the KnownTo list
|
|
||||||
for _, recipient := range msg.KnownTo {
|
|
||||||
if recipient == msg.Role || recipient == cfg.ToolRole {
|
|
||||||
// weird cases, skip
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
// Skip if this is the user character (user handles their own turn)
|
|
||||||
// If user is in KnownTo, stop processing - it's the user's turn
|
|
||||||
if recipient == cfg.UserRole || recipient == cfg.WriteNextMsgAs {
|
|
||||||
return "", false
|
|
||||||
}
|
|
||||||
return recipient, true
|
|
||||||
}
|
|
||||||
return "", false
|
|
||||||
}
|
|
||||||
|
|
||||||
// triggerPrivateMessageResponses checks if a message was sent privately to specific characters
|
|
||||||
// and triggers those non-user characters to respond
|
|
||||||
func triggerPrivateMessageResponses(msg *models.RoleMsg) {
|
|
||||||
recipient, ok := getValidKnowToRecipient(msg)
|
|
||||||
if !ok || recipient == "" {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
// Trigger the recipient character to respond
|
|
||||||
triggerMsg := recipient + ":\n"
|
|
||||||
// Send empty message so LLM continues naturally from the conversation
|
|
||||||
crr := &models.ChatRoundReq{
|
|
||||||
UserMsg: triggerMsg,
|
|
||||||
Role: recipient,
|
|
||||||
Resume: true,
|
|
||||||
}
|
|
||||||
fmt.Fprintf(textView, "\n[-:-:b](%d) ", len(chatBody.Messages))
|
|
||||||
fmt.Fprint(textView, roleToIcon(recipient))
|
|
||||||
fmt.Fprint(textView, "[-:-:-]\n")
|
|
||||||
chatRoundChan <- crr
|
|
||||||
}
|
}
|
||||||
|
|||||||
74
cli-tests/sort-img/check.sh
Executable file
74
cli-tests/sort-img/check.sh
Executable file
@@ -0,0 +1,74 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -e
|
||||||
|
|
||||||
|
SCRIPT_DIR="$(cd "$(dirname "$0")" && pwd)"
|
||||||
|
LOG_FILE=$(ls -t "$SCRIPT_DIR"/*_run.log 2>/dev/null | head -1)
|
||||||
|
|
||||||
|
PASS=0
|
||||||
|
FAIL=0
|
||||||
|
|
||||||
|
log_pass() {
|
||||||
|
echo "[PASS] $1"
|
||||||
|
PASS=$((PASS + 1))
|
||||||
|
}
|
||||||
|
|
||||||
|
log_fail() {
|
||||||
|
echo "[FAIL] $1"
|
||||||
|
FAIL=$((FAIL + 1))
|
||||||
|
}
|
||||||
|
|
||||||
|
echo "=== Checking results ==="
|
||||||
|
echo ""
|
||||||
|
|
||||||
|
# Check has-animals directory exists
|
||||||
|
if [ -d "/tmp/sort-img/has-animals" ]; then
|
||||||
|
log_pass "has-animals directory exists"
|
||||||
|
else
|
||||||
|
log_fail "has-animals directory missing"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Check no-animals directory exists
|
||||||
|
if [ -d "/tmp/sort-img/no-animals" ]; then
|
||||||
|
log_pass "no-animals directory exists"
|
||||||
|
else
|
||||||
|
log_fail "no-animals directory missing"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Check has-animals contains at least one image
|
||||||
|
HAS_ANIMALS_FILES=$(ls -1 /tmp/sort-img/has-animals 2>/dev/null | wc -l)
|
||||||
|
if [ "$HAS_ANIMALS_FILES" -gt 0 ]; then
|
||||||
|
log_pass "has-animals contains images ($HAS_ANIMALS_FILES files)"
|
||||||
|
else
|
||||||
|
log_fail "has-animals is empty"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Check no-animals contains at least one image
|
||||||
|
NO_ANIMALS_FILES=$(ls -1 /tmp/sort-img/no-animals 2>/dev/null | wc -l)
|
||||||
|
if [ "$NO_ANIMALS_FILES" -gt 0 ]; then
|
||||||
|
log_pass "no-animals contains images ($NO_ANIMALS_FILES files)"
|
||||||
|
else
|
||||||
|
log_fail "no-animals is empty"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Check total files sorted correctly (3 original files should be in subdirs)
|
||||||
|
TOTAL_SORTED=$((HAS_ANIMALS_FILES + NO_ANIMALS_FILES))
|
||||||
|
if [ "$TOTAL_SORTED" -eq 3 ]; then
|
||||||
|
log_pass "all 3 files sorted into subdirectories"
|
||||||
|
else
|
||||||
|
log_fail "expected 3 files sorted, got $TOTAL_SORTED"
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo ""
|
||||||
|
echo "=== Summary ==="
|
||||||
|
echo "PASSED: $PASS"
|
||||||
|
echo "FAILED: $FAIL"
|
||||||
|
|
||||||
|
if [ $FAIL -gt 0 ]; then
|
||||||
|
echo ""
|
||||||
|
echo "Log file: $LOG_FILE"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo ""
|
||||||
|
echo "All tests passed!"
|
||||||
|
exit 0
|
||||||
25
cli-tests/sort-img/run.sh
Executable file
25
cli-tests/sort-img/run.sh
Executable file
@@ -0,0 +1,25 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -e
|
||||||
|
|
||||||
|
SCRIPT_DIR="$(cd "$(dirname "$0")" && pwd)"
|
||||||
|
TIMESTAMP=$(date +%Y%m%d_%H%M%S)
|
||||||
|
LOG_FILE="$SCRIPT_DIR/${TIMESTAMP}_run.log"
|
||||||
|
|
||||||
|
exec > "$LOG_FILE" 2>&1
|
||||||
|
|
||||||
|
echo "=== Running teardown ==="
|
||||||
|
"$SCRIPT_DIR/teardown.sh"
|
||||||
|
|
||||||
|
echo ""
|
||||||
|
echo "=== Running setup ==="
|
||||||
|
"$SCRIPT_DIR/setup.sh"
|
||||||
|
|
||||||
|
echo ""
|
||||||
|
echo "=== Running task ==="
|
||||||
|
TASK=$(cat "$SCRIPT_DIR/task.txt")
|
||||||
|
cd /home/grail/projects/plays/goplays/gf-lt
|
||||||
|
go run . -cli -msg "$TASK"
|
||||||
|
|
||||||
|
echo ""
|
||||||
|
echo "=== Done ==="
|
||||||
|
echo "Log file: $LOG_FILE"
|
||||||
9
cli-tests/sort-img/setup.sh
Executable file
9
cli-tests/sort-img/setup.sh
Executable file
@@ -0,0 +1,9 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
set -e
|
||||||
|
|
||||||
|
mkdir -p /tmp/sort-img
|
||||||
|
|
||||||
|
cp ../../../assets/ex01.png /tmp/sort-img/file1.png
|
||||||
|
cp ../../../assets/helppage.png /tmp/sort-img/file2.png
|
||||||
|
cp ../../../assets/yt_thumb.jpg /tmp/sort-img/file3.jpg
|
||||||
2
cli-tests/sort-img/task.txt
Normal file
2
cli-tests/sort-img/task.txt
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
go to /tmp/sort-img, create directories: has-animals, no-animals
|
||||||
|
sort images in /tmp/sort-img into created directories by content
|
||||||
4
cli-tests/sort-img/teardown.sh
Executable file
4
cli-tests/sort-img/teardown.sh
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -e
|
||||||
|
|
||||||
|
rm -rf /tmp/sort-img
|
||||||
91
cli-tests/sort-text/check.sh
Executable file
91
cli-tests/sort-text/check.sh
Executable file
@@ -0,0 +1,91 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -e
|
||||||
|
|
||||||
|
SCRIPT_DIR="$(cd "$(dirname "$0")" && pwd)"
|
||||||
|
LOG_FILE=$(ls -t "$SCRIPT_DIR"/*_run.log 2>/dev/null | head -1)
|
||||||
|
|
||||||
|
PASS=0
|
||||||
|
FAIL=0
|
||||||
|
|
||||||
|
log_pass() {
|
||||||
|
echo "[PASS] $1"
|
||||||
|
PASS=$((PASS + 1))
|
||||||
|
}
|
||||||
|
|
||||||
|
log_fail() {
|
||||||
|
echo "[FAIL] $1"
|
||||||
|
FAIL=$((FAIL + 1))
|
||||||
|
}
|
||||||
|
|
||||||
|
echo "=== Checking results ==="
|
||||||
|
echo ""
|
||||||
|
|
||||||
|
# Check animals directory exists
|
||||||
|
if [ -d "/tmp/sort-text/animals" ]; then
|
||||||
|
log_pass "animals directory exists"
|
||||||
|
else
|
||||||
|
log_fail "animals directory missing"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Check colors directory exists
|
||||||
|
if [ -d "/tmp/sort-text/colors" ]; then
|
||||||
|
log_pass "colors directory exists"
|
||||||
|
else
|
||||||
|
log_fail "colors directory missing"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Check animals contain cat/dog
|
||||||
|
ANIMALS_FILES=$(ls -1 /tmp/sort-text/animals 2>/dev/null | tr '\n' ' ')
|
||||||
|
if echo "$ANIMALS_FILES" | grep -q "file1.txt" && echo "$ANIMALS_FILES" | grep -q "file3.txt"; then
|
||||||
|
log_pass "animals contains animal files"
|
||||||
|
else
|
||||||
|
log_fail "animals missing animal files (got: $ANIMALS_FILES)"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Check colors contain red/blue
|
||||||
|
COLORS_FILES=$(ls -1 /tmp/sort-text/colors 2>/dev/null | tr '\n' ' ')
|
||||||
|
if echo "$COLORS_FILES" | grep -q "file2.txt" && echo "$COLORS_FILES" | grep -q "file4.txt"; then
|
||||||
|
log_pass "colors contains color files"
|
||||||
|
else
|
||||||
|
log_fail "colors missing color files (got: $COLORS_FILES)"
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Verify content
|
||||||
|
if grep -q "cat" /tmp/sort-text/animals/file1.txt 2>/dev/null; then
|
||||||
|
log_pass "file1.txt contains 'cat'"
|
||||||
|
else
|
||||||
|
log_fail "file1.txt missing 'cat'"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if grep -q "dog" /tmp/sort-text/animals/file3.txt 2>/dev/null; then
|
||||||
|
log_pass "file3.txt contains 'dog'"
|
||||||
|
else
|
||||||
|
log_fail "file3.txt missing 'dog'"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if grep -q "red" /tmp/sort-text/colors/file2.txt 2>/dev/null; then
|
||||||
|
log_pass "file2.txt contains 'red'"
|
||||||
|
else
|
||||||
|
log_fail "file2.txt missing 'red'"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if grep -q "blue" /tmp/sort-text/colors/file4.txt 2>/dev/null; then
|
||||||
|
log_pass "file4.txt contains 'blue'"
|
||||||
|
else
|
||||||
|
log_fail "file4.txt missing 'blue'"
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo ""
|
||||||
|
echo "=== Summary ==="
|
||||||
|
echo "PASSED: $PASS"
|
||||||
|
echo "FAILED: $FAIL"
|
||||||
|
|
||||||
|
if [ $FAIL -gt 0 ]; then
|
||||||
|
echo ""
|
||||||
|
echo "Log file: $LOG_FILE"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo ""
|
||||||
|
echo "All tests passed!"
|
||||||
|
exit 0
|
||||||
25
cli-tests/sort-text/run.sh
Executable file
25
cli-tests/sort-text/run.sh
Executable file
@@ -0,0 +1,25 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -e
|
||||||
|
|
||||||
|
SCRIPT_DIR="$(cd "$(dirname "$0")" && pwd)"
|
||||||
|
TIMESTAMP=$(date +%Y%m%d_%H%M%S)
|
||||||
|
LOG_FILE="$SCRIPT_DIR/${TIMESTAMP}_run.log"
|
||||||
|
|
||||||
|
exec > "$LOG_FILE" 2>&1
|
||||||
|
|
||||||
|
echo "=== Running teardown ==="
|
||||||
|
"$SCRIPT_DIR/teardown.sh"
|
||||||
|
|
||||||
|
echo ""
|
||||||
|
echo "=== Running setup ==="
|
||||||
|
"$SCRIPT_DIR/setup.sh"
|
||||||
|
|
||||||
|
echo ""
|
||||||
|
echo "=== Running task ==="
|
||||||
|
TASK=$(cat "$SCRIPT_DIR/task.txt")
|
||||||
|
cd /home/grail/projects/plays/goplays/gf-lt
|
||||||
|
go run . -cli -msg "$TASK"
|
||||||
|
|
||||||
|
echo ""
|
||||||
|
echo "=== Done ==="
|
||||||
|
echo "Log file: $LOG_FILE"
|
||||||
10
cli-tests/sort-text/setup.sh
Executable file
10
cli-tests/sort-text/setup.sh
Executable file
@@ -0,0 +1,10 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
set -e
|
||||||
|
|
||||||
|
mkdir -p /tmp/sort-text
|
||||||
|
|
||||||
|
printf "cat" > /tmp/sort-text/file1.txt
|
||||||
|
printf "red" > /tmp/sort-text/file2.txt
|
||||||
|
printf "dog" > /tmp/sort-text/file3.txt
|
||||||
|
printf "blue" > /tmp/sort-text/file4.txt
|
||||||
2
cli-tests/sort-text/task.txt
Normal file
2
cli-tests/sort-text/task.txt
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
go to /tmp/sort-text, create directories: animals, colors
|
||||||
|
sort /tmp/sort-text/*.txt into created directories by text content
|
||||||
4
cli-tests/sort-text/teardown.sh
Executable file
4
cli-tests/sort-text/teardown.sh
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -e
|
||||||
|
|
||||||
|
rm -rf /tmp/sort-text
|
||||||
@@ -18,7 +18,7 @@ EmbedTokenizerPath = "onnx/embedgemma/tokenizer.json"
|
|||||||
EmbedDims = 768
|
EmbedDims = 768
|
||||||
#
|
#
|
||||||
ShowSys = true
|
ShowSys = true
|
||||||
LogFile = "log.txt"
|
LogFile = "log.log"
|
||||||
UserRole = "user"
|
UserRole = "user"
|
||||||
ToolRole = "tool"
|
ToolRole = "tool"
|
||||||
AssistantRole = "assistant"
|
AssistantRole = "assistant"
|
||||||
@@ -28,8 +28,8 @@ AutoScrollEnabled = true
|
|||||||
AutoCleanToolCallsFromCtx = false
|
AutoCleanToolCallsFromCtx = false
|
||||||
# rag settings
|
# rag settings
|
||||||
RAGBatchSize = 1
|
RAGBatchSize = 1
|
||||||
RAGWordLimit = 80
|
RAGWordLimit = 250
|
||||||
RAGOverlapWords = 16
|
RAGOverlapWords = 25
|
||||||
RAGDir = "ragimport"
|
RAGDir = "ragimport"
|
||||||
# extra tts
|
# extra tts
|
||||||
TTS_ENABLED = false
|
TTS_ENABLED = false
|
||||||
@@ -42,7 +42,7 @@ STT_ENABLED = false
|
|||||||
STT_TYPE = "WHISPER_SERVER" # WHISPER_SERVER or WHISPER_BINARY
|
STT_TYPE = "WHISPER_SERVER" # WHISPER_SERVER or WHISPER_BINARY
|
||||||
STT_URL = "http://localhost:8081/inference"
|
STT_URL = "http://localhost:8081/inference"
|
||||||
WhisperBinaryPath = "./batteries/whisper.cpp/build/bin/whisper-cli" # Path to whisper binary (for WHISPER_BINARY mode)
|
WhisperBinaryPath = "./batteries/whisper.cpp/build/bin/whisper-cli" # Path to whisper binary (for WHISPER_BINARY mode)
|
||||||
WhisperModelPath = "./batteries/whisper.cpp/ggml-large-v3-turbo-q5_0.bin" # Path to whisper model file (for WHISPER_BINARY mode)
|
WhisperModelPath = "./batteries/whisper.cpp/models/ggml-large-v3-turbo-q5_0.bin" # Path to whisper model file (for WHISPER_BINARY mode)
|
||||||
STT_LANG = "en" # Language for speech recognition (for WHISPER_BINARY mode)
|
STT_LANG = "en" # Language for speech recognition (for WHISPER_BINARY mode)
|
||||||
STT_SR = 16000 # Sample rate for audio recording
|
STT_SR = 16000 # Sample rate for audio recording
|
||||||
#
|
#
|
||||||
@@ -61,4 +61,4 @@ StripThinkingFromAPI = true # Strip <think> blocks from messages before sending
|
|||||||
ReasoningEffort = "medium"
|
ReasoningEffort = "medium"
|
||||||
# playwright tools
|
# playwright tools
|
||||||
PlaywrightEnabled = false
|
PlaywrightEnabled = false
|
||||||
PlaywrightDebug = false
|
PlaywrightDebug = false # when true opens in gui mode (headless=false)
|
||||||
|
|||||||
@@ -27,7 +27,6 @@ type Config struct {
|
|||||||
WriteNextMsgAs string
|
WriteNextMsgAs string
|
||||||
WriteNextMsgAsCompletionAgent string
|
WriteNextMsgAsCompletionAgent string
|
||||||
SkipLLMResp bool
|
SkipLLMResp bool
|
||||||
AutoCleanToolCallsFromCtx bool `toml:"AutoCleanToolCallsFromCtx"`
|
|
||||||
DBPATH string `toml:"DBPATH"`
|
DBPATH string `toml:"DBPATH"`
|
||||||
FilePickerDir string `toml:"FilePickerDir"`
|
FilePickerDir string `toml:"FilePickerDir"`
|
||||||
FilePickerExts string `toml:"FilePickerExts"`
|
FilePickerExts string `toml:"FilePickerExts"`
|
||||||
@@ -61,10 +60,6 @@ type Config struct {
|
|||||||
TTS_SPEED float32 `toml:"TTS_SPEED"`
|
TTS_SPEED float32 `toml:"TTS_SPEED"`
|
||||||
TTS_PROVIDER string `toml:"TTS_PROVIDER"`
|
TTS_PROVIDER string `toml:"TTS_PROVIDER"`
|
||||||
TTS_LANGUAGE string `toml:"TTS_LANGUAGE"`
|
TTS_LANGUAGE string `toml:"TTS_LANGUAGE"`
|
||||||
// Kokoro ONNX TTS
|
|
||||||
KokoroModelPath string `toml:"KokoroModelPath"`
|
|
||||||
KokoroVoicesPath string `toml:"KokoroVoicesPath"`
|
|
||||||
KokoroVoice string `toml:"KokoroVoice"`
|
|
||||||
// STT
|
// STT
|
||||||
STT_TYPE string `toml:"STT_TYPE"` // WHISPER_SERVER, WHISPER_BINARY
|
STT_TYPE string `toml:"STT_TYPE"` // WHISPER_SERVER, WHISPER_BINARY
|
||||||
STT_URL string `toml:"STT_URL"`
|
STT_URL string `toml:"STT_URL"`
|
||||||
@@ -80,6 +75,8 @@ type Config struct {
|
|||||||
// playwright browser
|
// playwright browser
|
||||||
PlaywrightEnabled bool `toml:"PlaywrightEnabled"`
|
PlaywrightEnabled bool `toml:"PlaywrightEnabled"`
|
||||||
PlaywrightDebug bool `toml:"PlaywrightDebug"` // !headless
|
PlaywrightDebug bool `toml:"PlaywrightDebug"` // !headless
|
||||||
|
// CLI mode
|
||||||
|
CLIMode bool
|
||||||
}
|
}
|
||||||
|
|
||||||
func LoadConfig(fn string) (*Config, error) {
|
func LoadConfig(fn string) (*Config, error) {
|
||||||
|
|||||||
@@ -63,9 +63,6 @@ This document explains how to set up and configure the application using the `co
|
|||||||
#### AutoScrollEnabled (`true`)
|
#### AutoScrollEnabled (`true`)
|
||||||
- Whether to automatically scroll chat window while llm streams its repsonse.
|
- Whether to automatically scroll chat window while llm streams its repsonse.
|
||||||
|
|
||||||
#### AutoCleanToolCallsFromCtx (`false`)
|
|
||||||
- Whether to automatically clean tool calls from the conversation context to manage token usage.
|
|
||||||
|
|
||||||
### RAG (Retrieval Augmented Generation) Settings
|
### RAG (Retrieval Augmented Generation) Settings
|
||||||
|
|
||||||
#### EmbedURL (`"http://localhost:8082/v1/embeddings"`)
|
#### EmbedURL (`"http://localhost:8082/v1/embeddings"`)
|
||||||
|
|||||||
218
extra/google_tts.go
Normal file
218
extra/google_tts.go
Normal file
@@ -0,0 +1,218 @@
|
|||||||
|
//go:build extra
|
||||||
|
// +build extra
|
||||||
|
|
||||||
|
package extra
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"gf-lt/models"
|
||||||
|
"io"
|
||||||
|
"log/slog"
|
||||||
|
"os/exec"
|
||||||
|
"strings"
|
||||||
|
"sync"
|
||||||
|
|
||||||
|
google_translate_tts "github.com/GrailFinder/google-translate-tts"
|
||||||
|
"github.com/neurosnap/sentences/english"
|
||||||
|
)
|
||||||
|
|
||||||
|
type GoogleTranslateOrator struct {
|
||||||
|
logger *slog.Logger
|
||||||
|
mu sync.Mutex
|
||||||
|
speech *google_translate_tts.Speech
|
||||||
|
// fields for playback control
|
||||||
|
cmd *exec.Cmd
|
||||||
|
cmdMu sync.Mutex
|
||||||
|
stopCh chan struct{}
|
||||||
|
// text buffer and interrupt flag
|
||||||
|
textBuffer strings.Builder
|
||||||
|
interrupt bool
|
||||||
|
Speed float32
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *GoogleTranslateOrator) stoproutine() {
|
||||||
|
for {
|
||||||
|
<-TTSDoneChan
|
||||||
|
o.logger.Debug("orator got done signal")
|
||||||
|
o.Stop()
|
||||||
|
for len(TTSTextChan) > 0 {
|
||||||
|
<-TTSTextChan
|
||||||
|
}
|
||||||
|
o.mu.Lock()
|
||||||
|
o.textBuffer.Reset()
|
||||||
|
o.interrupt = true
|
||||||
|
o.mu.Unlock()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *GoogleTranslateOrator) readroutine() {
|
||||||
|
tokenizer, _ := english.NewSentenceTokenizer(nil)
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case chunk := <-TTSTextChan:
|
||||||
|
o.mu.Lock()
|
||||||
|
o.interrupt = false
|
||||||
|
_, err := o.textBuffer.WriteString(chunk)
|
||||||
|
if err != nil {
|
||||||
|
o.logger.Warn("failed to write to stringbuilder", "error", err)
|
||||||
|
o.mu.Unlock()
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
text := o.textBuffer.String()
|
||||||
|
sentences := tokenizer.Tokenize(text)
|
||||||
|
o.logger.Debug("adding chunk", "chunk", chunk, "text", text, "sen-len", len(sentences))
|
||||||
|
if len(sentences) <= 1 {
|
||||||
|
o.mu.Unlock()
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
completeSentences := sentences[:len(sentences)-1]
|
||||||
|
remaining := sentences[len(sentences)-1].Text
|
||||||
|
o.textBuffer.Reset()
|
||||||
|
o.textBuffer.WriteString(remaining)
|
||||||
|
o.mu.Unlock()
|
||||||
|
for _, sentence := range completeSentences {
|
||||||
|
o.mu.Lock()
|
||||||
|
interrupted := o.interrupt
|
||||||
|
o.mu.Unlock()
|
||||||
|
if interrupted {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
cleanedText := models.CleanText(sentence.Text)
|
||||||
|
if cleanedText == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
o.logger.Debug("calling Speak with sentence", "sent", cleanedText)
|
||||||
|
if err := o.Speak(cleanedText); err != nil {
|
||||||
|
o.logger.Error("tts failed", "sentence", cleanedText, "error", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case <-TTSFlushChan:
|
||||||
|
o.logger.Debug("got flushchan signal start")
|
||||||
|
// lln is done get the whole message out
|
||||||
|
if len(TTSTextChan) > 0 { // otherwise might get stuck
|
||||||
|
for chunk := range TTSTextChan {
|
||||||
|
o.mu.Lock()
|
||||||
|
_, err := o.textBuffer.WriteString(chunk)
|
||||||
|
o.mu.Unlock()
|
||||||
|
if err != nil {
|
||||||
|
o.logger.Warn("failed to write to stringbuilder", "error", err)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if len(TTSTextChan) == 0 {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
o.mu.Lock()
|
||||||
|
remaining := o.textBuffer.String()
|
||||||
|
remaining = models.CleanText(remaining)
|
||||||
|
o.textBuffer.Reset()
|
||||||
|
o.mu.Unlock()
|
||||||
|
if remaining == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
o.logger.Debug("calling Speak with remainder", "rem", remaining)
|
||||||
|
sentencesRem := tokenizer.Tokenize(remaining)
|
||||||
|
for _, rs := range sentencesRem { // to avoid dumping large volume of text
|
||||||
|
o.mu.Lock()
|
||||||
|
interrupt := o.interrupt
|
||||||
|
o.mu.Unlock()
|
||||||
|
if interrupt {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
if err := o.Speak(rs.Text); err != nil {
|
||||||
|
o.logger.Error("tts failed", "sentence", rs.Text, "error", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *GoogleTranslateOrator) GetLogger() *slog.Logger {
|
||||||
|
return o.logger
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *GoogleTranslateOrator) Speak(text string) error {
|
||||||
|
o.logger.Debug("fn: Speak is called", "text-len", len(text))
|
||||||
|
// Generate MP3 data directly as an io.Reader
|
||||||
|
reader, err := o.speech.GenerateSpeech(text)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("generate speech failed: %w", err)
|
||||||
|
}
|
||||||
|
// Wrap in io.NopCloser since GenerateSpeech returns io.Reader (no close needed)
|
||||||
|
body := io.NopCloser(reader)
|
||||||
|
defer body.Close()
|
||||||
|
// Build ffplay command with optional speed filter
|
||||||
|
args := []string{"-nodisp", "-autoexit"}
|
||||||
|
if o.Speed > 0.1 && o.Speed != 1.0 {
|
||||||
|
// atempo range is 0.5 to 2.0; you might clamp it here
|
||||||
|
args = append(args, "-af", fmt.Sprintf("atempo=%.2f", o.Speed))
|
||||||
|
}
|
||||||
|
args = append(args, "-i", "pipe:0")
|
||||||
|
cmd := exec.Command("ffplay", args...)
|
||||||
|
stdin, err := cmd.StdinPipe()
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to get stdin pipe: %w", err)
|
||||||
|
}
|
||||||
|
o.cmdMu.Lock()
|
||||||
|
o.cmd = cmd
|
||||||
|
o.stopCh = make(chan struct{})
|
||||||
|
o.cmdMu.Unlock()
|
||||||
|
if err := cmd.Start(); err != nil {
|
||||||
|
return fmt.Errorf("failed to start ffplay: %w", err)
|
||||||
|
}
|
||||||
|
copyErr := make(chan error, 1)
|
||||||
|
go func() {
|
||||||
|
_, err := io.Copy(stdin, body)
|
||||||
|
stdin.Close()
|
||||||
|
copyErr <- err
|
||||||
|
}()
|
||||||
|
done := make(chan error, 1)
|
||||||
|
go func() {
|
||||||
|
done <- cmd.Wait()
|
||||||
|
}()
|
||||||
|
select {
|
||||||
|
case <-o.stopCh:
|
||||||
|
if o.cmd != nil && o.cmd.Process != nil {
|
||||||
|
o.cmd.Process.Kill()
|
||||||
|
}
|
||||||
|
<-done
|
||||||
|
return nil
|
||||||
|
case copyErrVal := <-copyErr:
|
||||||
|
if copyErrVal != nil {
|
||||||
|
if o.cmd != nil && o.cmd.Process != nil {
|
||||||
|
o.cmd.Process.Kill()
|
||||||
|
}
|
||||||
|
<-done
|
||||||
|
return copyErrVal
|
||||||
|
}
|
||||||
|
return <-done
|
||||||
|
case err := <-done:
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *GoogleTranslateOrator) Stop() {
|
||||||
|
o.cmdMu.Lock()
|
||||||
|
defer o.cmdMu.Unlock()
|
||||||
|
// Signal any running Speak to stop
|
||||||
|
if o.stopCh != nil {
|
||||||
|
select {
|
||||||
|
case <-o.stopCh: // already closed
|
||||||
|
default:
|
||||||
|
close(o.stopCh)
|
||||||
|
}
|
||||||
|
o.stopCh = nil
|
||||||
|
}
|
||||||
|
// Kill the external player process if it's still running
|
||||||
|
if o.cmd != nil && o.cmd.Process != nil {
|
||||||
|
o.cmd.Process.Kill()
|
||||||
|
o.cmd.Wait() // clean up zombie process
|
||||||
|
o.cmd = nil
|
||||||
|
}
|
||||||
|
// Also reset text buffer and interrupt flag (with o.mu)
|
||||||
|
o.mu.Lock()
|
||||||
|
o.textBuffer.Reset()
|
||||||
|
o.interrupt = true
|
||||||
|
o.mu.Unlock()
|
||||||
|
}
|
||||||
259
extra/kokoro.go
Normal file
259
extra/kokoro.go
Normal file
@@ -0,0 +1,259 @@
|
|||||||
|
//go:build extra
|
||||||
|
// +build extra
|
||||||
|
|
||||||
|
package extra
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
"gf-lt/models"
|
||||||
|
"io"
|
||||||
|
"log/slog"
|
||||||
|
"net/http"
|
||||||
|
"os/exec"
|
||||||
|
"strings"
|
||||||
|
"sync"
|
||||||
|
|
||||||
|
"github.com/neurosnap/sentences/english"
|
||||||
|
)
|
||||||
|
|
||||||
|
type KokoroOrator struct {
|
||||||
|
logger *slog.Logger
|
||||||
|
mu sync.Mutex
|
||||||
|
URL string
|
||||||
|
Format models.AudioFormat
|
||||||
|
Stream bool
|
||||||
|
Speed float32
|
||||||
|
Language string
|
||||||
|
Voice string
|
||||||
|
// fields for playback control
|
||||||
|
cmd *exec.Cmd
|
||||||
|
cmdMu sync.Mutex
|
||||||
|
stopCh chan struct{}
|
||||||
|
// textBuffer, interrupt etc. remain the same
|
||||||
|
textBuffer strings.Builder
|
||||||
|
interrupt bool
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *KokoroOrator) GetLogger() *slog.Logger {
|
||||||
|
return o.logger
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *KokoroOrator) Speak(text string) error {
|
||||||
|
o.logger.Debug("fn: Speak is called", "text-len", len(text))
|
||||||
|
body, err := o.requestSound(text)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("request failed: %w", err)
|
||||||
|
}
|
||||||
|
defer body.Close()
|
||||||
|
cmd := exec.Command("ffplay", "-nodisp", "-autoexit", "-i", "pipe:0")
|
||||||
|
stdin, err := cmd.StdinPipe()
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to get stdin pipe: %w", err)
|
||||||
|
}
|
||||||
|
o.cmdMu.Lock()
|
||||||
|
o.cmd = cmd
|
||||||
|
o.stopCh = make(chan struct{})
|
||||||
|
o.cmdMu.Unlock()
|
||||||
|
if err := cmd.Start(); err != nil {
|
||||||
|
return fmt.Errorf("failed to start ffplay: %w", err)
|
||||||
|
}
|
||||||
|
// Copy audio in background
|
||||||
|
copyErr := make(chan error, 1)
|
||||||
|
go func() {
|
||||||
|
_, err := io.Copy(stdin, body)
|
||||||
|
stdin.Close()
|
||||||
|
copyErr <- err
|
||||||
|
}()
|
||||||
|
// Wait for player in background
|
||||||
|
done := make(chan error, 1)
|
||||||
|
go func() {
|
||||||
|
done <- cmd.Wait()
|
||||||
|
}()
|
||||||
|
// Wait for BOTH copy and player, but ensure we block until done
|
||||||
|
select {
|
||||||
|
case <-o.stopCh:
|
||||||
|
// Stop requested: kill player and wait for it to exit
|
||||||
|
if o.cmd != nil && o.cmd.Process != nil {
|
||||||
|
o.cmd.Process.Kill()
|
||||||
|
}
|
||||||
|
<-done // Wait for process to actually exit
|
||||||
|
return nil
|
||||||
|
case copyErrVal := <-copyErr:
|
||||||
|
if copyErrVal != nil {
|
||||||
|
// Copy failed: kill player and wait
|
||||||
|
if o.cmd != nil && o.cmd.Process != nil {
|
||||||
|
o.cmd.Process.Kill()
|
||||||
|
}
|
||||||
|
<-done
|
||||||
|
return copyErrVal
|
||||||
|
}
|
||||||
|
// Copy succeeded, now wait for playback to complete
|
||||||
|
return <-done
|
||||||
|
case err := <-done:
|
||||||
|
// Playback finished normally (copy must have succeeded or player would have exited early)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
func (o *KokoroOrator) requestSound(text string) (io.ReadCloser, error) {
|
||||||
|
if o.URL == "" {
|
||||||
|
return nil, fmt.Errorf("TTS URL is empty")
|
||||||
|
}
|
||||||
|
payload := map[string]interface{}{
|
||||||
|
"input": text,
|
||||||
|
"voice": o.Voice,
|
||||||
|
"response_format": o.Format,
|
||||||
|
"download_format": o.Format,
|
||||||
|
"stream": o.Stream,
|
||||||
|
"speed": o.Speed,
|
||||||
|
// "return_download_link": true,
|
||||||
|
"lang_code": o.Language,
|
||||||
|
}
|
||||||
|
payloadBytes, err := json.Marshal(payload)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("failed to marshal payload: %w", err)
|
||||||
|
}
|
||||||
|
req, err := http.NewRequest("POST", o.URL, bytes.NewBuffer(payloadBytes)) //nolint:noctx
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("failed to create request: %w", err)
|
||||||
|
}
|
||||||
|
req.Header.Set("accept", "application/json")
|
||||||
|
req.Header.Set("Content-Type", "application/json")
|
||||||
|
resp, err := http.DefaultClient.Do(req)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("request failed: %w", err)
|
||||||
|
}
|
||||||
|
if resp.StatusCode != http.StatusOK {
|
||||||
|
defer resp.Body.Close()
|
||||||
|
return nil, fmt.Errorf("unexpected status code: %d", resp.StatusCode)
|
||||||
|
}
|
||||||
|
return resp.Body, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *KokoroOrator) stoproutine() {
|
||||||
|
for {
|
||||||
|
<-TTSDoneChan
|
||||||
|
o.logger.Debug("orator got done signal")
|
||||||
|
// 1. Stop any ongoing playback (kills external player, closes stopCh)
|
||||||
|
o.Stop()
|
||||||
|
// 2. Drain any pending text chunks
|
||||||
|
for len(TTSTextChan) > 0 {
|
||||||
|
<-TTSTextChan
|
||||||
|
}
|
||||||
|
// 3. Reset internal state
|
||||||
|
o.mu.Lock()
|
||||||
|
o.textBuffer.Reset()
|
||||||
|
o.interrupt = true
|
||||||
|
o.mu.Unlock()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *KokoroOrator) Stop() {
|
||||||
|
o.cmdMu.Lock()
|
||||||
|
defer o.cmdMu.Unlock()
|
||||||
|
// Signal any running Speak to stop
|
||||||
|
if o.stopCh != nil {
|
||||||
|
select {
|
||||||
|
case <-o.stopCh: // already closed
|
||||||
|
default:
|
||||||
|
close(o.stopCh)
|
||||||
|
}
|
||||||
|
o.stopCh = nil
|
||||||
|
}
|
||||||
|
// Kill the external player process if it's still running
|
||||||
|
if o.cmd != nil && o.cmd.Process != nil {
|
||||||
|
o.cmd.Process.Kill()
|
||||||
|
o.cmd.Wait() // clean up zombie process
|
||||||
|
o.cmd = nil
|
||||||
|
}
|
||||||
|
// Also reset text buffer and interrupt flag (with o.mu)
|
||||||
|
o.mu.Lock()
|
||||||
|
o.textBuffer.Reset()
|
||||||
|
o.interrupt = true
|
||||||
|
o.mu.Unlock()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (o *KokoroOrator) readroutine() {
|
||||||
|
tokenizer, _ := english.NewSentenceTokenizer(nil)
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case chunk := <-TTSTextChan:
|
||||||
|
o.mu.Lock()
|
||||||
|
o.interrupt = false
|
||||||
|
_, err := o.textBuffer.WriteString(chunk)
|
||||||
|
if err != nil {
|
||||||
|
o.logger.Warn("failed to write to stringbuilder", "error", err)
|
||||||
|
o.mu.Unlock()
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
text := o.textBuffer.String()
|
||||||
|
sentences := tokenizer.Tokenize(text)
|
||||||
|
o.logger.Debug("adding chunk", "chunk", chunk, "text", text, "sen-len", len(sentences))
|
||||||
|
if len(sentences) <= 1 {
|
||||||
|
o.mu.Unlock()
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
completeSentences := sentences[:len(sentences)-1]
|
||||||
|
remaining := sentences[len(sentences)-1].Text
|
||||||
|
o.textBuffer.Reset()
|
||||||
|
o.textBuffer.WriteString(remaining)
|
||||||
|
o.mu.Unlock()
|
||||||
|
for _, sentence := range completeSentences {
|
||||||
|
o.mu.Lock()
|
||||||
|
interrupted := o.interrupt
|
||||||
|
o.mu.Unlock()
|
||||||
|
if interrupted {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
cleanedText := models.CleanText(sentence.Text)
|
||||||
|
if cleanedText == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
o.logger.Debug("calling Speak with sentence", "sent", cleanedText)
|
||||||
|
if err := o.Speak(cleanedText); err != nil {
|
||||||
|
o.logger.Error("tts failed", "sentence", cleanedText, "error", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case <-TTSFlushChan:
|
||||||
|
o.logger.Debug("got flushchan signal start")
|
||||||
|
// lln is done get the whole message out
|
||||||
|
if len(TTSTextChan) > 0 { // otherwise might get stuck
|
||||||
|
for chunk := range TTSTextChan {
|
||||||
|
o.mu.Lock()
|
||||||
|
_, err := o.textBuffer.WriteString(chunk)
|
||||||
|
o.mu.Unlock()
|
||||||
|
if err != nil {
|
||||||
|
o.logger.Warn("failed to write to stringbuilder", "error", err)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if len(TTSTextChan) == 0 {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// flush remaining text
|
||||||
|
o.mu.Lock()
|
||||||
|
remaining := o.textBuffer.String()
|
||||||
|
remaining = models.CleanText(remaining)
|
||||||
|
o.textBuffer.Reset()
|
||||||
|
o.mu.Unlock()
|
||||||
|
if remaining == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
o.logger.Debug("calling Speak with remainder", "rem", remaining)
|
||||||
|
sentencesRem := tokenizer.Tokenize(remaining)
|
||||||
|
for _, rs := range sentencesRem { // to avoid dumping large volume of text
|
||||||
|
o.mu.Lock()
|
||||||
|
interrupt := o.interrupt
|
||||||
|
o.mu.Unlock()
|
||||||
|
if interrupt {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
if err := o.Speak(rs.Text); err != nil {
|
||||||
|
o.logger.Error("tts failed", "sentence", rs, "error", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,421 +0,0 @@
|
|||||||
//go:build extra
|
|
||||||
// +build extra
|
|
||||||
|
|
||||||
package extra
|
|
||||||
|
|
||||||
import (
|
|
||||||
"bytes"
|
|
||||||
"fmt"
|
|
||||||
"gf-lt/models"
|
|
||||||
"gf-lt/onnx"
|
|
||||||
"log/slog"
|
|
||||||
"os/exec"
|
|
||||||
"strings"
|
|
||||||
"sync"
|
|
||||||
"time"
|
|
||||||
|
|
||||||
"github.com/gopxl/beep/v2"
|
|
||||||
"github.com/gopxl/beep/v2/speaker"
|
|
||||||
"github.com/gopxl/beep/v2/wav"
|
|
||||||
"github.com/neurosnap/sentences/english"
|
|
||||||
"github.com/yalue/onnxruntime_go"
|
|
||||||
)
|
|
||||||
|
|
||||||
// KokoroONNXOrator implements Kokoro TTS using ONNX runtime
|
|
||||||
type KokoroONNXOrator struct {
|
|
||||||
logger *slog.Logger
|
|
||||||
mu sync.Mutex
|
|
||||||
session *onnxruntime_go.DynamicAdvancedSession
|
|
||||||
phonemeMap map[string]int
|
|
||||||
espeakCmd string
|
|
||||||
voice string
|
|
||||||
speed float32
|
|
||||||
styleVector []float32
|
|
||||||
currentStream *beep.Ctrl
|
|
||||||
currentDone chan bool
|
|
||||||
textBuffer strings.Builder
|
|
||||||
interrupt bool
|
|
||||||
modelLoaded bool
|
|
||||||
modelPath string
|
|
||||||
voicesPath string
|
|
||||||
}
|
|
||||||
|
|
||||||
// Phoneme to token ID mapping from Kokoro tokenizer.json
|
|
||||||
var kokoroPhonemeMap = map[string]int{
|
|
||||||
"$": 0, ";": 1, ":": 2, ",": 3, ".": 4, "!": 5, "?": 6, "—": 9, "…": 10, "\"": 11, "(": 12, ")": 13, "“": 14, "”": 15, " ": 16, "̃": 17, "ˢ": 18, "ˤ": 19, "˦": 20, "˨": 21, "ᾝ": 22, "⭧": 23,
|
|
||||||
"A": 24, "I": 25, "O": 31, "Q": 33, "S": 35, "T": 36, "W": 39, "Y": 41, "ʲ": 42,
|
|
||||||
"a": 43, "b": 44, "c": 45, "d": 46, "e": 47, "f": 48, "h": 50, "i": 51, "j": 52, "k": 53, "l": 54, "m": 55, "n": 56, "o": 57, "p": 58, "q": 59, "r": 60, "s": 61, "t": 62, "u": 63, "v": 64, "w": 65, "x": 66, "y": 67, "z": 68,
|
|
||||||
"ɑ": 69, "ɐ": 70, "ɒ": 71, "æ": 72, "β": 75, "ɔ": 76, "ɕ": 77, "ç": 78, "ɖ": 80, "ð": 81, "˔": 82, "ə": 83, "ɚ": 85, "ɛ": 86, "ɜ": 87, "ɟ": 90, "ɡ": 92, "ɥ": 99, "ɨ": 101, "ɪ": 102, "ɝ": 103, "ɯ": 110, "ɰ": 111, "ŋ": 112, "ɳ": 113, "ɲ": 114, "ɴ": 115, "ø": 116, "ɸ": 118, "θ": 119, "œ": 120, "ɹ": 123, "ɾ": 125, "ɺ": 126, "ʁ": 128, "ɽ": 129, "ʂ": 130, "ʃ": 131, "ʈ": 132, "˧": 133, "ʊ": 135, "ʋ": 136, "ʌ": 138, "ɢ": 139, "ɣ": 140, "χ": 142, "ʎ": 143, "ʒ": 147, "ʔ": 148,
|
|
||||||
"ˈ": 156, "ˌ": 157, "ː": 158, "̰": 162, "̊": 164, "↕": 169, "→": 171, "↗": 172, "↘": 173, "ᶻ": 177,
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroONNXOrator) ensureInitialized(modelPath string) error {
|
|
||||||
if o.modelLoaded {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
o.mu.Lock()
|
|
||||||
defer o.mu.Unlock()
|
|
||||||
if o.modelLoaded {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
if modelPath == "" {
|
|
||||||
o.logger.Error("modelPath is empty, cannot load ONNX model")
|
|
||||||
return fmt.Errorf("modelPath is empty, set KokoroModelPath in config")
|
|
||||||
}
|
|
||||||
// Initialize ONNX runtime (shared with embedder)
|
|
||||||
if err := onnx.Init(); err != nil {
|
|
||||||
o.logger.Error("ONNX init failed", "error", err)
|
|
||||||
return fmt.Errorf("ONNX init failed: %w", err)
|
|
||||||
}
|
|
||||||
if onnx.HasCUDASupport() {
|
|
||||||
o.logger.Info("ONNX using CUDA")
|
|
||||||
} else {
|
|
||||||
o.logger.Info("ONNX using CPU fallback")
|
|
||||||
}
|
|
||||||
if o.phonemeMap == nil {
|
|
||||||
o.phonemeMap = kokoroPhonemeMap
|
|
||||||
}
|
|
||||||
if o.espeakCmd == "" {
|
|
||||||
o.espeakCmd = "espeak-ng"
|
|
||||||
if _, err := exec.LookPath(o.espeakCmd); err != nil {
|
|
||||||
o.espeakCmd = "espeak"
|
|
||||||
if _, err := exec.LookPath(o.espeakCmd); err != nil {
|
|
||||||
return fmt.Errorf("espeak-ng or espeak not found. Install with: sudo apt-get install espeak-ng")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
o.logger.Info("using espeak command", "cmd", o.espeakCmd)
|
|
||||||
// Load voice embedding if not already loaded
|
|
||||||
if o.styleVector == nil {
|
|
||||||
voiceName := o.voice
|
|
||||||
if voiceName == "" {
|
|
||||||
voiceName = "af_bella"
|
|
||||||
}
|
|
||||||
if o.voicesPath != "" {
|
|
||||||
styleVec, err := onnx.LoadVoice(o.voicesPath, voiceName)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Warn("failed to load voice, using zeros", "error", err, "voice", voiceName)
|
|
||||||
o.styleVector = make([]float32, 256)
|
|
||||||
} else {
|
|
||||||
// Shape is (510, 1, 256), we want the last 256 values (or first? let's use mean or just pick one)
|
|
||||||
// Actually, let's average across all 510 to get a single 256-dim vector
|
|
||||||
if len(styleVec) != 510*256 {
|
|
||||||
o.logger.Error("voice embedding has unexpected size", "len", len(styleVec))
|
|
||||||
err = fmt.Errorf("voice embedding has unexpected size", "len", len(styleVec))
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
o.styleVector = make([]float32, 256)
|
|
||||||
for i := 0; i < 256; i++ {
|
|
||||||
var sum float32
|
|
||||||
for j := 0; j < 510; j++ {
|
|
||||||
sum += styleVec[j*256+i]
|
|
||||||
}
|
|
||||||
o.styleVector[i] = sum / 510.0
|
|
||||||
}
|
|
||||||
o.logger.Info("loaded voice embedding", "voice", voiceName)
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
o.logger.Warn("no voices path configured, using zeros for style")
|
|
||||||
o.styleVector = make([]float32, 256)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
opts, err := onnx.NewSessionOptions()
|
|
||||||
if err != nil {
|
|
||||||
return fmt.Errorf("failed to create session options: %w", err)
|
|
||||||
}
|
|
||||||
defer func() { _ = opts.Destroy() }()
|
|
||||||
if onnx.HasCUDASupport() {
|
|
||||||
o.logger.Info("session options created with CUDA")
|
|
||||||
} else {
|
|
||||||
o.logger.Info("session options created with CPU")
|
|
||||||
}
|
|
||||||
session, err := onnxruntime_go.NewDynamicAdvancedSession(
|
|
||||||
modelPath,
|
|
||||||
[]string{"input_ids", "style", "speed"},
|
|
||||||
[]string{"waveform"},
|
|
||||||
opts,
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("failed to create ONNX session", "error", err)
|
|
||||||
return fmt.Errorf("failed to create ONNX session: %w", err)
|
|
||||||
}
|
|
||||||
o.session = session
|
|
||||||
o.modelLoaded = true
|
|
||||||
o.logger.Info("Kokoro ONNX model loaded successfully", "model", modelPath)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroONNXOrator) textToPhonemes(text string) (string, error) {
|
|
||||||
cmd := exec.Command(o.espeakCmd, "-x", "-q", text)
|
|
||||||
output, err := cmd.Output()
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("espeak failed", "error", err, "cmd", o.espeakCmd, "text", text)
|
|
||||||
return "", fmt.Errorf("espeak failed: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
phonemeStr := strings.TrimSpace(string(output))
|
|
||||||
return phonemeStr, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroONNXOrator) phonemesToTokens(phonemeStr string) ([]int, error) {
|
|
||||||
if phonemeStr == "" {
|
|
||||||
o.logger.Error("empty phoneme string")
|
|
||||||
return nil, fmt.Errorf("empty phoneme string")
|
|
||||||
}
|
|
||||||
// Iterate over each character in the phoneme string
|
|
||||||
tokens := make([]int, 0)
|
|
||||||
for _, ch := range phonemeStr {
|
|
||||||
chStr := string(ch)
|
|
||||||
if tokenID, ok := o.phonemeMap[chStr]; ok {
|
|
||||||
tokens = append(tokens, tokenID)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if len(tokens) == 0 {
|
|
||||||
o.logger.Error("no phonemes mapped to tokens", "phonemeStr", phonemeStr)
|
|
||||||
return nil, fmt.Errorf("no valid phonemes mapped to tokens")
|
|
||||||
}
|
|
||||||
return tokens, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroONNXOrator) generateAudio(text string) ([]float32, error) {
|
|
||||||
if err := o.ensureInitialized(o.modelPath); err != nil {
|
|
||||||
o.logger.Error("ensureInitialized failed", "error", err)
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
phonemeStr, err := o.textToPhonemes(text)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("phoneme conversion failed", "error", err)
|
|
||||||
return nil, fmt.Errorf("phoneme conversion failed: %w", err)
|
|
||||||
}
|
|
||||||
tokens, err := o.phonemesToTokens(phonemeStr)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("token conversion failed", "error", err)
|
|
||||||
return nil, fmt.Errorf("token conversion failed: %w", err)
|
|
||||||
}
|
|
||||||
if len(tokens) > 510 {
|
|
||||||
return nil, fmt.Errorf("text too long: %d tokens (max 510)", len(tokens))
|
|
||||||
}
|
|
||||||
tokens = append([]int{0}, tokens...)
|
|
||||||
tokens = append(tokens, 0)
|
|
||||||
inputIDs := make([]int64, len(tokens))
|
|
||||||
for i, t := range tokens {
|
|
||||||
inputIDs[i] = int64(t)
|
|
||||||
}
|
|
||||||
inputTensor, err := onnxruntime_go.NewTensor[int64](
|
|
||||||
onnxruntime_go.NewShape(1, int64(len(inputIDs))),
|
|
||||||
inputIDs,
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("failed to create input tensor", "error", err)
|
|
||||||
return nil, fmt.Errorf("failed to create input tensor: %w", err)
|
|
||||||
}
|
|
||||||
defer func() { _ = inputTensor.Destroy() }()
|
|
||||||
styleTensor, err := onnxruntime_go.NewTensor[float32](
|
|
||||||
onnxruntime_go.NewShape(1, 256),
|
|
||||||
o.styleVector,
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("failed to create style tensor", "error", err)
|
|
||||||
return nil, fmt.Errorf("failed to create style tensor: %w", err)
|
|
||||||
}
|
|
||||||
defer func() { _ = styleTensor.Destroy() }()
|
|
||||||
speedTensor, err := onnxruntime_go.NewTensor[float32](
|
|
||||||
onnxruntime_go.NewShape(1),
|
|
||||||
[]float32{o.speed},
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("failed to create speed tensor", "error", err)
|
|
||||||
return nil, fmt.Errorf("failed to create speed tensor: %w", err)
|
|
||||||
}
|
|
||||||
defer func() { _ = speedTensor.Destroy() }()
|
|
||||||
outputTensor, err := onnxruntime_go.NewEmptyTensor[float32](
|
|
||||||
onnxruntime_go.NewShape(1, 512),
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("failed to create output tensor", "error", err)
|
|
||||||
return nil, fmt.Errorf("failed to create output tensor: %w", err)
|
|
||||||
}
|
|
||||||
defer func() { _ = outputTensor.Destroy() }()
|
|
||||||
err = o.session.Run(
|
|
||||||
[]onnxruntime_go.Value{inputTensor, styleTensor, speedTensor},
|
|
||||||
[]onnxruntime_go.Value{outputTensor},
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("ONNX inference failed", "error", err)
|
|
||||||
return nil, fmt.Errorf("ONNX inference failed: %w", err)
|
|
||||||
}
|
|
||||||
audioData := outputTensor.GetData()
|
|
||||||
if len(audioData) == 0 {
|
|
||||||
o.logger.Error("empty audio output from ONNX")
|
|
||||||
return nil, fmt.Errorf("empty audio output")
|
|
||||||
}
|
|
||||||
audio := make([]float32, len(audioData))
|
|
||||||
copy(audio, audioData)
|
|
||||||
return audio, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroONNXOrator) Speak(text string) error {
|
|
||||||
audio, err := o.generateAudio(text)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("audio generation failed", "error", err)
|
|
||||||
return fmt.Errorf("audio generation failed: %w", err)
|
|
||||||
}
|
|
||||||
// Create streamer for encoding
|
|
||||||
encodeStreamer := beep.StreamerFunc(func(samples [][2]float64) (n int, ok bool) {
|
|
||||||
for i := range samples {
|
|
||||||
if i >= len(audio) {
|
|
||||||
return i, false
|
|
||||||
}
|
|
||||||
samples[i][0] = float64(audio[i])
|
|
||||||
samples[i][1] = float64(audio[i])
|
|
||||||
}
|
|
||||||
return len(audio), true
|
|
||||||
})
|
|
||||||
buf := &seekableBuffer{new(bytes.Buffer)}
|
|
||||||
err = wav.Encode(buf, encodeStreamer, beep.Format{
|
|
||||||
SampleRate: 24000,
|
|
||||||
NumChannels: 1,
|
|
||||||
Precision: 2,
|
|
||||||
})
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("wav encoding failed", "error", err)
|
|
||||||
return fmt.Errorf("wav encoding failed: %w", err)
|
|
||||||
}
|
|
||||||
decodedStreamer, format, err := wav.Decode(bytes.NewReader(buf.Bytes()))
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("wav decode failed", "error", err)
|
|
||||||
return fmt.Errorf("wav decode failed: %w", err)
|
|
||||||
}
|
|
||||||
defer decodedStreamer.Close()
|
|
||||||
if err := speaker.Init(format.SampleRate, format.SampleRate.N(time.Second/10)); err != nil {
|
|
||||||
o.logger.Error("speaker init failed", "error", err)
|
|
||||||
return fmt.Errorf("speaker init failed: %w", err)
|
|
||||||
}
|
|
||||||
o.logger.Info("playing audio", "sampleRate", format.SampleRate, "channels", format.NumChannels)
|
|
||||||
done := make(chan bool)
|
|
||||||
o.mu.Lock()
|
|
||||||
o.currentDone = done
|
|
||||||
o.currentStream = &beep.Ctrl{Streamer: beep.Seq(decodedStreamer, beep.Callback(func() {
|
|
||||||
o.mu.Lock()
|
|
||||||
close(done)
|
|
||||||
o.currentStream = nil
|
|
||||||
o.currentDone = nil
|
|
||||||
o.mu.Unlock()
|
|
||||||
})), Paused: false}
|
|
||||||
o.mu.Unlock()
|
|
||||||
speaker.Play(o.currentStream)
|
|
||||||
<-done
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroONNXOrator) Stop() {
|
|
||||||
speaker.Lock()
|
|
||||||
defer speaker.Unlock()
|
|
||||||
o.mu.Lock()
|
|
||||||
defer o.mu.Unlock()
|
|
||||||
if o.currentStream != nil {
|
|
||||||
o.currentStream.Streamer = nil
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroONNXOrator) GetLogger() *slog.Logger {
|
|
||||||
return o.logger
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroONNXOrator) stoproutine() {
|
|
||||||
for {
|
|
||||||
<-TTSDoneChan
|
|
||||||
o.Stop()
|
|
||||||
for len(TTSTextChan) > 0 {
|
|
||||||
<-TTSTextChan
|
|
||||||
}
|
|
||||||
o.mu.Lock()
|
|
||||||
o.textBuffer.Reset()
|
|
||||||
if o.currentDone != nil {
|
|
||||||
select {
|
|
||||||
case o.currentDone <- true:
|
|
||||||
default:
|
|
||||||
}
|
|
||||||
}
|
|
||||||
o.interrupt = true
|
|
||||||
o.mu.Unlock()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroONNXOrator) readroutine() {
|
|
||||||
tokenizer, _ := english.NewSentenceTokenizer(nil)
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case chunk := <-TTSTextChan:
|
|
||||||
o.mu.Lock()
|
|
||||||
o.interrupt = false
|
|
||||||
_, err := o.textBuffer.WriteString(chunk)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Warn("failed to write to buffer", "error", err)
|
|
||||||
o.mu.Unlock()
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
text := o.textBuffer.String()
|
|
||||||
sentences := tokenizer.Tokenize(text)
|
|
||||||
if len(sentences) <= 1 {
|
|
||||||
o.mu.Unlock()
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
completeSentences := sentences[:len(sentences)-1]
|
|
||||||
remaining := sentences[len(sentences)-1].Text
|
|
||||||
o.textBuffer.Reset()
|
|
||||||
o.textBuffer.WriteString(remaining)
|
|
||||||
o.mu.Unlock()
|
|
||||||
for _, sentence := range completeSentences {
|
|
||||||
o.mu.Lock()
|
|
||||||
interrupted := o.interrupt
|
|
||||||
o.mu.Unlock()
|
|
||||||
if interrupted {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
cleanedText := models.CleanText(sentence.Text)
|
|
||||||
if cleanedText == "" {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
o.logger.Info("KokoroONNX speak", "text", cleanedText)
|
|
||||||
if err := o.Speak(cleanedText); err != nil {
|
|
||||||
o.logger.Error("KokoroONNX tts failed", "text", cleanedText, "error", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
case <-TTSFlushChan:
|
|
||||||
if len(TTSTextChan) > 0 {
|
|
||||||
for chunk := range TTSTextChan {
|
|
||||||
o.mu.Lock()
|
|
||||||
_, err := o.textBuffer.WriteString(chunk)
|
|
||||||
o.mu.Unlock()
|
|
||||||
if err != nil {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
if len(TTSTextChan) == 0 {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
o.mu.Lock()
|
|
||||||
remaining := o.textBuffer.String()
|
|
||||||
remaining = models.CleanText(remaining)
|
|
||||||
o.textBuffer.Reset()
|
|
||||||
o.mu.Unlock()
|
|
||||||
if remaining == "" {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
sentencesRem := tokenizer.Tokenize(remaining)
|
|
||||||
for _, rs := range sentencesRem {
|
|
||||||
o.mu.Lock()
|
|
||||||
interrupt := o.interrupt
|
|
||||||
o.mu.Unlock()
|
|
||||||
if interrupt {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
if err := o.Speak(rs.Text); err != nil {
|
|
||||||
o.logger.Error("tts failed", "text", rs.Text, "error", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
132
extra/stt.go
132
extra/stt.go
@@ -6,18 +6,10 @@ package extra
|
|||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"encoding/binary"
|
"encoding/binary"
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"gf-lt/config"
|
"gf-lt/config"
|
||||||
"io"
|
"io"
|
||||||
"log/slog"
|
"log/slog"
|
||||||
"mime/multipart"
|
|
||||||
"net/http"
|
|
||||||
"regexp"
|
"regexp"
|
||||||
"strings"
|
|
||||||
"syscall"
|
|
||||||
|
|
||||||
"github.com/gordonklaus/portaudio"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
var specialRE = regexp.MustCompile(`\[.*?\]`)
|
var specialRE = regexp.MustCompile(`\[.*?\]`)
|
||||||
@@ -44,14 +36,6 @@ func NewSTT(logger *slog.Logger, cfg *config.Config) STT {
|
|||||||
return NewWhisperServer(logger, cfg)
|
return NewWhisperServer(logger, cfg)
|
||||||
}
|
}
|
||||||
|
|
||||||
type WhisperServer struct {
|
|
||||||
logger *slog.Logger
|
|
||||||
ServerURL string
|
|
||||||
SampleRate int
|
|
||||||
AudioBuffer *bytes.Buffer
|
|
||||||
recording bool
|
|
||||||
}
|
|
||||||
|
|
||||||
func NewWhisperServer(logger *slog.Logger, cfg *config.Config) *WhisperServer {
|
func NewWhisperServer(logger *slog.Logger, cfg *config.Config) *WhisperServer {
|
||||||
return &WhisperServer{
|
return &WhisperServer{
|
||||||
logger: logger,
|
logger: logger,
|
||||||
@@ -61,69 +45,6 @@ func NewWhisperServer(logger *slog.Logger, cfg *config.Config) *WhisperServer {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (stt *WhisperServer) StartRecording() error {
|
|
||||||
if err := stt.microphoneStream(stt.SampleRate); err != nil {
|
|
||||||
return fmt.Errorf("failed to init microphone: %w", err)
|
|
||||||
}
|
|
||||||
stt.recording = true
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (stt *WhisperServer) StopRecording() (string, error) {
|
|
||||||
stt.recording = false
|
|
||||||
// wait loop to finish?
|
|
||||||
if stt.AudioBuffer == nil {
|
|
||||||
err := errors.New("unexpected nil AudioBuffer")
|
|
||||||
stt.logger.Error(err.Error())
|
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
// Create WAV header first
|
|
||||||
body := &bytes.Buffer{}
|
|
||||||
writer := multipart.NewWriter(body)
|
|
||||||
// Add audio file part
|
|
||||||
part, err := writer.CreateFormFile("file", "recording.wav")
|
|
||||||
if err != nil {
|
|
||||||
stt.logger.Error("fn: StopRecording", "error", err)
|
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
// Stream directly to multipart writer: header + raw data
|
|
||||||
dataSize := stt.AudioBuffer.Len()
|
|
||||||
stt.writeWavHeader(part, dataSize)
|
|
||||||
if _, err := io.Copy(part, stt.AudioBuffer); err != nil {
|
|
||||||
stt.logger.Error("fn: StopRecording", "error", err)
|
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
// Reset buffer for next recording
|
|
||||||
stt.AudioBuffer.Reset()
|
|
||||||
// Add response format field
|
|
||||||
err = writer.WriteField("response_format", "text")
|
|
||||||
if err != nil {
|
|
||||||
stt.logger.Error("fn: StopRecording", "error", err)
|
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
if writer.Close() != nil {
|
|
||||||
stt.logger.Error("fn: StopRecording", "error", err)
|
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
// Send request
|
|
||||||
resp, err := http.Post(stt.ServerURL, writer.FormDataContentType(), body) //nolint:noctx
|
|
||||||
if err != nil {
|
|
||||||
stt.logger.Error("fn: StopRecording", "error", err)
|
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
defer resp.Body.Close()
|
|
||||||
// Read and print response
|
|
||||||
responseTextBytes, err := io.ReadAll(resp.Body)
|
|
||||||
if err != nil {
|
|
||||||
stt.logger.Error("fn: StopRecording", "error", err)
|
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
resptext := strings.TrimRight(string(responseTextBytes), "\n")
|
|
||||||
// in case there are special tokens like [_BEG_]
|
|
||||||
resptext = specialRE.ReplaceAllString(resptext, "")
|
|
||||||
return strings.TrimSpace(strings.ReplaceAll(resptext, "\n ", "\n")), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (stt *WhisperServer) writeWavHeader(w io.Writer, dataSize int) {
|
func (stt *WhisperServer) writeWavHeader(w io.Writer, dataSize int) {
|
||||||
header := make([]byte, 44)
|
header := make([]byte, 44)
|
||||||
copy(header[0:4], "RIFF")
|
copy(header[0:4], "RIFF")
|
||||||
@@ -147,56 +68,3 @@ func (stt *WhisperServer) writeWavHeader(w io.Writer, dataSize int) {
|
|||||||
func (stt *WhisperServer) IsRecording() bool {
|
func (stt *WhisperServer) IsRecording() bool {
|
||||||
return stt.recording
|
return stt.recording
|
||||||
}
|
}
|
||||||
|
|
||||||
func (stt *WhisperServer) microphoneStream(sampleRate int) error {
|
|
||||||
// Temporarily redirect stderr to suppress ALSA warnings during PortAudio init
|
|
||||||
origStderr, errDup := syscall.Dup(syscall.Stderr)
|
|
||||||
if errDup != nil {
|
|
||||||
return fmt.Errorf("failed to dup stderr: %w", errDup)
|
|
||||||
}
|
|
||||||
nullFD, err := syscall.Open("/dev/null", syscall.O_WRONLY, 0)
|
|
||||||
if err != nil {
|
|
||||||
_ = syscall.Close(origStderr) // Close the dup'd fd if open fails
|
|
||||||
return fmt.Errorf("failed to open /dev/null: %w", err)
|
|
||||||
}
|
|
||||||
// redirect stderr
|
|
||||||
_ = syscall.Dup2(nullFD, syscall.Stderr)
|
|
||||||
// Initialize PortAudio (this is where ALSA warnings occur)
|
|
||||||
defer func() {
|
|
||||||
// Restore stderr
|
|
||||||
_ = syscall.Dup2(origStderr, syscall.Stderr)
|
|
||||||
_ = syscall.Close(origStderr)
|
|
||||||
_ = syscall.Close(nullFD)
|
|
||||||
}()
|
|
||||||
if err := portaudio.Initialize(); err != nil {
|
|
||||||
return fmt.Errorf("portaudio init failed: %w", err)
|
|
||||||
}
|
|
||||||
in := make([]int16, 64)
|
|
||||||
stream, err := portaudio.OpenDefaultStream(1, 0, float64(sampleRate), len(in), in)
|
|
||||||
if err != nil {
|
|
||||||
if paErr := portaudio.Terminate(); paErr != nil {
|
|
||||||
return fmt.Errorf("failed to open microphone: %w; terminate error: %w", err, paErr)
|
|
||||||
}
|
|
||||||
return fmt.Errorf("failed to open microphone: %w", err)
|
|
||||||
}
|
|
||||||
go func(stream *portaudio.Stream) {
|
|
||||||
if err := stream.Start(); err != nil {
|
|
||||||
stt.logger.Error("microphoneStream", "error", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
for {
|
|
||||||
if !stt.IsRecording() {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if err := stream.Read(); err != nil {
|
|
||||||
stt.logger.Error("reading stream", "error", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if err := binary.Write(stt.AudioBuffer, binary.LittleEndian, in); err != nil {
|
|
||||||
stt.logger.Error("writing to buffer", "error", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}(stream)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|||||||
426
extra/tts.go
426
extra/tts.go
@@ -4,25 +4,13 @@
|
|||||||
package extra
|
package extra
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
|
||||||
"encoding/json"
|
|
||||||
"fmt"
|
|
||||||
"gf-lt/config"
|
"gf-lt/config"
|
||||||
"gf-lt/models"
|
"gf-lt/models"
|
||||||
"io"
|
|
||||||
"log/slog"
|
"log/slog"
|
||||||
"net/http"
|
|
||||||
"os"
|
"os"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
|
||||||
"time"
|
|
||||||
|
|
||||||
google_translate_tts "github.com/GrailFinder/google-translate-tts"
|
google_translate_tts "github.com/GrailFinder/google-translate-tts"
|
||||||
"github.com/GrailFinder/google-translate-tts/handlers"
|
|
||||||
"github.com/gopxl/beep/v2"
|
|
||||||
"github.com/gopxl/beep/v2/mp3"
|
|
||||||
"github.com/gopxl/beep/v2/speaker"
|
|
||||||
"github.com/neurosnap/sentences/english"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
@@ -32,14 +20,6 @@ var (
|
|||||||
// endsWithPunctuation = regexp.MustCompile(`[;.!?]$`)
|
// endsWithPunctuation = regexp.MustCompile(`[;.!?]$`)
|
||||||
)
|
)
|
||||||
|
|
||||||
type seekableBuffer struct {
|
|
||||||
*bytes.Buffer
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *seekableBuffer) Seek(offset int64, whence int) (int64, error) {
|
|
||||||
return 0, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
type Orator interface {
|
type Orator interface {
|
||||||
Speak(text string) error
|
Speak(text string) error
|
||||||
Stop()
|
Stop()
|
||||||
@@ -47,142 +27,6 @@ type Orator interface {
|
|||||||
GetLogger() *slog.Logger
|
GetLogger() *slog.Logger
|
||||||
}
|
}
|
||||||
|
|
||||||
// impl https://github.com/remsky/Kokoro-FastAPI
|
|
||||||
type KokoroOrator struct {
|
|
||||||
logger *slog.Logger
|
|
||||||
mu sync.Mutex
|
|
||||||
URL string
|
|
||||||
Format models.AudioFormat
|
|
||||||
Stream bool
|
|
||||||
Speed float32
|
|
||||||
Language string
|
|
||||||
Voice string
|
|
||||||
currentStream *beep.Ctrl // Added for playback control
|
|
||||||
currentDone chan bool
|
|
||||||
textBuffer strings.Builder
|
|
||||||
interrupt bool
|
|
||||||
// textBuffer bytes.Buffer
|
|
||||||
}
|
|
||||||
|
|
||||||
// Google Translate TTS implementation
|
|
||||||
type GoogleTranslateOrator struct {
|
|
||||||
logger *slog.Logger
|
|
||||||
mu sync.Mutex
|
|
||||||
speech *google_translate_tts.Speech
|
|
||||||
currentStream *beep.Ctrl
|
|
||||||
currentDone chan bool
|
|
||||||
textBuffer strings.Builder
|
|
||||||
interrupt bool
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroOrator) stoproutine() {
|
|
||||||
for {
|
|
||||||
<-TTSDoneChan
|
|
||||||
o.logger.Debug("orator got done signal")
|
|
||||||
o.Stop()
|
|
||||||
// drain the channel
|
|
||||||
for len(TTSTextChan) > 0 {
|
|
||||||
<-TTSTextChan
|
|
||||||
}
|
|
||||||
o.mu.Lock()
|
|
||||||
o.textBuffer.Reset()
|
|
||||||
if o.currentDone != nil {
|
|
||||||
select {
|
|
||||||
case o.currentDone <- true:
|
|
||||||
default:
|
|
||||||
// Channel might be closed, ignore
|
|
||||||
}
|
|
||||||
}
|
|
||||||
o.interrupt = true
|
|
||||||
o.mu.Unlock()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroOrator) readroutine() {
|
|
||||||
tokenizer, _ := english.NewSentenceTokenizer(nil)
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case chunk := <-TTSTextChan:
|
|
||||||
o.mu.Lock()
|
|
||||||
o.interrupt = false
|
|
||||||
_, err := o.textBuffer.WriteString(chunk)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Warn("failed to write to stringbuilder", "error", err)
|
|
||||||
o.mu.Unlock()
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
text := o.textBuffer.String()
|
|
||||||
sentences := tokenizer.Tokenize(text)
|
|
||||||
o.logger.Debug("adding chunk", "chunk", chunk, "text", text, "sen-len", len(sentences))
|
|
||||||
if len(sentences) <= 1 {
|
|
||||||
o.mu.Unlock()
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
completeSentences := sentences[:len(sentences)-1]
|
|
||||||
remaining := sentences[len(sentences)-1].Text
|
|
||||||
o.textBuffer.Reset()
|
|
||||||
o.textBuffer.WriteString(remaining)
|
|
||||||
o.mu.Unlock()
|
|
||||||
|
|
||||||
for _, sentence := range completeSentences {
|
|
||||||
o.mu.Lock()
|
|
||||||
interrupted := o.interrupt
|
|
||||||
o.mu.Unlock()
|
|
||||||
if interrupted {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
cleanedText := models.CleanText(sentence.Text)
|
|
||||||
if cleanedText == "" {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
o.logger.Debug("calling Speak with sentence", "sent", cleanedText)
|
|
||||||
if err := o.Speak(cleanedText); err != nil {
|
|
||||||
o.logger.Error("tts failed", "sentence", cleanedText, "error", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
case <-TTSFlushChan:
|
|
||||||
o.logger.Debug("got flushchan signal start")
|
|
||||||
// lln is done get the whole message out
|
|
||||||
if len(TTSTextChan) > 0 { // otherwise might get stuck
|
|
||||||
for chunk := range TTSTextChan {
|
|
||||||
o.mu.Lock()
|
|
||||||
_, err := o.textBuffer.WriteString(chunk)
|
|
||||||
o.mu.Unlock()
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Warn("failed to write to stringbuilder", "error", err)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
if len(TTSTextChan) == 0 {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// flush remaining text
|
|
||||||
o.mu.Lock()
|
|
||||||
remaining := o.textBuffer.String()
|
|
||||||
remaining = models.CleanText(remaining)
|
|
||||||
o.textBuffer.Reset()
|
|
||||||
o.mu.Unlock()
|
|
||||||
if remaining == "" {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
o.logger.Debug("calling Speak with remainder", "rem", remaining)
|
|
||||||
sentencesRem := tokenizer.Tokenize(remaining)
|
|
||||||
for _, rs := range sentencesRem { // to avoid dumping large volume of text
|
|
||||||
o.mu.Lock()
|
|
||||||
interrupt := o.interrupt
|
|
||||||
o.mu.Unlock()
|
|
||||||
if interrupt {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
if err := o.Speak(rs.Text); err != nil {
|
|
||||||
o.logger.Error("tts failed", "sentence", rs, "error", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func NewOrator(log *slog.Logger, cfg *config.Config) Orator {
|
func NewOrator(log *slog.Logger, cfg *config.Config) Orator {
|
||||||
provider := cfg.TTS_PROVIDER
|
provider := cfg.TTS_PROVIDER
|
||||||
if provider == "" {
|
if provider == "" {
|
||||||
@@ -202,18 +46,6 @@ func NewOrator(log *slog.Logger, cfg *config.Config) Orator {
|
|||||||
go orator.readroutine()
|
go orator.readroutine()
|
||||||
go orator.stoproutine()
|
go orator.stoproutine()
|
||||||
return orator
|
return orator
|
||||||
case "kokoro_onnx":
|
|
||||||
log.Info("Initializing Kokoro ONNX TTS", "modelPath", cfg.KokoroModelPath, "voicesPath", cfg.KokoroVoicesPath, "voice", cfg.KokoroVoice, "speed", cfg.TTS_SPEED)
|
|
||||||
orator := &KokoroONNXOrator{
|
|
||||||
logger: log,
|
|
||||||
modelPath: cfg.KokoroModelPath,
|
|
||||||
voicesPath: cfg.KokoroVoicesPath,
|
|
||||||
speed: cfg.TTS_SPEED,
|
|
||||||
voice: cfg.KokoroVoice,
|
|
||||||
}
|
|
||||||
go orator.readroutine()
|
|
||||||
go orator.stoproutine()
|
|
||||||
return orator
|
|
||||||
default:
|
default:
|
||||||
language := cfg.TTS_LANGUAGE
|
language := cfg.TTS_LANGUAGE
|
||||||
if language == "" {
|
if language == "" {
|
||||||
@@ -224,270 +56,14 @@ func NewOrator(log *slog.Logger, cfg *config.Config) Orator {
|
|||||||
Language: language,
|
Language: language,
|
||||||
Proxy: "", // Proxy not supported
|
Proxy: "", // Proxy not supported
|
||||||
Speed: cfg.TTS_SPEED,
|
Speed: cfg.TTS_SPEED,
|
||||||
Handler: &handlers.Beep{},
|
|
||||||
}
|
}
|
||||||
orator := &GoogleTranslateOrator{
|
orator := &GoogleTranslateOrator{
|
||||||
logger: log,
|
logger: log,
|
||||||
speech: speech,
|
speech: speech,
|
||||||
|
Speed: cfg.TTS_SPEED,
|
||||||
}
|
}
|
||||||
go orator.readroutine()
|
go orator.readroutine()
|
||||||
go orator.stoproutine()
|
go orator.stoproutine()
|
||||||
return orator
|
return orator
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (o *KokoroOrator) GetLogger() *slog.Logger {
|
|
||||||
return o.logger
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroOrator) requestSound(text string) (io.ReadCloser, error) {
|
|
||||||
if o.URL == "" {
|
|
||||||
return nil, fmt.Errorf("TTS URL is empty")
|
|
||||||
}
|
|
||||||
payload := map[string]interface{}{
|
|
||||||
"input": text,
|
|
||||||
"voice": o.Voice,
|
|
||||||
"response_format": o.Format,
|
|
||||||
"download_format": o.Format,
|
|
||||||
"stream": o.Stream,
|
|
||||||
"speed": o.Speed,
|
|
||||||
// "return_download_link": true,
|
|
||||||
"lang_code": o.Language,
|
|
||||||
}
|
|
||||||
payloadBytes, err := json.Marshal(payload)
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("failed to marshal payload: %w", err)
|
|
||||||
}
|
|
||||||
req, err := http.NewRequest("POST", o.URL, bytes.NewBuffer(payloadBytes)) //nolint:noctx
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("failed to create request: %w", err)
|
|
||||||
}
|
|
||||||
req.Header.Set("accept", "application/json")
|
|
||||||
req.Header.Set("Content-Type", "application/json")
|
|
||||||
resp, err := http.DefaultClient.Do(req)
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("request failed: %w", err)
|
|
||||||
}
|
|
||||||
if resp.StatusCode != http.StatusOK {
|
|
||||||
defer resp.Body.Close()
|
|
||||||
return nil, fmt.Errorf("unexpected status code: %d", resp.StatusCode)
|
|
||||||
}
|
|
||||||
return resp.Body, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroOrator) Speak(text string) error {
|
|
||||||
o.logger.Debug("fn: Speak is called", "text-len", len(text))
|
|
||||||
body, err := o.requestSound(text)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("request failed", "error", err)
|
|
||||||
return fmt.Errorf("request failed: %w", err)
|
|
||||||
}
|
|
||||||
defer body.Close()
|
|
||||||
// Decode the mp3 audio from response body
|
|
||||||
streamer, format, err := mp3.Decode(body)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("mp3 decode failed", "error", err)
|
|
||||||
return fmt.Errorf("mp3 decode failed: %w", err)
|
|
||||||
}
|
|
||||||
defer streamer.Close()
|
|
||||||
// here it spams with errors that speaker cannot be initialized more than once, but how would we deal with many audio records then?
|
|
||||||
if err := speaker.Init(format.SampleRate, format.SampleRate.N(time.Second/10)); err != nil {
|
|
||||||
o.logger.Debug("failed to init speaker", "error", err)
|
|
||||||
}
|
|
||||||
done := make(chan bool)
|
|
||||||
o.mu.Lock()
|
|
||||||
o.currentDone = done
|
|
||||||
o.currentStream = &beep.Ctrl{Streamer: beep.Seq(streamer, beep.Callback(func() {
|
|
||||||
o.mu.Lock()
|
|
||||||
close(done)
|
|
||||||
o.currentStream = nil
|
|
||||||
o.currentDone = nil
|
|
||||||
o.mu.Unlock()
|
|
||||||
})), Paused: false}
|
|
||||||
o.mu.Unlock()
|
|
||||||
speaker.Play(o.currentStream)
|
|
||||||
<-done
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *KokoroOrator) Stop() {
|
|
||||||
// speaker.Clear()
|
|
||||||
o.logger.Debug("attempted to stop orator", "orator", o)
|
|
||||||
speaker.Lock()
|
|
||||||
defer speaker.Unlock()
|
|
||||||
o.mu.Lock()
|
|
||||||
defer o.mu.Unlock()
|
|
||||||
if o.currentStream != nil {
|
|
||||||
// o.currentStream.Paused = true
|
|
||||||
o.currentStream.Streamer = nil
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *GoogleTranslateOrator) stoproutine() {
|
|
||||||
for {
|
|
||||||
<-TTSDoneChan
|
|
||||||
o.logger.Debug("orator got done signal")
|
|
||||||
o.Stop()
|
|
||||||
// drain the channel
|
|
||||||
for len(TTSTextChan) > 0 {
|
|
||||||
<-TTSTextChan
|
|
||||||
}
|
|
||||||
o.mu.Lock()
|
|
||||||
o.textBuffer.Reset()
|
|
||||||
if o.currentDone != nil {
|
|
||||||
select {
|
|
||||||
case o.currentDone <- true:
|
|
||||||
default:
|
|
||||||
// Channel might be closed, ignore
|
|
||||||
}
|
|
||||||
}
|
|
||||||
o.interrupt = true
|
|
||||||
o.mu.Unlock()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *GoogleTranslateOrator) readroutine() {
|
|
||||||
tokenizer, _ := english.NewSentenceTokenizer(nil)
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case chunk := <-TTSTextChan:
|
|
||||||
o.mu.Lock()
|
|
||||||
o.interrupt = false
|
|
||||||
_, err := o.textBuffer.WriteString(chunk)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Warn("failed to write to stringbuilder", "error", err)
|
|
||||||
o.mu.Unlock()
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
text := o.textBuffer.String()
|
|
||||||
sentences := tokenizer.Tokenize(text)
|
|
||||||
o.logger.Debug("adding chunk", "chunk", chunk, "text", text, "sen-len", len(sentences))
|
|
||||||
if len(sentences) <= 1 {
|
|
||||||
o.mu.Unlock()
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
completeSentences := sentences[:len(sentences)-1]
|
|
||||||
remaining := sentences[len(sentences)-1].Text
|
|
||||||
o.textBuffer.Reset()
|
|
||||||
o.textBuffer.WriteString(remaining)
|
|
||||||
o.mu.Unlock()
|
|
||||||
|
|
||||||
for _, sentence := range completeSentences {
|
|
||||||
o.mu.Lock()
|
|
||||||
interrupted := o.interrupt
|
|
||||||
o.mu.Unlock()
|
|
||||||
if interrupted {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
cleanedText := models.CleanText(sentence.Text)
|
|
||||||
if cleanedText == "" {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
o.logger.Debug("calling Speak with sentence", "sent", cleanedText)
|
|
||||||
if err := o.Speak(cleanedText); err != nil {
|
|
||||||
o.logger.Error("tts failed", "sentence", cleanedText, "error", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
case <-TTSFlushChan:
|
|
||||||
o.logger.Debug("got flushchan signal start")
|
|
||||||
// lln is done get the whole message out
|
|
||||||
if len(TTSTextChan) > 0 { // otherwise might get stuck
|
|
||||||
for chunk := range TTSTextChan {
|
|
||||||
o.mu.Lock()
|
|
||||||
_, err := o.textBuffer.WriteString(chunk)
|
|
||||||
o.mu.Unlock()
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Warn("failed to write to stringbuilder", "error", err)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
if len(TTSTextChan) == 0 {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
o.mu.Lock()
|
|
||||||
remaining := o.textBuffer.String()
|
|
||||||
remaining = models.CleanText(remaining)
|
|
||||||
o.textBuffer.Reset()
|
|
||||||
o.mu.Unlock()
|
|
||||||
if remaining == "" {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
o.logger.Debug("calling Speak with remainder", "rem", remaining)
|
|
||||||
sentencesRem := tokenizer.Tokenize(remaining)
|
|
||||||
for _, rs := range sentencesRem { // to avoid dumping large volume of text
|
|
||||||
o.mu.Lock()
|
|
||||||
interrupt := o.interrupt
|
|
||||||
o.mu.Unlock()
|
|
||||||
if interrupt {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
if err := o.Speak(rs.Text); err != nil {
|
|
||||||
o.logger.Error("tts failed", "sentence", rs.Text, "error", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *GoogleTranslateOrator) GetLogger() *slog.Logger {
|
|
||||||
return o.logger
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *GoogleTranslateOrator) Speak(text string) error {
|
|
||||||
o.logger.Debug("fn: Speak is called", "text-len", len(text))
|
|
||||||
// Generate MP3 data using google-translate-tts
|
|
||||||
reader, err := o.speech.GenerateSpeech(text)
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("generate speech failed", "error", err)
|
|
||||||
return fmt.Errorf("generate speech failed: %w", err)
|
|
||||||
}
|
|
||||||
// Decode the mp3 audio from reader (wrap with NopCloser for io.ReadCloser)
|
|
||||||
streamer, format, err := mp3.Decode(io.NopCloser(reader))
|
|
||||||
if err != nil {
|
|
||||||
o.logger.Error("mp3 decode failed", "error", err)
|
|
||||||
return fmt.Errorf("mp3 decode failed: %w", err)
|
|
||||||
}
|
|
||||||
defer streamer.Close()
|
|
||||||
playbackStreamer := beep.Streamer(streamer)
|
|
||||||
speed := o.speech.Speed
|
|
||||||
if speed <= 0 {
|
|
||||||
speed = 1.0
|
|
||||||
}
|
|
||||||
if speed != 1.0 {
|
|
||||||
playbackStreamer = beep.ResampleRatio(3, float64(speed), streamer)
|
|
||||||
}
|
|
||||||
// Initialize speaker with the format's sample rate
|
|
||||||
if err := speaker.Init(format.SampleRate, format.SampleRate.N(time.Second/10)); err != nil {
|
|
||||||
o.logger.Debug("failed to init speaker", "error", err)
|
|
||||||
}
|
|
||||||
done := make(chan bool)
|
|
||||||
o.mu.Lock()
|
|
||||||
o.currentDone = done
|
|
||||||
o.currentStream = &beep.Ctrl{Streamer: beep.Seq(playbackStreamer, beep.Callback(func() {
|
|
||||||
o.mu.Lock()
|
|
||||||
close(done)
|
|
||||||
o.currentStream = nil
|
|
||||||
o.currentDone = nil
|
|
||||||
o.mu.Unlock()
|
|
||||||
})), Paused: false}
|
|
||||||
o.mu.Unlock()
|
|
||||||
speaker.Play(o.currentStream)
|
|
||||||
<-done // wait for playback to complete
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (o *GoogleTranslateOrator) Stop() {
|
|
||||||
o.logger.Debug("attempted to stop google translate orator")
|
|
||||||
speaker.Lock()
|
|
||||||
defer speaker.Unlock()
|
|
||||||
o.mu.Lock()
|
|
||||||
defer o.mu.Unlock()
|
|
||||||
if o.currentStream != nil {
|
|
||||||
o.currentStream.Streamer = nil
|
|
||||||
}
|
|
||||||
// Also stop the speech handler if possible
|
|
||||||
if o.speech != nil {
|
|
||||||
_ = o.speech.Stop()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -9,15 +9,13 @@ import (
|
|||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"gf-lt/config"
|
"gf-lt/config"
|
||||||
"io"
|
|
||||||
"log/slog"
|
"log/slog"
|
||||||
"os"
|
"os"
|
||||||
"os/exec"
|
"os/exec"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
"syscall"
|
"syscall"
|
||||||
|
"time"
|
||||||
"github.com/gordonklaus/portaudio"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
type WhisperBinary struct {
|
type WhisperBinary struct {
|
||||||
@@ -25,11 +23,143 @@ type WhisperBinary struct {
|
|||||||
whisperPath string
|
whisperPath string
|
||||||
modelPath string
|
modelPath string
|
||||||
lang string
|
lang string
|
||||||
ctx context.Context
|
// Per-recording fields (protected by mu)
|
||||||
cancel context.CancelFunc
|
|
||||||
mu sync.Mutex
|
mu sync.Mutex
|
||||||
recording bool
|
recording bool
|
||||||
audioBuffer []int16
|
tempFile string
|
||||||
|
ctx context.Context
|
||||||
|
cancel context.CancelFunc
|
||||||
|
cmd *exec.Cmd
|
||||||
|
cmdMu sync.Mutex
|
||||||
|
}
|
||||||
|
|
||||||
|
func (w *WhisperBinary) StartRecording() error {
|
||||||
|
w.mu.Lock()
|
||||||
|
defer w.mu.Unlock()
|
||||||
|
if w.recording {
|
||||||
|
return errors.New("recording is already in progress")
|
||||||
|
}
|
||||||
|
// Fresh context for this recording
|
||||||
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
|
w.ctx = ctx
|
||||||
|
w.cancel = cancel
|
||||||
|
// Create temporary file
|
||||||
|
tempFile, err := os.CreateTemp("", "recording_*.wav")
|
||||||
|
if err != nil {
|
||||||
|
cancel()
|
||||||
|
return fmt.Errorf("failed to create temp file: %w", err)
|
||||||
|
}
|
||||||
|
tempFile.Close()
|
||||||
|
w.tempFile = tempFile.Name()
|
||||||
|
// ffmpeg command: capture from default microphone, write WAV
|
||||||
|
args := []string{
|
||||||
|
"-f", "alsa", // or "pulse" if preferred
|
||||||
|
"-i", "default",
|
||||||
|
"-acodec", "pcm_s16le",
|
||||||
|
"-ar", "16000",
|
||||||
|
"-ac", "1",
|
||||||
|
"-y", // overwrite output file
|
||||||
|
w.tempFile,
|
||||||
|
}
|
||||||
|
cmd := exec.CommandContext(w.ctx, "ffmpeg", args...)
|
||||||
|
// Capture stderr for debugging (optional, but useful for diagnosing)
|
||||||
|
stderr, err := cmd.StderrPipe()
|
||||||
|
if err != nil {
|
||||||
|
cancel()
|
||||||
|
os.Remove(w.tempFile)
|
||||||
|
return fmt.Errorf("failed to create stderr pipe: %w", err)
|
||||||
|
}
|
||||||
|
go func() {
|
||||||
|
buf := make([]byte, 1024)
|
||||||
|
for {
|
||||||
|
n, err := stderr.Read(buf)
|
||||||
|
if n > 0 {
|
||||||
|
w.logger.Debug("ffmpeg stderr", "output", string(buf[:n]))
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
w.cmdMu.Lock()
|
||||||
|
w.cmd = cmd
|
||||||
|
w.cmdMu.Unlock()
|
||||||
|
if err := cmd.Start(); err != nil {
|
||||||
|
cancel()
|
||||||
|
os.Remove(w.tempFile)
|
||||||
|
return fmt.Errorf("failed to start ffmpeg: %w", err)
|
||||||
|
}
|
||||||
|
w.recording = true
|
||||||
|
w.logger.Debug("Recording started", "file", w.tempFile)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (w *WhisperBinary) StopRecording() (string, error) {
|
||||||
|
w.mu.Lock()
|
||||||
|
defer w.mu.Unlock()
|
||||||
|
if !w.recording {
|
||||||
|
return "", errors.New("not currently recording")
|
||||||
|
}
|
||||||
|
w.recording = false
|
||||||
|
// Gracefully stop ffmpeg
|
||||||
|
w.cmdMu.Lock()
|
||||||
|
if w.cmd != nil && w.cmd.Process != nil {
|
||||||
|
w.logger.Debug("Sending SIGTERM to ffmpeg")
|
||||||
|
w.cmd.Process.Signal(syscall.SIGTERM)
|
||||||
|
// Wait for process to exit (up to 2 seconds)
|
||||||
|
done := make(chan error, 1)
|
||||||
|
go func() {
|
||||||
|
done <- w.cmd.Wait()
|
||||||
|
}()
|
||||||
|
select {
|
||||||
|
case <-done:
|
||||||
|
w.logger.Debug("ffmpeg exited after SIGTERM")
|
||||||
|
case <-time.After(2 * time.Second):
|
||||||
|
w.logger.Warn("ffmpeg did not exit, sending SIGKILL")
|
||||||
|
w.cmd.Process.Kill()
|
||||||
|
<-done
|
||||||
|
}
|
||||||
|
}
|
||||||
|
w.cmdMu.Unlock()
|
||||||
|
// Cancel context (already done, but for cleanliness)
|
||||||
|
if w.cancel != nil {
|
||||||
|
w.cancel()
|
||||||
|
}
|
||||||
|
// Validate temp file
|
||||||
|
if w.tempFile == "" {
|
||||||
|
return "", errors.New("no recording file")
|
||||||
|
}
|
||||||
|
defer os.Remove(w.tempFile)
|
||||||
|
info, err := os.Stat(w.tempFile)
|
||||||
|
if err != nil {
|
||||||
|
return "", fmt.Errorf("failed to stat temp file: %w", err)
|
||||||
|
}
|
||||||
|
if info.Size() < 44 { // WAV header is 44 bytes
|
||||||
|
// Log ffmpeg stderr? Already captured in debug logs.
|
||||||
|
return "", fmt.Errorf("recording file too small (%d bytes), possibly no audio captured", info.Size())
|
||||||
|
}
|
||||||
|
// Run whisper.cpp binary
|
||||||
|
cmd := exec.Command(w.whisperPath, "-m", w.modelPath, "-l", w.lang, w.tempFile)
|
||||||
|
var outBuf, errBuf bytes.Buffer
|
||||||
|
cmd.Stdout = &outBuf
|
||||||
|
cmd.Stderr = &errBuf
|
||||||
|
if err := cmd.Run(); err != nil {
|
||||||
|
w.logger.Error("whisper binary failed",
|
||||||
|
"error", err,
|
||||||
|
"stderr", errBuf.String(),
|
||||||
|
"file_size", info.Size())
|
||||||
|
return "", fmt.Errorf("whisper binary failed: %w (stderr: %s)", err, errBuf.String())
|
||||||
|
}
|
||||||
|
result := strings.TrimRight(outBuf.String(), "\n")
|
||||||
|
result = specialRE.ReplaceAllString(result, "")
|
||||||
|
return strings.TrimSpace(strings.ReplaceAll(result, "\n ", "\n")), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// IsRecording returns true if a recording is in progress.
|
||||||
|
func (w *WhisperBinary) IsRecording() bool {
|
||||||
|
w.mu.Lock()
|
||||||
|
defer w.mu.Unlock()
|
||||||
|
return w.recording
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewWhisperBinary(logger *slog.Logger, cfg *config.Config) *WhisperBinary {
|
func NewWhisperBinary(logger *slog.Logger, cfg *config.Config) *WhisperBinary {
|
||||||
@@ -44,283 +174,3 @@ func NewWhisperBinary(logger *slog.Logger, cfg *config.Config) *WhisperBinary {
|
|||||||
cancel: cancel,
|
cancel: cancel,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (w *WhisperBinary) StartRecording() error {
|
|
||||||
w.mu.Lock()
|
|
||||||
defer w.mu.Unlock()
|
|
||||||
if w.recording {
|
|
||||||
return errors.New("recording is already in progress")
|
|
||||||
}
|
|
||||||
// If context is cancelled, create a new one for the next recording session
|
|
||||||
if w.ctx.Err() != nil {
|
|
||||||
w.logger.Debug("Context cancelled, creating new context")
|
|
||||||
w.ctx, w.cancel = context.WithCancel(context.Background())
|
|
||||||
}
|
|
||||||
// Temporarily redirect stderr to suppress ALSA warnings during PortAudio init
|
|
||||||
origStderr, errDup := syscall.Dup(syscall.Stderr)
|
|
||||||
if errDup != nil {
|
|
||||||
return fmt.Errorf("failed to dup stderr: %w", errDup)
|
|
||||||
}
|
|
||||||
nullFD, err := syscall.Open("/dev/null", syscall.O_WRONLY, 0)
|
|
||||||
if err != nil {
|
|
||||||
_ = syscall.Close(origStderr) // Close the dup'd fd if open fails
|
|
||||||
return fmt.Errorf("failed to open /dev/null: %w", err)
|
|
||||||
}
|
|
||||||
// redirect stderr
|
|
||||||
_ = syscall.Dup2(nullFD, syscall.Stderr)
|
|
||||||
// Initialize PortAudio (this is where ALSA warnings occur)
|
|
||||||
portaudioErr := portaudio.Initialize()
|
|
||||||
defer func() {
|
|
||||||
// Restore stderr
|
|
||||||
_ = syscall.Dup2(origStderr, syscall.Stderr)
|
|
||||||
_ = syscall.Close(origStderr)
|
|
||||||
_ = syscall.Close(nullFD)
|
|
||||||
}()
|
|
||||||
if portaudioErr != nil {
|
|
||||||
return fmt.Errorf("portaudio init failed: %w", portaudioErr)
|
|
||||||
}
|
|
||||||
// Initialize audio buffer
|
|
||||||
w.audioBuffer = make([]int16, 0)
|
|
||||||
in := make([]int16, 1024) // buffer size
|
|
||||||
stream, err := portaudio.OpenDefaultStream(1, 0, 16000.0, len(in), in)
|
|
||||||
if err != nil {
|
|
||||||
if paErr := portaudio.Terminate(); paErr != nil {
|
|
||||||
return fmt.Errorf("failed to open microphone: %w; terminate error: %w", err, paErr)
|
|
||||||
}
|
|
||||||
return fmt.Errorf("failed to open microphone: %w", err)
|
|
||||||
}
|
|
||||||
go w.recordAudio(stream, in)
|
|
||||||
w.recording = true
|
|
||||||
w.logger.Debug("Recording started")
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (w *WhisperBinary) recordAudio(stream *portaudio.Stream, in []int16) {
|
|
||||||
defer func() {
|
|
||||||
w.logger.Debug("recordAudio defer function called")
|
|
||||||
_ = stream.Stop() // Stop the stream
|
|
||||||
_ = portaudio.Terminate() // ignoring error as we're shutting down
|
|
||||||
w.logger.Debug("recordAudio terminated")
|
|
||||||
}()
|
|
||||||
w.logger.Debug("Starting audio stream")
|
|
||||||
if err := stream.Start(); err != nil {
|
|
||||||
w.logger.Error("Failed to start audio stream", "error", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
w.logger.Debug("Audio stream started, entering recording loop")
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case <-w.ctx.Done():
|
|
||||||
w.logger.Debug("Context done, exiting recording loop")
|
|
||||||
return
|
|
||||||
default:
|
|
||||||
// Check recording status with minimal lock time
|
|
||||||
w.mu.Lock()
|
|
||||||
recording := w.recording
|
|
||||||
w.mu.Unlock()
|
|
||||||
|
|
||||||
if !recording {
|
|
||||||
w.logger.Debug("Recording flag is false, exiting recording loop")
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if err := stream.Read(); err != nil {
|
|
||||||
w.logger.Error("Error reading from stream", "error", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
// Append samples to buffer - only acquire lock when necessary
|
|
||||||
w.mu.Lock()
|
|
||||||
if w.audioBuffer == nil {
|
|
||||||
w.audioBuffer = make([]int16, 0)
|
|
||||||
}
|
|
||||||
// Make a copy of the input buffer to avoid overwriting
|
|
||||||
tempBuffer := make([]int16, len(in))
|
|
||||||
copy(tempBuffer, in)
|
|
||||||
w.audioBuffer = append(w.audioBuffer, tempBuffer...)
|
|
||||||
w.mu.Unlock()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (w *WhisperBinary) StopRecording() (string, error) {
|
|
||||||
w.logger.Debug("StopRecording called")
|
|
||||||
w.mu.Lock()
|
|
||||||
if !w.recording {
|
|
||||||
w.mu.Unlock()
|
|
||||||
return "", errors.New("not currently recording")
|
|
||||||
}
|
|
||||||
w.logger.Debug("Setting recording to false and cancelling context")
|
|
||||||
w.recording = false
|
|
||||||
w.cancel() // This will stop the recording goroutine
|
|
||||||
w.mu.Unlock()
|
|
||||||
// // Small delay to allow the recording goroutine to react to context cancellation
|
|
||||||
// time.Sleep(20 * time.Millisecond)
|
|
||||||
// Save the recorded audio to a temporary file
|
|
||||||
tempFile, err := w.saveAudioToTempFile()
|
|
||||||
if err != nil {
|
|
||||||
w.logger.Error("Error saving audio to temp file", "error", err)
|
|
||||||
return "", fmt.Errorf("failed to save audio to temp file: %w", err)
|
|
||||||
}
|
|
||||||
w.logger.Debug("Saved audio to temp file", "file", tempFile)
|
|
||||||
// Run the whisper binary with a separate context to avoid cancellation during transcription
|
|
||||||
cmd := exec.Command(w.whisperPath, "-m", w.modelPath, "-l", w.lang, tempFile, "2>/dev/null")
|
|
||||||
var outBuf bytes.Buffer
|
|
||||||
cmd.Stdout = &outBuf
|
|
||||||
// Redirect stderr to suppress ALSA warnings and other stderr output
|
|
||||||
cmd.Stderr = io.Discard // Suppress stderr output from whisper binary
|
|
||||||
w.logger.Debug("Running whisper binary command")
|
|
||||||
if err := cmd.Run(); err != nil {
|
|
||||||
// Clean up audio buffer
|
|
||||||
w.mu.Lock()
|
|
||||||
w.audioBuffer = nil
|
|
||||||
w.mu.Unlock()
|
|
||||||
// Since we're suppressing stderr, we'll just log that the command failed
|
|
||||||
w.logger.Error("Error running whisper binary", "error", err)
|
|
||||||
return "", fmt.Errorf("whisper binary failed: %w", err)
|
|
||||||
}
|
|
||||||
result := outBuf.String()
|
|
||||||
w.logger.Debug("Whisper binary completed", "result", result)
|
|
||||||
// Clean up audio buffer
|
|
||||||
w.mu.Lock()
|
|
||||||
w.audioBuffer = nil
|
|
||||||
w.mu.Unlock()
|
|
||||||
// Clean up the temporary file after transcription
|
|
||||||
w.logger.Debug("StopRecording completed")
|
|
||||||
os.Remove(tempFile)
|
|
||||||
result = strings.TrimRight(result, "\n")
|
|
||||||
// in case there are special tokens like [_BEG_]
|
|
||||||
result = specialRE.ReplaceAllString(result, "")
|
|
||||||
return strings.TrimSpace(strings.ReplaceAll(result, "\n ", "\n")), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// saveAudioToTempFile saves the recorded audio data to a temporary WAV file
|
|
||||||
func (w *WhisperBinary) saveAudioToTempFile() (string, error) {
|
|
||||||
w.logger.Debug("saveAudioToTempFile called")
|
|
||||||
// Create temporary WAV file
|
|
||||||
tempFile, err := os.CreateTemp("", "recording_*.wav")
|
|
||||||
if err != nil {
|
|
||||||
w.logger.Error("Failed to create temp file", "error", err)
|
|
||||||
return "", fmt.Errorf("failed to create temp file: %w", err)
|
|
||||||
}
|
|
||||||
w.logger.Debug("Created temp file", "file", tempFile.Name())
|
|
||||||
defer tempFile.Close()
|
|
||||||
|
|
||||||
// Write WAV header and data
|
|
||||||
w.logger.Debug("About to write WAV file", "file", tempFile.Name())
|
|
||||||
err = w.writeWAVFile(tempFile.Name())
|
|
||||||
if err != nil {
|
|
||||||
w.logger.Error("Error writing WAV file", "error", err)
|
|
||||||
return "", fmt.Errorf("failed to write WAV file: %w", err)
|
|
||||||
}
|
|
||||||
w.logger.Debug("WAV file written successfully", "file", tempFile.Name())
|
|
||||||
|
|
||||||
return tempFile.Name(), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// writeWAVFile creates a WAV file from the recorded audio data
|
|
||||||
func (w *WhisperBinary) writeWAVFile(filename string) error {
|
|
||||||
w.logger.Debug("writeWAVFile called", "filename", filename)
|
|
||||||
// Open file for writing
|
|
||||||
file, err := os.Create(filename)
|
|
||||||
if err != nil {
|
|
||||||
w.logger.Error("Error creating file", "error", err)
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
defer file.Close()
|
|
||||||
|
|
||||||
w.logger.Debug("About to acquire mutex in writeWAVFile")
|
|
||||||
w.mu.Lock()
|
|
||||||
w.logger.Debug("Locked mutex, copying audio buffer")
|
|
||||||
audioData := make([]int16, len(w.audioBuffer))
|
|
||||||
copy(audioData, w.audioBuffer)
|
|
||||||
w.mu.Unlock()
|
|
||||||
w.logger.Debug("Unlocked mutex", "audio_data_length", len(audioData))
|
|
||||||
|
|
||||||
if len(audioData) == 0 {
|
|
||||||
w.logger.Warn("No audio data to write")
|
|
||||||
return errors.New("no audio data to write")
|
|
||||||
}
|
|
||||||
|
|
||||||
// Calculate data size (number of samples * size of int16)
|
|
||||||
dataSize := len(audioData) * 2 // 2 bytes per int16 sample
|
|
||||||
w.logger.Debug("Calculated data size", "size", dataSize)
|
|
||||||
|
|
||||||
// Write WAV header with the correct data size
|
|
||||||
header := w.createWAVHeader(16000, 1, 16, dataSize)
|
|
||||||
_, err = file.Write(header)
|
|
||||||
if err != nil {
|
|
||||||
w.logger.Error("Error writing WAV header", "error", err)
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
w.logger.Debug("WAV header written successfully")
|
|
||||||
|
|
||||||
// Write audio data
|
|
||||||
w.logger.Debug("About to write audio data samples")
|
|
||||||
for i, sample := range audioData {
|
|
||||||
// Write little-endian 16-bit sample
|
|
||||||
_, err := file.Write([]byte{byte(sample), byte(sample >> 8)})
|
|
||||||
if err != nil {
|
|
||||||
w.logger.Error("Error writing sample", "index", i, "error", err)
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
// Log progress every 10000 samples to avoid too much output
|
|
||||||
if i%10000 == 0 {
|
|
||||||
w.logger.Debug("Written samples", "count", i)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
w.logger.Debug("All audio data written successfully")
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// createWAVHeader creates a WAV file header
|
|
||||||
func (w *WhisperBinary) createWAVHeader(sampleRate, channels, bitsPerSample int, dataSize int) []byte {
|
|
||||||
header := make([]byte, 44)
|
|
||||||
copy(header[0:4], "RIFF")
|
|
||||||
// Total file size will be updated later
|
|
||||||
copy(header[8:12], "WAVE")
|
|
||||||
copy(header[12:16], "fmt ")
|
|
||||||
// fmt chunk size (16 for PCM)
|
|
||||||
header[16] = 16
|
|
||||||
header[17] = 0
|
|
||||||
header[18] = 0
|
|
||||||
header[19] = 0
|
|
||||||
// Audio format (1 = PCM)
|
|
||||||
header[20] = 1
|
|
||||||
header[21] = 0
|
|
||||||
// Number of channels
|
|
||||||
header[22] = byte(channels)
|
|
||||||
header[23] = 0
|
|
||||||
// Sample rate
|
|
||||||
header[24] = byte(sampleRate)
|
|
||||||
header[25] = byte(sampleRate >> 8)
|
|
||||||
header[26] = byte(sampleRate >> 16)
|
|
||||||
header[27] = byte(sampleRate >> 24)
|
|
||||||
// Byte rate
|
|
||||||
byteRate := sampleRate * channels * bitsPerSample / 8
|
|
||||||
header[28] = byte(byteRate)
|
|
||||||
header[29] = byte(byteRate >> 8)
|
|
||||||
header[30] = byte(byteRate >> 16)
|
|
||||||
header[31] = byte(byteRate >> 24)
|
|
||||||
// Block align
|
|
||||||
blockAlign := channels * bitsPerSample / 8
|
|
||||||
header[32] = byte(blockAlign)
|
|
||||||
header[33] = 0
|
|
||||||
// Bits per sample
|
|
||||||
header[34] = byte(bitsPerSample)
|
|
||||||
header[35] = 0
|
|
||||||
// "data" subchunk
|
|
||||||
copy(header[36:40], "data")
|
|
||||||
// Data size
|
|
||||||
header[40] = byte(dataSize)
|
|
||||||
header[41] = byte(dataSize >> 8)
|
|
||||||
header[42] = byte(dataSize >> 16)
|
|
||||||
header[43] = byte(dataSize >> 24)
|
|
||||||
|
|
||||||
return header
|
|
||||||
}
|
|
||||||
|
|
||||||
func (w *WhisperBinary) IsRecording() bool {
|
|
||||||
w.mu.Lock()
|
|
||||||
defer w.mu.Unlock()
|
|
||||||
return w.recording
|
|
||||||
}
|
|
||||||
|
|||||||
156
extra/whisper_server.go
Normal file
156
extra/whisper_server.go
Normal file
@@ -0,0 +1,156 @@
|
|||||||
|
//go:build extra
|
||||||
|
// +build extra
|
||||||
|
|
||||||
|
package extra
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"log/slog"
|
||||||
|
"mime/multipart"
|
||||||
|
"net/http"
|
||||||
|
"os/exec"
|
||||||
|
"strings"
|
||||||
|
"sync"
|
||||||
|
)
|
||||||
|
|
||||||
|
type WhisperServer struct {
|
||||||
|
logger *slog.Logger
|
||||||
|
ServerURL string
|
||||||
|
SampleRate int
|
||||||
|
AudioBuffer *bytes.Buffer
|
||||||
|
recording bool // protected by mu
|
||||||
|
mu sync.Mutex // protects recording & AudioBuffer
|
||||||
|
cmd *exec.Cmd // protected by cmdMu
|
||||||
|
stopCh chan struct{} // protected by cmdMu
|
||||||
|
cmdMu sync.Mutex // protects cmd and stopCh
|
||||||
|
}
|
||||||
|
|
||||||
|
func (stt *WhisperServer) StartRecording() error {
|
||||||
|
stt.mu.Lock()
|
||||||
|
defer stt.mu.Unlock()
|
||||||
|
if stt.recording {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
// Build ffmpeg command for microphone capture
|
||||||
|
args := []string{
|
||||||
|
"-f", "alsa",
|
||||||
|
"-i", "default",
|
||||||
|
"-acodec", "pcm_s16le",
|
||||||
|
"-ar", fmt.Sprint(stt.SampleRate),
|
||||||
|
"-ac", "1",
|
||||||
|
"-f", "s16le",
|
||||||
|
"-",
|
||||||
|
}
|
||||||
|
cmd := exec.Command("ffmpeg", args...)
|
||||||
|
stdout, err := cmd.StdoutPipe()
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("failed to get stdout pipe: %w", err)
|
||||||
|
}
|
||||||
|
stt.cmdMu.Lock()
|
||||||
|
stt.cmd = cmd
|
||||||
|
stt.stopCh = make(chan struct{})
|
||||||
|
stt.cmdMu.Unlock()
|
||||||
|
if err := cmd.Start(); err != nil {
|
||||||
|
return fmt.Errorf("failed to start ffmpeg: %w", err)
|
||||||
|
}
|
||||||
|
stt.recording = true
|
||||||
|
stt.AudioBuffer.Reset()
|
||||||
|
// Read PCM data in goroutine
|
||||||
|
go func() {
|
||||||
|
buf := make([]byte, 4096)
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-stt.stopCh:
|
||||||
|
return
|
||||||
|
default:
|
||||||
|
n, err := stdout.Read(buf)
|
||||||
|
if n > 0 {
|
||||||
|
stt.mu.Lock()
|
||||||
|
stt.AudioBuffer.Write(buf[:n])
|
||||||
|
stt.mu.Unlock()
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
if err != io.EOF {
|
||||||
|
stt.logger.Error("recording read error", "error", err)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (stt *WhisperServer) StopRecording() (string, error) {
|
||||||
|
stt.mu.Lock()
|
||||||
|
defer stt.mu.Unlock()
|
||||||
|
if !stt.recording {
|
||||||
|
return "", errors.New("not recording")
|
||||||
|
}
|
||||||
|
stt.recording = false
|
||||||
|
// Stop ffmpeg
|
||||||
|
stt.cmdMu.Lock()
|
||||||
|
if stt.cmd != nil && stt.cmd.Process != nil {
|
||||||
|
stt.cmd.Process.Kill()
|
||||||
|
stt.cmd.Wait()
|
||||||
|
}
|
||||||
|
close(stt.stopCh)
|
||||||
|
stt.cmdMu.Unlock()
|
||||||
|
// Rest of StopRecording unchanged (WAV header + HTTP upload)
|
||||||
|
// ...
|
||||||
|
stt.recording = false
|
||||||
|
// wait loop to finish?
|
||||||
|
if stt.AudioBuffer == nil {
|
||||||
|
err := errors.New("unexpected nil AudioBuffer")
|
||||||
|
stt.logger.Error(err.Error())
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
// Create WAV header first
|
||||||
|
body := &bytes.Buffer{}
|
||||||
|
writer := multipart.NewWriter(body)
|
||||||
|
// Add audio file part
|
||||||
|
part, err := writer.CreateFormFile("file", "recording.wav")
|
||||||
|
if err != nil {
|
||||||
|
stt.logger.Error("fn: StopRecording", "error", err)
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
// Stream directly to multipart writer: header + raw data
|
||||||
|
dataSize := stt.AudioBuffer.Len()
|
||||||
|
stt.writeWavHeader(part, dataSize)
|
||||||
|
if _, err := io.Copy(part, stt.AudioBuffer); err != nil {
|
||||||
|
stt.logger.Error("fn: StopRecording", "error", err)
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
// Reset buffer for next recording
|
||||||
|
stt.AudioBuffer.Reset()
|
||||||
|
// Add response format field
|
||||||
|
err = writer.WriteField("response_format", "text")
|
||||||
|
if err != nil {
|
||||||
|
stt.logger.Error("fn: StopRecording", "error", err)
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
if writer.Close() != nil {
|
||||||
|
stt.logger.Error("fn: StopRecording", "error", err)
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
// Send request
|
||||||
|
resp, err := http.Post(stt.ServerURL, writer.FormDataContentType(), body) //nolint:noctx
|
||||||
|
if err != nil {
|
||||||
|
stt.logger.Error("fn: StopRecording", "error", err)
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
// Read and print response
|
||||||
|
responseTextBytes, err := io.ReadAll(resp.Body)
|
||||||
|
if err != nil {
|
||||||
|
stt.logger.Error("fn: StopRecording", "error", err)
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
resptext := strings.TrimRight(string(responseTextBytes), "\n")
|
||||||
|
// in case there are special tokens like [_BEG_]
|
||||||
|
resptext = specialRE.ReplaceAllString(resptext, "")
|
||||||
|
return strings.TrimSpace(strings.ReplaceAll(resptext, "\n ", "\n")), nil
|
||||||
|
}
|
||||||
8
go.mod
8
go.mod
@@ -4,13 +4,11 @@ go 1.25.1
|
|||||||
|
|
||||||
require (
|
require (
|
||||||
github.com/BurntSushi/toml v1.5.0
|
github.com/BurntSushi/toml v1.5.0
|
||||||
github.com/GrailFinder/google-translate-tts v0.1.3
|
github.com/GrailFinder/google-translate-tts v0.1.4
|
||||||
github.com/GrailFinder/searchagent v0.2.0
|
github.com/GrailFinder/searchagent v0.2.0
|
||||||
github.com/PuerkitoBio/goquery v1.11.0
|
github.com/PuerkitoBio/goquery v1.11.0
|
||||||
github.com/gdamore/tcell/v2 v2.13.2
|
github.com/gdamore/tcell/v2 v2.13.2
|
||||||
github.com/glebarez/go-sqlite v1.22.0
|
github.com/glebarez/go-sqlite v1.22.0
|
||||||
github.com/gopxl/beep/v2 v2.1.1
|
|
||||||
github.com/gordonklaus/portaudio v0.0.0-20250206071425-98a94950218b
|
|
||||||
github.com/jmoiron/sqlx v1.4.0
|
github.com/jmoiron/sqlx v1.4.0
|
||||||
github.com/ledongthuc/pdf v0.0.0-20250511090121-5959a4027728
|
github.com/ledongthuc/pdf v0.0.0-20250511090121-5959a4027728
|
||||||
github.com/neurosnap/sentences v1.1.2
|
github.com/neurosnap/sentences v1.1.2
|
||||||
@@ -25,21 +23,17 @@ require (
|
|||||||
github.com/andybalholm/cascadia v1.3.3 // indirect
|
github.com/andybalholm/cascadia v1.3.3 // indirect
|
||||||
github.com/deckarep/golang-set/v2 v2.8.0 // indirect
|
github.com/deckarep/golang-set/v2 v2.8.0 // indirect
|
||||||
github.com/dustin/go-humanize v1.0.1 // indirect
|
github.com/dustin/go-humanize v1.0.1 // indirect
|
||||||
github.com/ebitengine/oto/v3 v3.4.0 // indirect
|
|
||||||
github.com/ebitengine/purego v0.9.1 // indirect
|
|
||||||
github.com/emirpasic/gods v1.18.1 // indirect
|
github.com/emirpasic/gods v1.18.1 // indirect
|
||||||
github.com/gdamore/encoding v1.0.1 // indirect
|
github.com/gdamore/encoding v1.0.1 // indirect
|
||||||
github.com/go-jose/go-jose/v3 v3.0.4 // indirect
|
github.com/go-jose/go-jose/v3 v3.0.4 // indirect
|
||||||
github.com/go-stack/stack v1.8.1 // indirect
|
github.com/go-stack/stack v1.8.1 // indirect
|
||||||
github.com/google/uuid v1.6.0 // indirect
|
github.com/google/uuid v1.6.0 // indirect
|
||||||
github.com/hajimehoshi/go-mp3 v0.3.4 // indirect
|
github.com/hajimehoshi/go-mp3 v0.3.4 // indirect
|
||||||
github.com/hajimehoshi/oto/v2 v2.3.1 // indirect
|
|
||||||
github.com/lucasb-eyer/go-colorful v1.3.0 // indirect
|
github.com/lucasb-eyer/go-colorful v1.3.0 // indirect
|
||||||
github.com/mattn/go-isatty v0.0.20 // indirect
|
github.com/mattn/go-isatty v0.0.20 // indirect
|
||||||
github.com/mitchellh/colorstring v0.0.0-20190213212951-d06e56a500db // indirect
|
github.com/mitchellh/colorstring v0.0.0-20190213212951-d06e56a500db // indirect
|
||||||
github.com/ncruces/go-strftime v1.0.0 // indirect
|
github.com/ncruces/go-strftime v1.0.0 // indirect
|
||||||
github.com/patrickmn/go-cache v2.1.0+incompatible // indirect
|
github.com/patrickmn/go-cache v2.1.0+incompatible // indirect
|
||||||
github.com/pkg/errors v0.9.1 // indirect
|
|
||||||
github.com/remyoudompheng/bigfft v0.0.0-20230129092748-24d4a6f8daec // indirect
|
github.com/remyoudompheng/bigfft v0.0.0-20230129092748-24d4a6f8daec // indirect
|
||||||
github.com/rivo/uniseg v0.4.7 // indirect
|
github.com/rivo/uniseg v0.4.7 // indirect
|
||||||
github.com/schollz/progressbar/v2 v2.15.0 // indirect
|
github.com/schollz/progressbar/v2 v2.15.0 // indirect
|
||||||
|
|||||||
15
go.sum
15
go.sum
@@ -2,8 +2,8 @@ filippo.io/edwards25519 v1.1.0 h1:FNf4tywRC1HmFuKW5xopWpigGjJKiJSV0Cqo0cJWDaA=
|
|||||||
filippo.io/edwards25519 v1.1.0/go.mod h1:BxyFTGdWcka3PhytdK4V28tE5sGfRvvvRV7EaN4VDT4=
|
filippo.io/edwards25519 v1.1.0/go.mod h1:BxyFTGdWcka3PhytdK4V28tE5sGfRvvvRV7EaN4VDT4=
|
||||||
github.com/BurntSushi/toml v1.5.0 h1:W5quZX/G/csjUnuI8SUYlsHs9M38FC7znL0lIO+DvMg=
|
github.com/BurntSushi/toml v1.5.0 h1:W5quZX/G/csjUnuI8SUYlsHs9M38FC7znL0lIO+DvMg=
|
||||||
github.com/BurntSushi/toml v1.5.0/go.mod h1:ukJfTF/6rtPPRCnwkur4qwRxa8vTRFBF0uk2lLoLwho=
|
github.com/BurntSushi/toml v1.5.0/go.mod h1:ukJfTF/6rtPPRCnwkur4qwRxa8vTRFBF0uk2lLoLwho=
|
||||||
github.com/GrailFinder/google-translate-tts v0.1.3 h1:Mww9tNzTWjjSh+OCbTPl/+21oMPKcUecXZfU7nTB/lA=
|
github.com/GrailFinder/google-translate-tts v0.1.4 h1:NJoPZUGfBrmouQMN19MUcNPNUx4tmf4a8OZRME4E4Mg=
|
||||||
github.com/GrailFinder/google-translate-tts v0.1.3/go.mod h1:YIOLKR7sObazdUCrSex3u9OVBovU55eYgWa25vsQJ18=
|
github.com/GrailFinder/google-translate-tts v0.1.4/go.mod h1:YIOLKR7sObazdUCrSex3u9OVBovU55eYgWa25vsQJ18=
|
||||||
github.com/GrailFinder/searchagent v0.2.0 h1:U2GVjLh/9xZt0xX9OcYk9Q2fMkyzyTiADPUmUisRdtQ=
|
github.com/GrailFinder/searchagent v0.2.0 h1:U2GVjLh/9xZt0xX9OcYk9Q2fMkyzyTiADPUmUisRdtQ=
|
||||||
github.com/GrailFinder/searchagent v0.2.0/go.mod h1:d66tn5+22LI8IGJREUsRBT60P0sFdgQgvQRqyvgItrs=
|
github.com/GrailFinder/searchagent v0.2.0/go.mod h1:d66tn5+22LI8IGJREUsRBT60P0sFdgQgvQRqyvgItrs=
|
||||||
github.com/PuerkitoBio/goquery v1.11.0 h1:jZ7pwMQXIITcUXNH83LLk+txlaEy6NVOfTuP43xxfqw=
|
github.com/PuerkitoBio/goquery v1.11.0 h1:jZ7pwMQXIITcUXNH83LLk+txlaEy6NVOfTuP43xxfqw=
|
||||||
@@ -17,10 +17,6 @@ github.com/deckarep/golang-set/v2 v2.8.0 h1:swm0rlPCmdWn9mESxKOjWk8hXSqoxOp+Zlfu
|
|||||||
github.com/deckarep/golang-set/v2 v2.8.0/go.mod h1:VAky9rY/yGXJOLEDv3OMci+7wtDpOF4IN+y82NBOac4=
|
github.com/deckarep/golang-set/v2 v2.8.0/go.mod h1:VAky9rY/yGXJOLEDv3OMci+7wtDpOF4IN+y82NBOac4=
|
||||||
github.com/dustin/go-humanize v1.0.1 h1:GzkhY7T5VNhEkwH0PVJgjz+fX1rhBrR7pRT3mDkpeCY=
|
github.com/dustin/go-humanize v1.0.1 h1:GzkhY7T5VNhEkwH0PVJgjz+fX1rhBrR7pRT3mDkpeCY=
|
||||||
github.com/dustin/go-humanize v1.0.1/go.mod h1:Mu1zIs6XwVuF/gI1OepvI0qD18qycQx+mFykh5fBlto=
|
github.com/dustin/go-humanize v1.0.1/go.mod h1:Mu1zIs6XwVuF/gI1OepvI0qD18qycQx+mFykh5fBlto=
|
||||||
github.com/ebitengine/oto/v3 v3.4.0 h1:br0PgASsEWaoWn38b2Goe7m1GKFYfNgnsjSd5Gg+/bQ=
|
|
||||||
github.com/ebitengine/oto/v3 v3.4.0/go.mod h1:IOleLVD0m+CMak3mRVwsYY8vTctQgOM0iiL6S7Ar7eI=
|
|
||||||
github.com/ebitengine/purego v0.9.1 h1:a/k2f2HQU3Pi399RPW1MOaZyhKJL9w/xFpKAg4q1s0A=
|
|
||||||
github.com/ebitengine/purego v0.9.1/go.mod h1:iIjxzd6CiRiOG0UyXP+V1+jWqUXVjPKLAI0mRfJZTmQ=
|
|
||||||
github.com/emirpasic/gods v1.18.1 h1:FXtiHYKDGKCW2KzwZKx0iC0PQmdlorYgdFG9jPXJ1Bc=
|
github.com/emirpasic/gods v1.18.1 h1:FXtiHYKDGKCW2KzwZKx0iC0PQmdlorYgdFG9jPXJ1Bc=
|
||||||
github.com/emirpasic/gods v1.18.1/go.mod h1:8tpGGwCnJ5H4r6BWwaV6OrWmMoPhUl5jm/FMNAnJvWQ=
|
github.com/emirpasic/gods v1.18.1/go.mod h1:8tpGGwCnJ5H4r6BWwaV6OrWmMoPhUl5jm/FMNAnJvWQ=
|
||||||
github.com/gdamore/encoding v1.0.1 h1:YzKZckdBL6jVt2Gc+5p82qhrGiqMdG/eNs6Wy0u3Uhw=
|
github.com/gdamore/encoding v1.0.1 h1:YzKZckdBL6jVt2Gc+5p82qhrGiqMdG/eNs6Wy0u3Uhw=
|
||||||
@@ -41,13 +37,8 @@ github.com/google/pprof v0.0.0-20250317173921-a4b03ec1a45e h1:ijClszYn+mADRFY17k
|
|||||||
github.com/google/pprof v0.0.0-20250317173921-a4b03ec1a45e/go.mod h1:boTsfXsheKC2y+lKOCMpSfarhxDeIzfZG1jqGcPl3cA=
|
github.com/google/pprof v0.0.0-20250317173921-a4b03ec1a45e/go.mod h1:boTsfXsheKC2y+lKOCMpSfarhxDeIzfZG1jqGcPl3cA=
|
||||||
github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0=
|
github.com/google/uuid v1.6.0 h1:NIvaJDMOsjHA8n1jAhLSgzrAzy1Hgr+hNrb57e+94F0=
|
||||||
github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo=
|
github.com/google/uuid v1.6.0/go.mod h1:TIyPZe4MgqvfeYDBFedMoGGpEw/LqOeaOT+nhxU+yHo=
|
||||||
github.com/gopxl/beep/v2 v2.1.1 h1:6FYIYMm2qPAdWkjX+7xwKrViS1x0Po5kDMdRkq8NVbU=
|
|
||||||
github.com/gopxl/beep/v2 v2.1.1/go.mod h1:ZAm9TGQ9lvpoiFLd4zf5B1IuyxZhgRACMId1XJbaW0E=
|
|
||||||
github.com/gordonklaus/portaudio v0.0.0-20250206071425-98a94950218b h1:WEuQWBxelOGHA6z9lABqaMLMrfwVyMdN3UgRLT+YUPo=
|
|
||||||
github.com/gordonklaus/portaudio v0.0.0-20250206071425-98a94950218b/go.mod h1:esZFQEUwqC+l76f2R8bIWSwXMaPbp79PppwZ1eJhFco=
|
|
||||||
github.com/hajimehoshi/go-mp3 v0.3.4 h1:NUP7pBYH8OguP4diaTZ9wJbUbk3tC0KlfzsEpWmYj68=
|
github.com/hajimehoshi/go-mp3 v0.3.4 h1:NUP7pBYH8OguP4diaTZ9wJbUbk3tC0KlfzsEpWmYj68=
|
||||||
github.com/hajimehoshi/go-mp3 v0.3.4/go.mod h1:fRtZraRFcWb0pu7ok0LqyFhCUrPeMsGRSVop0eemFmo=
|
github.com/hajimehoshi/go-mp3 v0.3.4/go.mod h1:fRtZraRFcWb0pu7ok0LqyFhCUrPeMsGRSVop0eemFmo=
|
||||||
github.com/hajimehoshi/oto/v2 v2.3.1 h1:qrLKpNus2UfD674oxckKjNJmesp9hMh7u7QCrStB3Rc=
|
|
||||||
github.com/hajimehoshi/oto/v2 v2.3.1/go.mod h1:seWLbgHH7AyUMYKfKYT9pg7PhUu9/SisyJvNTT+ASQo=
|
github.com/hajimehoshi/oto/v2 v2.3.1/go.mod h1:seWLbgHH7AyUMYKfKYT9pg7PhUu9/SisyJvNTT+ASQo=
|
||||||
github.com/hashicorp/golang-lru/v2 v2.0.7 h1:a+bsQ5rvGLjzHuww6tVxozPZFVghXaHOwFs4luLUK2k=
|
github.com/hashicorp/golang-lru/v2 v2.0.7 h1:a+bsQ5rvGLjzHuww6tVxozPZFVghXaHOwFs4luLUK2k=
|
||||||
github.com/hashicorp/golang-lru/v2 v2.0.7/go.mod h1:QeFd9opnmA6QUJc5vARoKUSoFhyfM2/ZepoAG6RGpeM=
|
github.com/hashicorp/golang-lru/v2 v2.0.7/go.mod h1:QeFd9opnmA6QUJc5vARoKUSoFhyfM2/ZepoAG6RGpeM=
|
||||||
@@ -71,8 +62,6 @@ github.com/neurosnap/sentences v1.1.2 h1:iphYOzx/XckXeBiLIUBkPu2EKMJ+6jDbz/sLJZ7
|
|||||||
github.com/neurosnap/sentences v1.1.2/go.mod h1:/pwU4E9XNL21ygMIkOIllv/SMy2ujHwpf8GQPu1YPbQ=
|
github.com/neurosnap/sentences v1.1.2/go.mod h1:/pwU4E9XNL21ygMIkOIllv/SMy2ujHwpf8GQPu1YPbQ=
|
||||||
github.com/patrickmn/go-cache v2.1.0+incompatible h1:HRMgzkcYKYpi3C8ajMPV8OFXaaRUnok+kx1WdO15EQc=
|
github.com/patrickmn/go-cache v2.1.0+incompatible h1:HRMgzkcYKYpi3C8ajMPV8OFXaaRUnok+kx1WdO15EQc=
|
||||||
github.com/patrickmn/go-cache v2.1.0+incompatible/go.mod h1:3Qf8kWWT7OJRJbdiICTKqZju1ZixQ/KpMGzzAfe6+WQ=
|
github.com/patrickmn/go-cache v2.1.0+incompatible/go.mod h1:3Qf8kWWT7OJRJbdiICTKqZju1ZixQ/KpMGzzAfe6+WQ=
|
||||||
github.com/pkg/errors v0.9.1 h1:FEBLx1zS214owpjy7qsBeixbURkuhQAwrK5UwLGTwt4=
|
|
||||||
github.com/pkg/errors v0.9.1/go.mod h1:bwawxfHBFNV+L2hUp1rHADufV3IMtnDRdf1r5NINEl0=
|
|
||||||
github.com/playwright-community/playwright-go v0.5700.1 h1:PNFb1byWqrTT720rEO0JL88C6Ju0EmUnR5deFLvtP/U=
|
github.com/playwright-community/playwright-go v0.5700.1 h1:PNFb1byWqrTT720rEO0JL88C6Ju0EmUnR5deFLvtP/U=
|
||||||
github.com/playwright-community/playwright-go v0.5700.1/go.mod h1:MlSn1dZrx8rszbCxY6x3qK89ZesJUYVx21B2JnkoNF0=
|
github.com/playwright-community/playwright-go v0.5700.1/go.mod h1:MlSn1dZrx8rszbCxY6x3qK89ZesJUYVx21B2JnkoNF0=
|
||||||
github.com/pmezard/go-difflib v1.0.0 h1:4DBwDE0NGyQoBHbLQYPwSUPoCMWR5BEzIk/f1lZbAQM=
|
github.com/pmezard/go-difflib v1.0.0 h1:4DBwDE0NGyQoBHbLQYPwSUPoCMWR5BEzIk/f1lZbAQM=
|
||||||
|
|||||||
106
helpfuncs.go
106
helpfuncs.go
@@ -5,7 +5,6 @@ import (
|
|||||||
"gf-lt/models"
|
"gf-lt/models"
|
||||||
"gf-lt/pngmeta"
|
"gf-lt/pngmeta"
|
||||||
"image"
|
"image"
|
||||||
"net/url"
|
|
||||||
"os"
|
"os"
|
||||||
"os/exec"
|
"os/exec"
|
||||||
"path"
|
"path"
|
||||||
@@ -13,6 +12,7 @@ import (
|
|||||||
"slices"
|
"slices"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
|
"sync/atomic"
|
||||||
"time"
|
"time"
|
||||||
"unicode"
|
"unicode"
|
||||||
|
|
||||||
@@ -20,7 +20,8 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
// Cached model color - updated by background goroutine
|
// Cached model color - updated by background goroutine
|
||||||
var cachedModelColor string = "orange"
|
// var cachedModelColor string = "orange"
|
||||||
|
var cachedModelColor atomic.Value
|
||||||
|
|
||||||
// startModelColorUpdater starts a background goroutine that periodically updates
|
// startModelColorUpdater starts a background goroutine that periodically updates
|
||||||
// the cached model color. Only runs HTTP requests for local llama.cpp APIs.
|
// the cached model color. Only runs HTTP requests for local llama.cpp APIs.
|
||||||
@@ -39,20 +40,20 @@ func startModelColorUpdater() {
|
|||||||
// updateCachedModelColor updates the global cachedModelColor variable
|
// updateCachedModelColor updates the global cachedModelColor variable
|
||||||
func updateCachedModelColor() {
|
func updateCachedModelColor() {
|
||||||
if !isLocalLlamacpp() {
|
if !isLocalLlamacpp() {
|
||||||
cachedModelColor = "orange"
|
cachedModelColor.Store("orange")
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
// Check if model is loaded
|
// Check if model is loaded
|
||||||
loaded, err := isModelLoaded(chatBody.Model)
|
loaded, err := isModelLoaded(chatBody.Model)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
// On error, assume not loaded (red)
|
// On error, assume not loaded (red)
|
||||||
cachedModelColor = "red"
|
cachedModelColor.Store("red")
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
if loaded {
|
if loaded {
|
||||||
cachedModelColor = "green"
|
cachedModelColor.Store("green")
|
||||||
} else {
|
} else {
|
||||||
cachedModelColor = "red"
|
cachedModelColor.Store("red")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -85,8 +86,8 @@ func stripThinkingFromMsg(msg *models.RoleMsg) *models.RoleMsg {
|
|||||||
}
|
}
|
||||||
// Strip thinking from assistant messages
|
// Strip thinking from assistant messages
|
||||||
msgText := msg.GetText()
|
msgText := msg.GetText()
|
||||||
if thinkRE.MatchString(msgText) {
|
if models.ThinkRE.MatchString(msgText) {
|
||||||
cleanedText := thinkRE.ReplaceAllString(msgText, "")
|
cleanedText := models.ThinkRE.ReplaceAllString(msgText, "")
|
||||||
cleanedText = strings.TrimSpace(cleanedText)
|
cleanedText = strings.TrimSpace(cleanedText)
|
||||||
msg.SetText(cleanedText)
|
msg.SetText(cleanedText)
|
||||||
}
|
}
|
||||||
@@ -108,7 +109,7 @@ func refreshChatDisplay() {
|
|||||||
textView.SetText(displayText)
|
textView.SetText(displayText)
|
||||||
colorText()
|
colorText()
|
||||||
updateStatusLine()
|
updateStatusLine()
|
||||||
if scrollToEndEnabled {
|
if cfg.AutoScrollEnabled {
|
||||||
textView.ScrollToEnd()
|
textView.ScrollToEnd()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -147,7 +148,7 @@ func colorText() {
|
|||||||
placeholderThink := "__THINK_BLOCK_%d__"
|
placeholderThink := "__THINK_BLOCK_%d__"
|
||||||
counterThink := 0
|
counterThink := 0
|
||||||
// Replace code blocks with placeholders and store their styled versions
|
// Replace code blocks with placeholders and store their styled versions
|
||||||
text = codeBlockRE.ReplaceAllStringFunc(text, func(match string) string {
|
text = models.CodeBlockRE.ReplaceAllStringFunc(text, func(match string) string {
|
||||||
// Style the code block and store it
|
// Style the code block and store it
|
||||||
styled := fmt.Sprintf("[red::i]%s[-:-:-]", match)
|
styled := fmt.Sprintf("[red::i]%s[-:-:-]", match)
|
||||||
codeBlocks = append(codeBlocks, styled)
|
codeBlocks = append(codeBlocks, styled)
|
||||||
@@ -156,7 +157,7 @@ func colorText() {
|
|||||||
counter++
|
counter++
|
||||||
return id
|
return id
|
||||||
})
|
})
|
||||||
text = thinkRE.ReplaceAllStringFunc(text, func(match string) string {
|
text = models.ThinkRE.ReplaceAllStringFunc(text, func(match string) string {
|
||||||
// Style the code block and store it
|
// Style the code block and store it
|
||||||
styled := fmt.Sprintf("[red::i]%s[-:-:-]", match)
|
styled := fmt.Sprintf("[red::i]%s[-:-:-]", match)
|
||||||
thinkBlocks = append(thinkBlocks, styled)
|
thinkBlocks = append(thinkBlocks, styled)
|
||||||
@@ -166,10 +167,10 @@ func colorText() {
|
|||||||
return id
|
return id
|
||||||
})
|
})
|
||||||
// Step 2: Apply other regex styles to the non-code parts
|
// Step 2: Apply other regex styles to the non-code parts
|
||||||
text = quotesRE.ReplaceAllString(text, `[orange::-]$1[-:-:-]`)
|
text = models.QuotesRE.ReplaceAllString(text, `[orange::-]$1[-:-:-]`)
|
||||||
text = starRE.ReplaceAllString(text, `[turquoise::i]$1[-:-:-]`)
|
text = models.StarRE.ReplaceAllString(text, `[turquoise::i]$1[-:-:-]`)
|
||||||
text = singleBacktickRE.ReplaceAllString(text, "`[pink::i]$1[-:-:-]`")
|
text = models.SingleBacktickRE.ReplaceAllString(text, "`[pink::i]$1[-:-:-]`")
|
||||||
// text = thinkRE.ReplaceAllString(text, `[yellow::i]$1[-:-:-]`)
|
// text = tools.ThinkRE.ReplaceAllString(text, `[yellow::i]$1[-:-:-]`)
|
||||||
// Step 3: Restore the styled code blocks from placeholders
|
// Step 3: Restore the styled code blocks from placeholders
|
||||||
for i, cb := range codeBlocks {
|
for i, cb := range codeBlocks {
|
||||||
text = strings.Replace(text, fmt.Sprintf(placeholder, i), cb, 1)
|
text = strings.Replace(text, fmt.Sprintf(placeholder, i), cb, 1)
|
||||||
@@ -187,7 +188,7 @@ func updateStatusLine() {
|
|||||||
|
|
||||||
func initSysCards() ([]string, error) {
|
func initSysCards() ([]string, error) {
|
||||||
labels := []string{}
|
labels := []string{}
|
||||||
labels = append(labels, sysLabels...)
|
labels = append(labels, models.SysLabels...)
|
||||||
cards, err := pngmeta.ReadDirCards(cfg.SysDir, cfg.UserRole, logger)
|
cards, err := pngmeta.ReadDirCards(cfg.SysDir, cfg.UserRole, logger)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
logger.Error("failed to read sys dir", "error", err)
|
logger.Error("failed to read sys dir", "error", err)
|
||||||
@@ -323,19 +324,17 @@ func strInSlice(s string, sl []string) bool {
|
|||||||
|
|
||||||
// isLocalLlamacpp checks if the current API is a local llama.cpp instance.
|
// isLocalLlamacpp checks if the current API is a local llama.cpp instance.
|
||||||
func isLocalLlamacpp() bool {
|
func isLocalLlamacpp() bool {
|
||||||
u, err := url.Parse(cfg.CurrentAPI)
|
if strings.Contains(cfg.CurrentAPI, "openrouter") || strings.Contains(cfg.CurrentAPI, "deepseek") {
|
||||||
if err != nil {
|
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
host := u.Hostname()
|
return true
|
||||||
return host == "localhost" || host == "127.0.0.1" || host == "::1"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// getModelColor returns the cached color tag for the model name.
|
// getModelColor returns the cached color tag for the model name.
|
||||||
// The cached value is updated by a background goroutine every 5 seconds.
|
// The cached value is updated by a background goroutine every 5 seconds.
|
||||||
// For non-local models, returns orange. For local llama.cpp models, returns green if loaded, red if not.
|
// For non-local models, returns orange. For local llama.cpp models, returns green if loaded, red if not.
|
||||||
func getModelColor() string {
|
func getModelColor() string {
|
||||||
return cachedModelColor
|
return cachedModelColor.Load().(string)
|
||||||
}
|
}
|
||||||
|
|
||||||
func makeStatusLine() string {
|
func makeStatusLine() string {
|
||||||
@@ -542,7 +541,7 @@ func executeCommandAndDisplay(cmdText string) {
|
|||||||
cmdText = strings.TrimSpace(cmdText)
|
cmdText = strings.TrimSpace(cmdText)
|
||||||
if cmdText == "" {
|
if cmdText == "" {
|
||||||
fmt.Fprintf(textView, "\n[red]Error: No command provided[-:-:-]\n")
|
fmt.Fprintf(textView, "\n[red]Error: No command provided[-:-:-]\n")
|
||||||
if scrollToEndEnabled {
|
if cfg.AutoScrollEnabled {
|
||||||
textView.ScrollToEnd()
|
textView.ScrollToEnd()
|
||||||
}
|
}
|
||||||
colorText()
|
colorText()
|
||||||
@@ -574,7 +573,7 @@ func executeCommandAndDisplay(cmdText string) {
|
|||||||
Content: "$ " + cmdText + "\n\n" + outputContent,
|
Content: "$ " + cmdText + "\n\n" + outputContent,
|
||||||
}
|
}
|
||||||
chatBody.Messages = append(chatBody.Messages, combinedMsg)
|
chatBody.Messages = append(chatBody.Messages, combinedMsg)
|
||||||
if scrollToEndEnabled {
|
if cfg.AutoScrollEnabled {
|
||||||
textView.ScrollToEnd()
|
textView.ScrollToEnd()
|
||||||
}
|
}
|
||||||
colorText()
|
colorText()
|
||||||
@@ -589,7 +588,7 @@ func executeCommandAndDisplay(cmdText string) {
|
|||||||
Content: "$ " + cmdText + "\n\n" + outputContent,
|
Content: "$ " + cmdText + "\n\n" + outputContent,
|
||||||
}
|
}
|
||||||
chatBody.Messages = append(chatBody.Messages, combinedMsg)
|
chatBody.Messages = append(chatBody.Messages, combinedMsg)
|
||||||
if scrollToEndEnabled {
|
if cfg.AutoScrollEnabled {
|
||||||
textView.ScrollToEnd()
|
textView.ScrollToEnd()
|
||||||
}
|
}
|
||||||
colorText()
|
colorText()
|
||||||
@@ -637,7 +636,7 @@ func executeCommandAndDisplay(cmdText string) {
|
|||||||
}
|
}
|
||||||
chatBody.Messages = append(chatBody.Messages, combinedMsg)
|
chatBody.Messages = append(chatBody.Messages, combinedMsg)
|
||||||
// Scroll to end and update colors
|
// Scroll to end and update colors
|
||||||
if scrollToEndEnabled {
|
if cfg.AutoScrollEnabled {
|
||||||
textView.ScrollToEnd()
|
textView.ScrollToEnd()
|
||||||
}
|
}
|
||||||
colorText()
|
colorText()
|
||||||
@@ -967,3 +966,60 @@ func extractDisplayPath(p, bp string) string {
|
|||||||
}
|
}
|
||||||
return p
|
return p
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func getValidKnowToRecipient(msg *models.RoleMsg) (string, bool) {
|
||||||
|
if cfg == nil || !cfg.CharSpecificContextEnabled {
|
||||||
|
return "", false
|
||||||
|
}
|
||||||
|
// case where all roles are in the tag => public message
|
||||||
|
cr := listChatRoles()
|
||||||
|
slices.Sort(cr)
|
||||||
|
slices.Sort(msg.KnownTo)
|
||||||
|
if slices.Equal(cr, msg.KnownTo) {
|
||||||
|
logger.Info("got msg with tag mentioning every role")
|
||||||
|
return "", false
|
||||||
|
}
|
||||||
|
// Check each character in the KnownTo list
|
||||||
|
for _, recipient := range msg.KnownTo {
|
||||||
|
if recipient == msg.Role || recipient == cfg.ToolRole {
|
||||||
|
// weird cases, skip
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// Skip if this is the user character (user handles their own turn)
|
||||||
|
// If user is in KnownTo, stop processing - it's the user's turn
|
||||||
|
if recipient == cfg.UserRole || recipient == cfg.WriteNextMsgAs {
|
||||||
|
return "", false
|
||||||
|
}
|
||||||
|
return recipient, true
|
||||||
|
}
|
||||||
|
return "", false
|
||||||
|
}
|
||||||
|
|
||||||
|
// triggerPrivateMessageResponses checks if a message was sent privately to specific characters
|
||||||
|
// and triggers those non-user characters to respond
|
||||||
|
func triggerPrivateMessageResponses(msg *models.RoleMsg) {
|
||||||
|
recipient, ok := getValidKnowToRecipient(msg)
|
||||||
|
if !ok || recipient == "" {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
// Trigger the recipient character to respond
|
||||||
|
triggerMsg := recipient + ":\n"
|
||||||
|
// Send empty message so LLM continues naturally from the conversation
|
||||||
|
crr := &models.ChatRoundReq{
|
||||||
|
UserMsg: triggerMsg,
|
||||||
|
Role: recipient,
|
||||||
|
Resume: true,
|
||||||
|
}
|
||||||
|
fmt.Fprintf(textView, "\n[-:-:b](%d) ", len(chatBody.Messages))
|
||||||
|
fmt.Fprint(textView, roleToIcon(recipient))
|
||||||
|
fmt.Fprint(textView, "[-:-:-]\n")
|
||||||
|
chatRoundChan <- crr
|
||||||
|
}
|
||||||
|
|
||||||
|
func GetCardByRole(role string) *models.CharCard {
|
||||||
|
cardID, ok := roleToID[role]
|
||||||
|
if !ok {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return sysMap[cardID]
|
||||||
|
}
|
||||||
|
|||||||
24
llm.go
24
llm.go
@@ -4,6 +4,7 @@ import (
|
|||||||
"bytes"
|
"bytes"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"gf-lt/models"
|
"gf-lt/models"
|
||||||
|
"gf-lt/tools"
|
||||||
"io"
|
"io"
|
||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
@@ -11,10 +12,10 @@ import (
|
|||||||
var imageAttachmentPath string // Global variable to track image attachment for next message
|
var imageAttachmentPath string // Global variable to track image attachment for next message
|
||||||
var lastImg string // for ctrl+j
|
var lastImg string // for ctrl+j
|
||||||
|
|
||||||
// containsToolSysMsg checks if the toolSysMsg already exists in the chat body
|
// containsToolSysMsg checks if the tools.ToolSysMsg already exists in the chat body
|
||||||
func containsToolSysMsg() bool {
|
func containsToolSysMsg() bool {
|
||||||
for i := range chatBody.Messages {
|
for i := range chatBody.Messages {
|
||||||
if chatBody.Messages[i].Role == cfg.ToolRole && chatBody.Messages[i].Content == toolSysMsg {
|
if chatBody.Messages[i].Role == cfg.ToolRole && chatBody.Messages[i].Content == tools.ToolSysMsg {
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -62,11 +63,11 @@ type ChunkParser interface {
|
|||||||
func choseChunkParser() {
|
func choseChunkParser() {
|
||||||
chunkParser = LCPCompletion{}
|
chunkParser = LCPCompletion{}
|
||||||
switch cfg.CurrentAPI {
|
switch cfg.CurrentAPI {
|
||||||
case "http://localhost:8080/completion":
|
case "http://localhost:8080/completion", "http://127.0.0.1:8080/completion":
|
||||||
chunkParser = LCPCompletion{}
|
chunkParser = LCPCompletion{}
|
||||||
logger.Debug("chosen lcpcompletion", "link", cfg.CurrentAPI)
|
logger.Debug("chosen lcpcompletion", "link", cfg.CurrentAPI)
|
||||||
return
|
return
|
||||||
case "http://localhost:8080/v1/chat/completions":
|
case "http://localhost:8080/v1/chat/completions", "http://127.0.0.1:8080/v1/chat/completions":
|
||||||
chunkParser = LCPChat{}
|
chunkParser = LCPChat{}
|
||||||
logger.Debug("chosen lcpchat", "link", cfg.CurrentAPI)
|
logger.Debug("chosen lcpchat", "link", cfg.CurrentAPI)
|
||||||
return
|
return
|
||||||
@@ -87,6 +88,11 @@ func choseChunkParser() {
|
|||||||
logger.Debug("chosen openrouterchat", "link", cfg.CurrentAPI)
|
logger.Debug("chosen openrouterchat", "link", cfg.CurrentAPI)
|
||||||
return
|
return
|
||||||
default:
|
default:
|
||||||
|
logger.Warn("unexpected case, assuming llama.cpp on non default address", "link", cfg.CurrentAPI)
|
||||||
|
if strings.Contains(cfg.CurrentAPI, "chat") {
|
||||||
|
chunkParser = LCPChat{}
|
||||||
|
return
|
||||||
|
}
|
||||||
chunkParser = LCPCompletion{}
|
chunkParser = LCPCompletion{}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -139,7 +145,7 @@ func (lcp LCPCompletion) FormMsg(msg, role string, resume bool) (io.Reader, erro
|
|||||||
}
|
}
|
||||||
// sending description of the tools and how to use them
|
// sending description of the tools and how to use them
|
||||||
if cfg.ToolUse && !resume && role == cfg.UserRole && !containsToolSysMsg() {
|
if cfg.ToolUse && !resume && role == cfg.UserRole && !containsToolSysMsg() {
|
||||||
chatBody.Messages = append(chatBody.Messages, models.RoleMsg{Role: cfg.ToolRole, Content: toolSysMsg})
|
chatBody.Messages = append(chatBody.Messages, models.RoleMsg{Role: cfg.ToolRole, Content: tools.ToolSysMsg})
|
||||||
}
|
}
|
||||||
filteredMessages, botPersona := filterMessagesForCurrentCharacter(chatBody.Messages)
|
filteredMessages, botPersona := filterMessagesForCurrentCharacter(chatBody.Messages)
|
||||||
// Build prompt and extract images inline as we process each message
|
// Build prompt and extract images inline as we process each message
|
||||||
@@ -326,7 +332,7 @@ func (op LCPChat) FormMsg(msg, role string, resume bool) (io.Reader, error) {
|
|||||||
Tools: nil,
|
Tools: nil,
|
||||||
}
|
}
|
||||||
if cfg.ToolUse && !resume && role != cfg.ToolRole {
|
if cfg.ToolUse && !resume && role != cfg.ToolRole {
|
||||||
req.Tools = baseTools // set tools to use
|
req.Tools = tools.BaseTools // set tools to use
|
||||||
}
|
}
|
||||||
data, err := json.Marshal(req)
|
data, err := json.Marshal(req)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -379,7 +385,7 @@ func (ds DeepSeekerCompletion) FormMsg(msg, role string, resume bool) (io.Reader
|
|||||||
}
|
}
|
||||||
// sending description of the tools and how to use them
|
// sending description of the tools and how to use them
|
||||||
if cfg.ToolUse && !resume && role == cfg.UserRole && !containsToolSysMsg() {
|
if cfg.ToolUse && !resume && role == cfg.UserRole && !containsToolSysMsg() {
|
||||||
chatBody.Messages = append(chatBody.Messages, models.RoleMsg{Role: cfg.ToolRole, Content: toolSysMsg})
|
chatBody.Messages = append(chatBody.Messages, models.RoleMsg{Role: cfg.ToolRole, Content: tools.ToolSysMsg})
|
||||||
}
|
}
|
||||||
filteredMessages, botPersona := filterMessagesForCurrentCharacter(chatBody.Messages)
|
filteredMessages, botPersona := filterMessagesForCurrentCharacter(chatBody.Messages)
|
||||||
messages := make([]string, len(filteredMessages))
|
messages := make([]string, len(filteredMessages))
|
||||||
@@ -531,7 +537,7 @@ func (or OpenRouterCompletion) FormMsg(msg, role string, resume bool) (io.Reader
|
|||||||
}
|
}
|
||||||
// sending description of the tools and how to use them
|
// sending description of the tools and how to use them
|
||||||
if cfg.ToolUse && !resume && role == cfg.UserRole && !containsToolSysMsg() {
|
if cfg.ToolUse && !resume && role == cfg.UserRole && !containsToolSysMsg() {
|
||||||
chatBody.Messages = append(chatBody.Messages, models.RoleMsg{Role: cfg.ToolRole, Content: toolSysMsg})
|
chatBody.Messages = append(chatBody.Messages, models.RoleMsg{Role: cfg.ToolRole, Content: tools.ToolSysMsg})
|
||||||
}
|
}
|
||||||
filteredMessages, botPersona := filterMessagesForCurrentCharacter(chatBody.Messages)
|
filteredMessages, botPersona := filterMessagesForCurrentCharacter(chatBody.Messages)
|
||||||
messages := make([]string, len(filteredMessages))
|
messages := make([]string, len(filteredMessages))
|
||||||
@@ -666,7 +672,7 @@ func (or OpenRouterChat) FormMsg(msg, role string, resume bool) (io.Reader, erro
|
|||||||
bodyCopy.Messages = consolidateAssistantMessages(bodyCopy.Messages)
|
bodyCopy.Messages = consolidateAssistantMessages(bodyCopy.Messages)
|
||||||
orBody := models.NewOpenRouterChatReq(*bodyCopy, defaultLCPProps, cfg.ReasoningEffort)
|
orBody := models.NewOpenRouterChatReq(*bodyCopy, defaultLCPProps, cfg.ReasoningEffort)
|
||||||
if cfg.ToolUse && !resume && role != cfg.ToolRole {
|
if cfg.ToolUse && !resume && role != cfg.ToolRole {
|
||||||
orBody.Tools = baseTools // set tools to use
|
orBody.Tools = tools.BaseTools // set tools to use
|
||||||
}
|
}
|
||||||
data, err := json.Marshal(orBody)
|
data, err := json.Marshal(orBody)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|||||||
261
main.go
261
main.go
@@ -1,13 +1,24 @@
|
|||||||
package main
|
package main
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"bufio"
|
||||||
|
"flag"
|
||||||
|
"fmt"
|
||||||
|
"gf-lt/models"
|
||||||
|
"gf-lt/pngmeta"
|
||||||
|
"os"
|
||||||
|
"slices"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
"sync/atomic"
|
||||||
|
|
||||||
"github.com/rivo/tview"
|
"github.com/rivo/tview"
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
boolColors = map[bool]string{true: "green", false: "red"}
|
boolColors = map[bool]string{true: "green", false: "red"}
|
||||||
botRespMode = false
|
botRespMode atomic.Bool
|
||||||
toolRunningMode = false
|
toolRunningMode atomic.Bool
|
||||||
editMode = false
|
editMode = false
|
||||||
roleEditMode = false
|
roleEditMode = false
|
||||||
injectRole = true
|
injectRole = true
|
||||||
@@ -20,9 +31,22 @@ var (
|
|||||||
statusLineTempl = "help (F12) | chat: [orange:-:b]%s[-:-:-] (F1) | [%s:-:b]tool use[-:-:-] (ctrl+k) | model: [%s:-:b]%s[-:-:-] (ctrl+l) | [%s:-:b]skip LLM resp[-:-:-] (F10) | API: [orange:-:b]%s[-:-:-] (ctrl+v)\nwriting as: [orange:-:b]%s[-:-:-] (ctrl+q) | bot will write as [orange:-:b]%s[-:-:-] (ctrl+x)"
|
statusLineTempl = "help (F12) | chat: [orange:-:b]%s[-:-:-] (F1) | [%s:-:b]tool use[-:-:-] (ctrl+k) | model: [%s:-:b]%s[-:-:-] (ctrl+l) | [%s:-:b]skip LLM resp[-:-:-] (F10) | API: [orange:-:b]%s[-:-:-] (ctrl+v)\nwriting as: [orange:-:b]%s[-:-:-] (ctrl+q) | bot will write as [orange:-:b]%s[-:-:-] (ctrl+x)"
|
||||||
focusSwitcher = map[tview.Primitive]tview.Primitive{}
|
focusSwitcher = map[tview.Primitive]tview.Primitive{}
|
||||||
app *tview.Application
|
app *tview.Application
|
||||||
|
cliCardPath string
|
||||||
|
cliContinue bool
|
||||||
|
cliMsg string
|
||||||
)
|
)
|
||||||
|
|
||||||
func main() {
|
func main() {
|
||||||
|
flag.BoolVar(&cfg.CLIMode, "cli", false, "Run in CLI mode without TUI")
|
||||||
|
flag.BoolVar(&cfg.ToolUse, "tools", true, "run with tools")
|
||||||
|
flag.StringVar(&cliCardPath, "card", "", "Path to syscard JSON file")
|
||||||
|
flag.BoolVar(&cliContinue, "continue", false, "Continue from last chat (by agent or card)")
|
||||||
|
flag.StringVar(&cliMsg, "msg", "", "Send message and exit (one-shot mode)")
|
||||||
|
flag.Parse()
|
||||||
|
if cfg.CLIMode {
|
||||||
|
runCLIMode()
|
||||||
|
return
|
||||||
|
}
|
||||||
pages.AddPage("main", flex, true, true)
|
pages.AddPage("main", flex, true, true)
|
||||||
if err := app.SetRoot(pages,
|
if err := app.SetRoot(pages,
|
||||||
true).EnableMouse(cfg.EnableMouse).EnablePaste(true).Run(); err != nil {
|
true).EnableMouse(cfg.EnableMouse).EnablePaste(true).Run(); err != nil {
|
||||||
@@ -30,3 +54,236 @@ func main() {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func runCLIMode() {
|
||||||
|
outputHandler = &CLIOutputHandler{}
|
||||||
|
cliRespDone = make(chan bool, 1)
|
||||||
|
if cliCardPath != "" {
|
||||||
|
card, err := pngmeta.ReadCardJson(cliCardPath)
|
||||||
|
if err != nil {
|
||||||
|
fmt.Fprintf(os.Stderr, "Failed to load syscard: %v\n", err)
|
||||||
|
os.Exit(1)
|
||||||
|
}
|
||||||
|
cfg.AssistantRole = card.Role
|
||||||
|
sysMap[card.ID] = card
|
||||||
|
roleToID[card.Role] = card.ID
|
||||||
|
charToStart(card.Role, false)
|
||||||
|
fmt.Printf("Loaded syscard: %s (%s)\n", card.Role, card.FilePath)
|
||||||
|
}
|
||||||
|
if cliContinue {
|
||||||
|
if cliCardPath != "" {
|
||||||
|
history, err := loadAgentsLastChat(cfg.AssistantRole)
|
||||||
|
if err != nil {
|
||||||
|
fmt.Printf("No previous chat found for %s, starting new chat\n", cfg.AssistantRole)
|
||||||
|
startNewCLIChat()
|
||||||
|
} else {
|
||||||
|
chatBody.Messages = history
|
||||||
|
fmt.Printf("Continued chat: %s\n", activeChatName)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
chatBody.Messages = loadOldChatOrGetNew()
|
||||||
|
fmt.Printf("Continued chat: %s\n", activeChatName)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
startNewCLIChat()
|
||||||
|
}
|
||||||
|
printCLIWelcome()
|
||||||
|
go func() {
|
||||||
|
<-ctx.Done()
|
||||||
|
os.Exit(0)
|
||||||
|
}()
|
||||||
|
if cliMsg != "" {
|
||||||
|
persona := cfg.UserRole
|
||||||
|
if cfg.WriteNextMsgAs != "" {
|
||||||
|
persona = cfg.WriteNextMsgAs
|
||||||
|
}
|
||||||
|
chatRoundChan <- &models.ChatRoundReq{Role: persona, UserMsg: cliMsg}
|
||||||
|
<-cliRespDone
|
||||||
|
fmt.Println()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
scanner := bufio.NewScanner(os.Stdin)
|
||||||
|
for {
|
||||||
|
fmt.Print("> ")
|
||||||
|
if !scanner.Scan() {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
msg := scanner.Text()
|
||||||
|
if msg == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if strings.HasPrefix(msg, "/") {
|
||||||
|
if !handleCLICommand(msg) {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
fmt.Println()
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
persona := cfg.UserRole
|
||||||
|
if cfg.WriteNextMsgAs != "" {
|
||||||
|
persona = cfg.WriteNextMsgAs
|
||||||
|
}
|
||||||
|
chatRoundChan <- &models.ChatRoundReq{Role: persona, UserMsg: msg}
|
||||||
|
<-cliRespDone
|
||||||
|
fmt.Println()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func printCLIWelcome() {
|
||||||
|
fmt.Println("CLI Mode started. Type your messages or commands.")
|
||||||
|
fmt.Println("Type /help for available commands.")
|
||||||
|
fmt.Println()
|
||||||
|
}
|
||||||
|
|
||||||
|
func printCLIHelp() {
|
||||||
|
fmt.Println("Available commands:")
|
||||||
|
fmt.Println(" /help, /h - Show this help message")
|
||||||
|
fmt.Println(" /new, /n - Start a new chat (clears conversation)")
|
||||||
|
fmt.Println(" /card <path>, /c <path> - Load a different syscard")
|
||||||
|
fmt.Println(" /undo, /u - Delete last message")
|
||||||
|
fmt.Println(" /history, /ls - List chat history")
|
||||||
|
fmt.Println(" /load <name> - Load a specific chat by name")
|
||||||
|
fmt.Println(" /model <name>, /m <name> - Switch model")
|
||||||
|
fmt.Println(" /api <index>, /a <index> - Switch API link (no index to list)")
|
||||||
|
fmt.Println(" /quit, /q, /exit - Exit CLI mode")
|
||||||
|
fmt.Println()
|
||||||
|
fmt.Printf("Current syscard: %s\n", cfg.AssistantRole)
|
||||||
|
fmt.Printf("Current model: %s\n", chatBody.Model)
|
||||||
|
fmt.Printf("Current API: %s\n", cfg.CurrentAPI)
|
||||||
|
fmt.Println()
|
||||||
|
}
|
||||||
|
|
||||||
|
func handleCLICommand(msg string) bool {
|
||||||
|
parts := strings.Fields(msg)
|
||||||
|
cmd := strings.ToLower(parts[0])
|
||||||
|
args := parts[1:]
|
||||||
|
|
||||||
|
switch cmd {
|
||||||
|
case "/help", "/h":
|
||||||
|
printCLIHelp()
|
||||||
|
case "/new", "/n":
|
||||||
|
startNewCLIChat()
|
||||||
|
fmt.Println("New chat started.")
|
||||||
|
fmt.Printf("Syscard: %s\n", cfg.AssistantRole)
|
||||||
|
fmt.Println()
|
||||||
|
case "/card", "/c":
|
||||||
|
if len(args) == 0 {
|
||||||
|
fmt.Println("Usage: /card <path>")
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
card, err := pngmeta.ReadCardJson(args[0])
|
||||||
|
if err != nil {
|
||||||
|
fmt.Fprintf(os.Stderr, "Failed to load syscard: %v\n", err)
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
cfg.AssistantRole = card.Role
|
||||||
|
sysMap[card.ID] = card
|
||||||
|
roleToID[card.Role] = card.ID
|
||||||
|
charToStart(card.Role, false)
|
||||||
|
startNewCLIChat()
|
||||||
|
fmt.Printf("Switched to syscard: %s (%s)\n", card.Role, card.FilePath)
|
||||||
|
case "/undo", "/u":
|
||||||
|
if len(chatBody.Messages) == 0 {
|
||||||
|
fmt.Println("No messages to delete.")
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
chatBody.Messages = chatBody.Messages[:len(chatBody.Messages)-1]
|
||||||
|
cliPrevOutput = ""
|
||||||
|
fmt.Println("Last message deleted.")
|
||||||
|
case "/history", "/ls":
|
||||||
|
fmt.Println("Chat history:")
|
||||||
|
for name := range chatMap {
|
||||||
|
marker := " "
|
||||||
|
if name == activeChatName {
|
||||||
|
marker = "* "
|
||||||
|
}
|
||||||
|
fmt.Printf("%s%s\n", marker, name)
|
||||||
|
}
|
||||||
|
fmt.Println()
|
||||||
|
case "/load":
|
||||||
|
if len(args) == 0 {
|
||||||
|
fmt.Println("Usage: /load <name>")
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
name := args[0]
|
||||||
|
chat, ok := chatMap[name]
|
||||||
|
if !ok {
|
||||||
|
fmt.Printf("Chat not found: %s\n", name)
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
history, err := chat.ToHistory()
|
||||||
|
if err != nil {
|
||||||
|
fmt.Printf("Failed to load chat: %v\n", err)
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
chatBody.Messages = history
|
||||||
|
activeChatName = name
|
||||||
|
cfg.AssistantRole = chat.Agent
|
||||||
|
fmt.Printf("Loaded chat: %s\n", name)
|
||||||
|
case "/model", "/m":
|
||||||
|
getModelListForAPI := func(api string) []string {
|
||||||
|
if strings.Contains(api, "api.deepseek.com/") {
|
||||||
|
return []string{"deepseek-chat", "deepseek-reasoner"}
|
||||||
|
} else if strings.Contains(api, "openrouter.ai") {
|
||||||
|
return ORFreeModels
|
||||||
|
}
|
||||||
|
return LocalModels
|
||||||
|
}
|
||||||
|
modelList := getModelListForAPI(cfg.CurrentAPI)
|
||||||
|
if len(args) == 0 {
|
||||||
|
fmt.Println("Models:")
|
||||||
|
for i, model := range modelList {
|
||||||
|
marker := " "
|
||||||
|
if model == chatBody.Model {
|
||||||
|
marker = "* "
|
||||||
|
}
|
||||||
|
fmt.Printf("%s%d: %s\n", marker, i, model)
|
||||||
|
}
|
||||||
|
fmt.Printf("\nCurrent model: %s\n", chatBody.Model)
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
// Try index first, then model name
|
||||||
|
if idx, err := strconv.Atoi(args[0]); err == nil && idx >= 0 && idx < len(modelList) {
|
||||||
|
chatBody.Model = modelList[idx]
|
||||||
|
fmt.Printf("Switched to model: %s\n", chatBody.Model)
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
if slices.Index(modelList, args[0]) < 0 {
|
||||||
|
fmt.Printf("Model '%s' not found. Use index or choose from:\n", args[0])
|
||||||
|
for i, model := range modelList {
|
||||||
|
fmt.Printf(" %d: %s\n", i, model)
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
chatBody.Model = args[0]
|
||||||
|
fmt.Printf("Switched to model: %s\n", args[0])
|
||||||
|
case "/api", "/a":
|
||||||
|
if len(args) == 0 {
|
||||||
|
fmt.Println("API Links:")
|
||||||
|
for i, link := range cfg.ApiLinks {
|
||||||
|
marker := " "
|
||||||
|
if link == cfg.CurrentAPI {
|
||||||
|
marker = "* "
|
||||||
|
}
|
||||||
|
fmt.Printf("%s%d: %s\n", marker, i, link)
|
||||||
|
}
|
||||||
|
fmt.Printf("\nCurrent API: %s\n", cfg.CurrentAPI)
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
idx := 0
|
||||||
|
fmt.Sscanf(args[0], "%d", &idx)
|
||||||
|
if idx < 0 || idx >= len(cfg.ApiLinks) {
|
||||||
|
fmt.Printf("Invalid index. Valid range: 0-%d\n", len(cfg.ApiLinks)-1)
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
cfg.CurrentAPI = cfg.ApiLinks[idx]
|
||||||
|
fmt.Printf("Switched to API: %s\n", cfg.CurrentAPI)
|
||||||
|
case "/quit", "/q", "/exit":
|
||||||
|
fmt.Println("Goodbye!")
|
||||||
|
return false
|
||||||
|
default:
|
||||||
|
fmt.Printf("Unknown command: %s\n", msg)
|
||||||
|
fmt.Println("Type /help for available commands.")
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|||||||
@@ -1,8 +1,12 @@
|
|||||||
package models
|
package models
|
||||||
|
|
||||||
|
import "regexp"
|
||||||
|
|
||||||
const (
|
const (
|
||||||
LoadedMark = "(loaded) "
|
LoadedMark = "(loaded) "
|
||||||
ToolRespMultyType = "multimodel_content"
|
ToolRespMultyType = "multimodel_content"
|
||||||
|
DefaultFirstMsg = "Hello! What can I do for you?"
|
||||||
|
BasicSysMsg = "Large Language Model that helps user with any of his requests."
|
||||||
)
|
)
|
||||||
|
|
||||||
type APIType int
|
type APIType int
|
||||||
@@ -11,3 +15,17 @@ const (
|
|||||||
APITypeChat APIType = iota
|
APITypeChat APIType = iota
|
||||||
APITypeCompletion
|
APITypeCompletion
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
ToolCallRE = regexp.MustCompile(`__tool_call__\s*([\s\S]*?)__tool_call__`)
|
||||||
|
QuotesRE = regexp.MustCompile(`(".*?")`)
|
||||||
|
StarRE = regexp.MustCompile(`(\*.*?\*)`)
|
||||||
|
ThinkRE = regexp.MustCompile(`(?s)<think>.*?</think>`)
|
||||||
|
CodeBlockRE = regexp.MustCompile(`(?s)\x60{3}(?:.*?)\n(.*?)\n\s*\x60{3}\s*`)
|
||||||
|
SingleBacktickRE = regexp.MustCompile(`\x60([^\x60]*)\x60`)
|
||||||
|
RoleRE = regexp.MustCompile(`^(\w+):`)
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
SysLabels = []string{"assistant"}
|
||||||
|
)
|
||||||
|
|||||||
@@ -139,7 +139,7 @@ func showAPILinkSelectionPopup() {
|
|||||||
apiListWidget.SetSelectedFunc(func(index int, mainText string, secondaryText string, shortcut rune) {
|
apiListWidget.SetSelectedFunc(func(index int, mainText string, secondaryText string, shortcut rune) {
|
||||||
// Update the API in config
|
// Update the API in config
|
||||||
cfg.CurrentAPI = mainText
|
cfg.CurrentAPI = mainText
|
||||||
// updateToolCapabilities()
|
// tools.UpdateToolCapabilities()
|
||||||
// Update model list based on new API
|
// Update model list based on new API
|
||||||
// Helper function to get model list for a given API (same as in props_table.go)
|
// Helper function to get model list for a given API (same as in props_table.go)
|
||||||
getModelListForAPI := func(api string) []string {
|
getModelListForAPI := func(api string) []string {
|
||||||
@@ -159,7 +159,7 @@ func showAPILinkSelectionPopup() {
|
|||||||
if len(newModelList) > 0 && !slices.Contains(newModelList, chatBody.Model) {
|
if len(newModelList) > 0 && !slices.Contains(newModelList, chatBody.Model) {
|
||||||
chatBody.Model = strings.TrimPrefix(newModelList[0], models.LoadedMark)
|
chatBody.Model = strings.TrimPrefix(newModelList[0], models.LoadedMark)
|
||||||
cfg.CurrentModel = chatBody.Model
|
cfg.CurrentModel = chatBody.Model
|
||||||
updateToolCapabilities()
|
UpdateToolCapabilities()
|
||||||
}
|
}
|
||||||
pages.RemovePage("apiLinkSelectionPopup")
|
pages.RemovePage("apiLinkSelectionPopup")
|
||||||
app.SetFocus(textArea)
|
app.SetFocus(textArea)
|
||||||
|
|||||||
@@ -115,15 +115,9 @@ func makePropsTable(props map[string]float32) *tview.Table {
|
|||||||
row++
|
row++
|
||||||
}
|
}
|
||||||
// Add checkboxes
|
// Add checkboxes
|
||||||
addCheckboxRow("Inject role", injectRole, func(checked bool) {
|
|
||||||
injectRole = checked
|
|
||||||
})
|
|
||||||
addCheckboxRow("TTS Enabled", cfg.TTS_ENABLED, func(checked bool) {
|
addCheckboxRow("TTS Enabled", cfg.TTS_ENABLED, func(checked bool) {
|
||||||
cfg.TTS_ENABLED = checked
|
cfg.TTS_ENABLED = checked
|
||||||
})
|
})
|
||||||
addCheckboxRow("Auto clean tool calls from context", cfg.AutoCleanToolCallsFromCtx, func(checked bool) {
|
|
||||||
cfg.AutoCleanToolCallsFromCtx = checked
|
|
||||||
})
|
|
||||||
addCheckboxRow("Enable Mouse", cfg.EnableMouse, func(checked bool) {
|
addCheckboxRow("Enable Mouse", cfg.EnableMouse, func(checked bool) {
|
||||||
cfg.EnableMouse = checked
|
cfg.EnableMouse = checked
|
||||||
// Reconfigure the app's mouse setting
|
// Reconfigure the app's mouse setting
|
||||||
|
|||||||
@@ -7,7 +7,6 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"gf-lt/config"
|
"gf-lt/config"
|
||||||
"gf-lt/models"
|
"gf-lt/models"
|
||||||
"gf-lt/onnx"
|
|
||||||
"log/slog"
|
"log/slog"
|
||||||
"net/http"
|
"net/http"
|
||||||
"os"
|
"os"
|
||||||
@@ -157,6 +156,43 @@ type ONNXEmbedder struct {
|
|||||||
modelPath string
|
modelPath string
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var onnxInitOnce sync.Once
|
||||||
|
var onnxReady bool
|
||||||
|
var onnxLibPath string
|
||||||
|
var cudaLibPath string
|
||||||
|
|
||||||
|
var onnxLibPaths = []string{
|
||||||
|
"/usr/lib/libonnxruntime.so",
|
||||||
|
"/usr/lib/libonnxruntime.so.1.24.2",
|
||||||
|
"/usr/local/lib/libonnxruntime.so",
|
||||||
|
"/usr/lib/x86_64-linux-gnu/libonnxruntime.so",
|
||||||
|
"/opt/onnxruntime/lib/libonnxruntime.so",
|
||||||
|
}
|
||||||
|
|
||||||
|
var cudaLibPaths = []string{
|
||||||
|
"/usr/lib/libonnxruntime_providers_cuda.so",
|
||||||
|
"/usr/local/lib/libonnxruntime_providers_cuda.so",
|
||||||
|
"/opt/onnxruntime/lib/libonnxruntime_providers_cuda.so",
|
||||||
|
}
|
||||||
|
|
||||||
|
func findONNXLibrary() string {
|
||||||
|
for _, path := range onnxLibPaths {
|
||||||
|
if _, err := os.Stat(path); err == nil {
|
||||||
|
return path
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
|
func findCUDALibrary() string {
|
||||||
|
for _, path := range cudaLibPaths {
|
||||||
|
if _, err := os.Stat(path); err == nil {
|
||||||
|
return path
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
|
||||||
func NewONNXEmbedder(modelPath, tokenizerPath string, dims int, logger *slog.Logger) (*ONNXEmbedder, error) {
|
func NewONNXEmbedder(modelPath, tokenizerPath string, dims int, logger *slog.Logger) (*ONNXEmbedder, error) {
|
||||||
// Check if model and tokenizer files exist
|
// Check if model and tokenizer files exist
|
||||||
if _, err := os.Stat(modelPath); err != nil {
|
if _, err := os.Stat(modelPath); err != nil {
|
||||||
@@ -166,16 +202,17 @@ func NewONNXEmbedder(modelPath, tokenizerPath string, dims int, logger *slog.Log
|
|||||||
return nil, fmt.Errorf("tokenizer not found: %w", err)
|
return nil, fmt.Errorf("tokenizer not found: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Initialize ONNX runtime
|
// Find ONNX library
|
||||||
if err := onnx.Init(); err != nil {
|
onnxLibPath = findONNXLibrary()
|
||||||
return nil, fmt.Errorf("ONNX init failed: %w", err)
|
if onnxLibPath == "" {
|
||||||
}
|
return nil, errors.New("ONNX runtime library not found in standard locations")
|
||||||
if onnx.HasCUDASupport() {
|
|
||||||
logger.Info("ONNX CUDA support enabled")
|
|
||||||
} else {
|
|
||||||
logger.Info("ONNX using CPU fallback")
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Find CUDA provider library (optional)
|
||||||
|
cudaLibPath = findCUDALibrary()
|
||||||
|
if cudaLibPath == "" {
|
||||||
|
fmt.Println("WARNING: CUDA provider library not found, will use CPU")
|
||||||
|
}
|
||||||
emb := &ONNXEmbedder{
|
emb := &ONNXEmbedder{
|
||||||
tokenizerPath: tokenizerPath,
|
tokenizerPath: tokenizerPath,
|
||||||
dims: dims,
|
dims: dims,
|
||||||
@@ -202,12 +239,26 @@ func (e *ONNXEmbedder) ensureInitialized() error {
|
|||||||
}
|
}
|
||||||
e.tokenizer = tok
|
e.tokenizer = tok
|
||||||
}
|
}
|
||||||
// ONNX runtime already initialized by onnx.Init() in NewONNXEmbedder
|
onnxInitOnce.Do(func() {
|
||||||
if !onnx.IsReady() {
|
onnxruntime_go.SetSharedLibraryPath(onnxLibPath)
|
||||||
|
if err := onnxruntime_go.InitializeEnvironment(); err != nil {
|
||||||
|
e.logger.Error("failed to initialize ONNX runtime", "error", err)
|
||||||
|
onnxReady = false
|
||||||
|
return
|
||||||
|
}
|
||||||
|
// Register CUDA provider if available
|
||||||
|
if cudaLibPath != "" {
|
||||||
|
if err := onnxruntime_go.RegisterExecutionProviderLibrary("CUDA", cudaLibPath); err != nil {
|
||||||
|
e.logger.Warn("failed to register CUDA provider", "error", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
onnxReady = true
|
||||||
|
})
|
||||||
|
if !onnxReady {
|
||||||
return errors.New("ONNX runtime not ready")
|
return errors.New("ONNX runtime not ready")
|
||||||
}
|
}
|
||||||
// Create session options
|
// Create session options
|
||||||
opts, err := onnx.NewSessionOptions()
|
opts, err := onnxruntime_go.NewSessionOptions()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to create session options: %w", err)
|
return fmt.Errorf("failed to create session options: %w", err)
|
||||||
}
|
}
|
||||||
@@ -215,7 +266,27 @@ func (e *ONNXEmbedder) ensureInitialized() error {
|
|||||||
_ = opts.Destroy()
|
_ = opts.Destroy()
|
||||||
}()
|
}()
|
||||||
|
|
||||||
if onnx.HasCUDASupport() {
|
// Try to add CUDA provider
|
||||||
|
useCUDA := cudaLibPath != ""
|
||||||
|
if useCUDA {
|
||||||
|
cudaOpts, err := onnxruntime_go.NewCUDAProviderOptions()
|
||||||
|
if err != nil {
|
||||||
|
e.logger.Warn("failed to create CUDA provider options, falling back to CPU", "error", err)
|
||||||
|
useCUDA = false
|
||||||
|
} else {
|
||||||
|
defer func() {
|
||||||
|
_ = cudaOpts.Destroy()
|
||||||
|
}()
|
||||||
|
if err := cudaOpts.Update(map[string]string{"device_id": "0"}); err != nil {
|
||||||
|
e.logger.Warn("failed to update CUDA options, falling back to CPU", "error", err)
|
||||||
|
useCUDA = false
|
||||||
|
} else if err := opts.AppendExecutionProviderCUDA(cudaOpts); err != nil {
|
||||||
|
e.logger.Warn("failed to append CUDA provider, falling back to CPU", "error", err)
|
||||||
|
useCUDA = false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if useCUDA {
|
||||||
e.logger.Info("Using CUDA for ONNX inference")
|
e.logger.Info("Using CUDA for ONNX inference")
|
||||||
} else {
|
} else {
|
||||||
e.logger.Info("Using CPU for ONNX inference")
|
e.logger.Info("Using CPU for ONNX inference")
|
||||||
|
|||||||
287
rag/rag.go
287
rag/rag.go
@@ -12,6 +12,7 @@ import (
|
|||||||
"regexp"
|
"regexp"
|
||||||
"runtime"
|
"runtime"
|
||||||
"sort"
|
"sort"
|
||||||
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
@@ -24,11 +25,120 @@ const ()
|
|||||||
var (
|
var (
|
||||||
// Status messages for TUI integration
|
// Status messages for TUI integration
|
||||||
LongJobStatusCh = make(chan string, 100) // Increased buffer size for parallel batch updates
|
LongJobStatusCh = make(chan string, 100) // Increased buffer size for parallel batch updates
|
||||||
FinishedRAGStatus = "finished loading RAG file; press Enter"
|
FinishedRAGStatus = "finished loading RAG file; press x to exit"
|
||||||
LoadedFileRAGStatus = "loaded file"
|
LoadedFileRAGStatus = "loaded file"
|
||||||
ErrRAGStatus = "some error occurred; failed to transfer data to vector db"
|
ErrRAGStatus = "some error occurred; failed to transfer data to vector db"
|
||||||
|
|
||||||
|
// stopWords are common words that can be removed from queries when not part of phrases
|
||||||
|
stopWords = []string{"the", "a", "an", "and", "or", "but", "in", "on", "at", "to", "for", "of", "with", "by", "from", "up", "down", "left", "right", "about", "like", "such", "than", "then", "also", "too"}
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// isStopWord checks if a word is in the stop words list
|
||||||
|
func isStopWord(word string) bool {
|
||||||
|
for _, stop := range stopWords {
|
||||||
|
if strings.EqualFold(word, stop) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// detectPhrases returns multi-word phrases from a query that should be treated as units
|
||||||
|
func detectPhrases(query string) []string {
|
||||||
|
words := strings.Fields(strings.ToLower(query))
|
||||||
|
var phrases []string
|
||||||
|
for i := 0; i < len(words)-1; i++ {
|
||||||
|
word1 := strings.Trim(words[i], ".,!?;:'\"()[]{}")
|
||||||
|
word2 := strings.Trim(words[i+1], ".,!?;:'\"()[]{}")
|
||||||
|
|
||||||
|
// Skip if either word is a stop word or too short
|
||||||
|
if isStopWord(word1) || isStopWord(word2) || len(word1) < 2 || len(word2) < 2 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check if this pair appears to be a meaningful phrase
|
||||||
|
// Simple heuristic: consecutive non-stop words of reasonable length
|
||||||
|
phrase := word1 + " " + word2
|
||||||
|
phrases = append(phrases, phrase)
|
||||||
|
|
||||||
|
// Optionally check for 3-word phrases
|
||||||
|
if i < len(words)-2 {
|
||||||
|
word3 := strings.Trim(words[i+2], ".,!?;:'\"()[]{}")
|
||||||
|
if !isStopWord(word3) && len(word3) >= 2 {
|
||||||
|
phrases = append(phrases, word1+" "+word2+" "+word3)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return phrases
|
||||||
|
}
|
||||||
|
|
||||||
|
// countPhraseMatches returns the number of query phrases found in text
|
||||||
|
func countPhraseMatches(text, query string) int {
|
||||||
|
phrases := detectPhrases(query)
|
||||||
|
if len(phrases) == 0 {
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
textLower := strings.ToLower(text)
|
||||||
|
count := 0
|
||||||
|
for _, phrase := range phrases {
|
||||||
|
if strings.Contains(textLower, phrase) {
|
||||||
|
count++
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return count
|
||||||
|
}
|
||||||
|
|
||||||
|
// parseSlugIndices extracts batch and chunk indices from a slug
|
||||||
|
// slug format: filename_batch_chunk (e.g., "kjv_bible.epub_1786_0")
|
||||||
|
func parseSlugIndices(slug string) (batch, chunk int, ok bool) {
|
||||||
|
// Find the last two numbers separated by underscores
|
||||||
|
re := regexp.MustCompile(`_(\d+)_(\d+)$`)
|
||||||
|
matches := re.FindStringSubmatch(slug)
|
||||||
|
if matches == nil || len(matches) != 3 {
|
||||||
|
return 0, 0, false
|
||||||
|
}
|
||||||
|
batch, err1 := strconv.Atoi(matches[1])
|
||||||
|
chunk, err2 := strconv.Atoi(matches[2])
|
||||||
|
if err1 != nil || err2 != nil {
|
||||||
|
return 0, 0, false
|
||||||
|
}
|
||||||
|
return batch, chunk, true
|
||||||
|
}
|
||||||
|
|
||||||
|
// areSlugsAdjacent returns true if two slugs are from the same file and have sequential indices
|
||||||
|
func areSlugsAdjacent(slug1, slug2 string) bool {
|
||||||
|
// Extract filename prefix (everything before the last underscore sequence)
|
||||||
|
parts1 := strings.Split(slug1, "_")
|
||||||
|
parts2 := strings.Split(slug2, "_")
|
||||||
|
if len(parts1) < 3 || len(parts2) < 3 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// Compare filename prefixes (all parts except last two)
|
||||||
|
prefix1 := strings.Join(parts1[:len(parts1)-2], "_")
|
||||||
|
prefix2 := strings.Join(parts2[:len(parts2)-2], "_")
|
||||||
|
if prefix1 != prefix2 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
batch1, chunk1, ok1 := parseSlugIndices(slug1)
|
||||||
|
batch2, chunk2, ok2 := parseSlugIndices(slug2)
|
||||||
|
if !ok1 || !ok2 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check if they're in same batch and chunks are sequential
|
||||||
|
if batch1 == batch2 && (chunk1 == chunk2+1 || chunk2 == chunk1+1) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check if they're in sequential batches and chunk indices suggest continuity
|
||||||
|
// This is heuristic but useful for cross-batch adjacency
|
||||||
|
if (batch1 == batch2+1 && chunk1 == 0) || (batch2 == batch1+1 && chunk2 == 0) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
type RAG struct {
|
type RAG struct {
|
||||||
logger *slog.Logger
|
logger *slog.Logger
|
||||||
store storage.FullRepo
|
store storage.FullRepo
|
||||||
@@ -155,8 +265,8 @@ func createChunks(sentences []string, wordLimit, overlapWords uint32) []string {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func sanitizeFTSQuery(query string) string {
|
func sanitizeFTSQuery(query string) string {
|
||||||
// Remove double quotes and other problematic characters for FTS5
|
// Keep double quotes for FTS5 phrase matching
|
||||||
query = strings.ReplaceAll(query, "\"", " ")
|
// Remove other problematic characters
|
||||||
query = strings.ReplaceAll(query, "'", " ")
|
query = strings.ReplaceAll(query, "'", " ")
|
||||||
query = strings.ReplaceAll(query, ";", " ")
|
query = strings.ReplaceAll(query, ";", " ")
|
||||||
query = strings.ReplaceAll(query, "\\", " ")
|
query = strings.ReplaceAll(query, "\\", " ")
|
||||||
@@ -549,7 +659,6 @@ func (r *RAG) RemoveFile(filename string) error {
|
|||||||
var (
|
var (
|
||||||
queryRefinementPattern = regexp.MustCompile(`(?i)(based on my (vector db|vector db|vector database|rags?|past (conversations?|chat|messages?))|from my (files?|documents?|data|information|memory)|search (in|my) (vector db|database|rags?)|rag search for)`)
|
queryRefinementPattern = regexp.MustCompile(`(?i)(based on my (vector db|vector db|vector database|rags?|past (conversations?|chat|messages?))|from my (files?|documents?|data|information|memory)|search (in|my) (vector db|database|rags?)|rag search for)`)
|
||||||
importantKeywords = []string{"project", "architecture", "code", "file", "chat", "conversation", "topic", "summary", "details", "history", "previous", "my", "user", "me"}
|
importantKeywords = []string{"project", "architecture", "code", "file", "chat", "conversation", "topic", "summary", "details", "history", "previous", "my", "user", "me"}
|
||||||
stopWords = []string{"the", "a", "an", "and", "or", "but", "in", "on", "at", "to", "for", "of", "with", "by", "from", "up", "down", "left", "right"}
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func (r *RAG) RefineQuery(query string) string {
|
func (r *RAG) RefineQuery(query string) string {
|
||||||
@@ -561,10 +670,27 @@ func (r *RAG) RefineQuery(query string) string {
|
|||||||
if len(query) <= 3 {
|
if len(query) <= 3 {
|
||||||
return original
|
return original
|
||||||
}
|
}
|
||||||
|
// If query already contains double quotes, assume it's a phrase query and skip refinement
|
||||||
|
if strings.Contains(query, "\"") {
|
||||||
|
return original
|
||||||
|
}
|
||||||
query = strings.ToLower(query)
|
query = strings.ToLower(query)
|
||||||
words := strings.Fields(query)
|
words := strings.Fields(query)
|
||||||
if len(words) >= 3 {
|
if len(words) >= 3 {
|
||||||
|
// Detect phrases and protect words that are part of phrases
|
||||||
|
phrases := detectPhrases(query)
|
||||||
|
protectedWords := make(map[string]bool)
|
||||||
|
for _, phrase := range phrases {
|
||||||
|
for _, word := range strings.Fields(phrase) {
|
||||||
|
protectedWords[word] = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Remove stop words that are not protected
|
||||||
for _, stopWord := range stopWords {
|
for _, stopWord := range stopWords {
|
||||||
|
if protectedWords[stopWord] {
|
||||||
|
continue
|
||||||
|
}
|
||||||
wordPattern := `\b` + stopWord + `\b`
|
wordPattern := `\b` + stopWord + `\b`
|
||||||
re := regexp.MustCompile(wordPattern)
|
re := regexp.MustCompile(wordPattern)
|
||||||
query = re.ReplaceAllString(query, "")
|
query = re.ReplaceAllString(query, "")
|
||||||
@@ -673,13 +799,56 @@ func (r *RAG) GenerateQueryVariations(query string) []string {
|
|||||||
if !strings.HasSuffix(query, " summary") {
|
if !strings.HasSuffix(query, " summary") {
|
||||||
variations = append(variations, query+" summary")
|
variations = append(variations, query+" summary")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Add phrase-quoted variations for better FTS5 matching
|
||||||
|
phrases := detectPhrases(query)
|
||||||
|
if len(phrases) > 0 {
|
||||||
|
// Sort phrases by length descending to prioritize longer phrases
|
||||||
|
sort.Slice(phrases, func(i, j int) bool {
|
||||||
|
return len(phrases[i]) > len(phrases[j])
|
||||||
|
})
|
||||||
|
|
||||||
|
// Create a version with all phrases quoted
|
||||||
|
quotedQuery := query
|
||||||
|
for _, phrase := range phrases {
|
||||||
|
// Only quote if not already quoted
|
||||||
|
quotedPhrase := "\"" + phrase + "\""
|
||||||
|
if !strings.Contains(strings.ToLower(quotedQuery), strings.ToLower(quotedPhrase)) {
|
||||||
|
// Case-insensitive replacement of phrase with quoted version
|
||||||
|
re := regexp.MustCompile(`(?i)\b` + regexp.QuoteMeta(phrase) + `\b`)
|
||||||
|
quotedQuery = re.ReplaceAllString(quotedQuery, quotedPhrase)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Disabled malformed quoted query for now
|
||||||
|
// if quotedQuery != query {
|
||||||
|
// variations = append(variations, quotedQuery)
|
||||||
|
// }
|
||||||
|
|
||||||
|
// Also add individual phrase variations for short queries
|
||||||
|
if len(phrases) <= 5 {
|
||||||
|
for _, phrase := range phrases {
|
||||||
|
// Create a focused query with just this phrase quoted
|
||||||
|
// Keep original context but emphasize this phrase
|
||||||
|
quotedPhrase := "\"" + phrase + "\""
|
||||||
|
re := regexp.MustCompile(`(?i)\b` + regexp.QuoteMeta(phrase) + `\b`)
|
||||||
|
focusedQuery := re.ReplaceAllString(query, quotedPhrase)
|
||||||
|
if focusedQuery != query && focusedQuery != quotedQuery {
|
||||||
|
variations = append(variations, focusedQuery)
|
||||||
|
}
|
||||||
|
// Add the phrase alone (quoted) as a separate variation
|
||||||
|
variations = append(variations, quotedPhrase)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
return variations
|
return variations
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RAG) RerankResults(results []models.VectorRow, query string) []models.VectorRow {
|
func (r *RAG) RerankResults(results []models.VectorRow, query string) []models.VectorRow {
|
||||||
|
phraseCount := len(detectPhrases(query))
|
||||||
type scoredResult struct {
|
type scoredResult struct {
|
||||||
row models.VectorRow
|
row models.VectorRow
|
||||||
distance float32
|
distance float32
|
||||||
|
phraseMatches int
|
||||||
}
|
}
|
||||||
scored := make([]scoredResult, 0, len(results))
|
scored := make([]scoredResult, 0, len(results))
|
||||||
for i := range results {
|
for i := range results {
|
||||||
@@ -704,18 +873,52 @@ func (r *RAG) RerankResults(results []models.VectorRow, query string) []models.V
|
|||||||
if row.FileName == "chat" || strings.Contains(strings.ToLower(row.FileName), "conversation") {
|
if row.FileName == "chat" || strings.Contains(strings.ToLower(row.FileName), "conversation") {
|
||||||
score += 3
|
score += 3
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Phrase match bonus: extra points for containing detected phrases
|
||||||
|
phraseMatches := countPhraseMatches(row.RawText, query)
|
||||||
|
if phraseMatches > 0 {
|
||||||
|
// Significant bonus per phrase to prioritize exact phrase matches
|
||||||
|
r.logger.Debug("phrase match bonus", "slug", row.Slug, "phraseMatches", phraseMatches, "score", score)
|
||||||
|
score += float32(phraseMatches) * 100
|
||||||
|
}
|
||||||
|
|
||||||
|
// Cross-chunk adjacency bonus: if this chunk has adjacent siblings in results,
|
||||||
|
// boost score to promote narrative continuity
|
||||||
|
adjacentCount := 0
|
||||||
|
for _, other := range results {
|
||||||
|
if other.Slug == row.Slug {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if areSlugsAdjacent(row.Slug, other.Slug) {
|
||||||
|
adjacentCount++
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if adjacentCount > 0 {
|
||||||
|
// Bonus per adjacent chunk, but diminishing returns
|
||||||
|
score += float32(adjacentCount) * 4
|
||||||
|
}
|
||||||
distance := row.Distance - score/100
|
distance := row.Distance - score/100
|
||||||
scored = append(scored, scoredResult{row: row, distance: distance})
|
scored = append(scored, scoredResult{row: row, distance: distance, phraseMatches: phraseMatches})
|
||||||
}
|
}
|
||||||
sort.Slice(scored, func(i, j int) bool {
|
sort.Slice(scored, func(i, j int) bool {
|
||||||
return scored[i].distance < scored[j].distance
|
return scored[i].distance < scored[j].distance
|
||||||
})
|
})
|
||||||
unique := make([]models.VectorRow, 0)
|
unique := make([]models.VectorRow, 0)
|
||||||
seen := make(map[string]bool)
|
seen := make(map[string]bool)
|
||||||
|
maxPerFile := 2
|
||||||
|
if phraseCount > 0 {
|
||||||
|
maxPerFile = 10
|
||||||
|
}
|
||||||
fileCounts := make(map[string]int)
|
fileCounts := make(map[string]int)
|
||||||
for i := range scored {
|
for i := range scored {
|
||||||
if !seen[scored[i].row.Slug] {
|
if !seen[scored[i].row.Slug] {
|
||||||
if fileCounts[scored[i].row.FileName] >= 2 {
|
// Allow phrase-matching chunks to bypass per-file limit (up to +5 extra)
|
||||||
|
allowed := fileCounts[scored[i].row.FileName] < maxPerFile
|
||||||
|
if !allowed && scored[i].phraseMatches > 0 {
|
||||||
|
// If chunk has phrase matches, allow extra slots (up to maxPerFile + 5)
|
||||||
|
allowed = fileCounts[scored[i].row.FileName] < maxPerFile+5
|
||||||
|
}
|
||||||
|
if !allowed {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
seen[scored[i].row.Slug] = true
|
seen[scored[i].row.Slug] = true
|
||||||
@@ -723,8 +926,8 @@ func (r *RAG) RerankResults(results []models.VectorRow, query string) []models.V
|
|||||||
unique = append(unique, scored[i].row)
|
unique = append(unique, scored[i].row)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if len(unique) > 10 {
|
if len(unique) > 30 {
|
||||||
unique = unique[:10]
|
unique = unique[:30]
|
||||||
}
|
}
|
||||||
return unique
|
return unique
|
||||||
}
|
}
|
||||||
@@ -793,6 +996,7 @@ func (r *RAG) Search(query string, limit int) ([]models.VectorRow, error) {
|
|||||||
r.resetIdleTimer()
|
r.resetIdleTimer()
|
||||||
refined := r.RefineQuery(query)
|
refined := r.RefineQuery(query)
|
||||||
variations := r.GenerateQueryVariations(refined)
|
variations := r.GenerateQueryVariations(refined)
|
||||||
|
r.logger.Debug("query variations", "original", query, "refined", refined, "variations", variations)
|
||||||
|
|
||||||
// Collect embedding search results from all variations
|
// Collect embedding search results from all variations
|
||||||
var embResults []models.VectorRow
|
var embResults []models.VectorRow
|
||||||
@@ -824,17 +1028,35 @@ func (r *RAG) Search(query string, limit int) ([]models.VectorRow, error) {
|
|||||||
return embResults[i].Distance < embResults[j].Distance
|
return embResults[i].Distance < embResults[j].Distance
|
||||||
})
|
})
|
||||||
|
|
||||||
// Perform keyword search
|
// Perform keyword search on all variations
|
||||||
kwResults, err := r.searchKeyword(refined, limit*2)
|
var kwResults []models.VectorRow
|
||||||
|
seenKw := make(map[string]bool)
|
||||||
|
for _, q := range variations {
|
||||||
|
results, err := r.searchKeyword(q, limit)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
r.logger.Warn("keyword search failed, using only embeddings", "error", err)
|
r.logger.Debug("keyword search failed for variation", "error", err, "query", q)
|
||||||
kwResults = nil
|
continue
|
||||||
}
|
}
|
||||||
// Sort keyword results by distance (already sorted by BM25 score)
|
for _, row := range results {
|
||||||
// kwResults already sorted by distance (lower is better)
|
if !seenKw[row.Slug] {
|
||||||
|
seenKw[row.Slug] = true
|
||||||
|
kwResults = append(kwResults, row)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Sort keyword results by distance (lower is better)
|
||||||
|
sort.Slice(kwResults, func(i, j int) bool {
|
||||||
|
return kwResults[i].Distance < kwResults[j].Distance
|
||||||
|
})
|
||||||
|
|
||||||
// Combine using Reciprocal Rank Fusion (RRF)
|
// Combine using Reciprocal Rank Fusion (RRF)
|
||||||
const rrfK = 60
|
// Use smaller K for phrase-heavy queries to give more weight to top ranks
|
||||||
|
phraseCount := len(detectPhrases(query))
|
||||||
|
rrfK := 60.0
|
||||||
|
if phraseCount > 0 {
|
||||||
|
rrfK = 30.0
|
||||||
|
}
|
||||||
|
r.logger.Debug("RRF parameters", "phraseCount", phraseCount, "rrfK", rrfK, "query", query)
|
||||||
type scoredRow struct {
|
type scoredRow struct {
|
||||||
row models.VectorRow
|
row models.VectorRow
|
||||||
score float64
|
score float64
|
||||||
@@ -844,11 +1066,22 @@ func (r *RAG) Search(query string, limit int) ([]models.VectorRow, error) {
|
|||||||
for rank, row := range embResults {
|
for rank, row := range embResults {
|
||||||
score := 1.0 / (float64(rank) + rrfK)
|
score := 1.0 / (float64(rank) + rrfK)
|
||||||
scoreMap[row.Slug] += score
|
scoreMap[row.Slug] += score
|
||||||
|
if row.Slug == "kjv_bible.epub_1786_0" {
|
||||||
|
r.logger.Debug("target chunk embedding rank", "rank", rank, "score", score)
|
||||||
}
|
}
|
||||||
// Add keyword results
|
}
|
||||||
|
// Add keyword results with weight boost when phrases are present
|
||||||
|
kwWeight := 1.0
|
||||||
|
if phraseCount > 0 {
|
||||||
|
kwWeight = 100.0
|
||||||
|
}
|
||||||
|
r.logger.Debug("keyword weight", "kwWeight", kwWeight, "phraseCount", phraseCount)
|
||||||
for rank, row := range kwResults {
|
for rank, row := range kwResults {
|
||||||
score := 1.0 / (float64(rank) + rrfK)
|
score := kwWeight * (1.0 / (float64(rank) + rrfK))
|
||||||
scoreMap[row.Slug] += score
|
scoreMap[row.Slug] += score
|
||||||
|
if row.Slug == "kjv_bible.epub_1786_0" {
|
||||||
|
r.logger.Debug("target chunk keyword rank", "rank", rank, "score", score, "kwWeight", kwWeight, "rrfK", rrfK)
|
||||||
|
}
|
||||||
// Ensure row exists in combined results
|
// Ensure row exists in combined results
|
||||||
if _, exists := seen[row.Slug]; !exists {
|
if _, exists := seen[row.Slug]; !exists {
|
||||||
embResults = append(embResults, row)
|
embResults = append(embResults, row)
|
||||||
@@ -860,6 +1093,18 @@ func (r *RAG) Search(query string, limit int) ([]models.VectorRow, error) {
|
|||||||
score := scoreMap[row.Slug]
|
score := scoreMap[row.Slug]
|
||||||
scoredRows = append(scoredRows, scoredRow{row: row, score: score})
|
scoredRows = append(scoredRows, scoredRow{row: row, score: score})
|
||||||
}
|
}
|
||||||
|
// Debug: log scores for target chunk and top chunks
|
||||||
|
if strings.Contains(strings.ToLower(query), "bald") || strings.Contains(strings.ToLower(query), "she bears") {
|
||||||
|
for _, sr := range scoredRows {
|
||||||
|
if sr.row.Slug == "kjv_bible.epub_1786_0" {
|
||||||
|
r.logger.Debug("target chunk score", "slug", sr.row.Slug, "score", sr.score, "distance", sr.row.Distance)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Log top 5 scores
|
||||||
|
for i := 0; i < len(scoredRows) && i < 5; i++ {
|
||||||
|
r.logger.Debug("top scored row", "rank", i+1, "slug", scoredRows[i].row.Slug, "score", scoredRows[i].score, "distance", scoredRows[i].row.Distance)
|
||||||
|
}
|
||||||
|
}
|
||||||
// Sort by descending RRF score
|
// Sort by descending RRF score
|
||||||
sort.Slice(scoredRows, func(i, j int) bool {
|
sort.Slice(scoredRows, func(i, j int) bool {
|
||||||
return scoredRows[i].score > scoredRows[j].score
|
return scoredRows[i].score > scoredRows[j].score
|
||||||
@@ -938,3 +1183,11 @@ func (r *RAG) Destroy() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// SetEmbedderForTesting replaces the internal embedder with a mock.
|
||||||
|
// This function is only available when compiling with the "test" build tag.
|
||||||
|
func (r *RAG) SetEmbedderForTesting(e Embedder) {
|
||||||
|
r.mu.Lock()
|
||||||
|
defer r.mu.Unlock()
|
||||||
|
r.embedder = e
|
||||||
|
}
|
||||||
|
|||||||
409
rag/rag_integration_test.go
Normal file
409
rag/rag_integration_test.go
Normal file
@@ -0,0 +1,409 @@
|
|||||||
|
package rag
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"gf-lt/config"
|
||||||
|
"gf-lt/models"
|
||||||
|
"gf-lt/storage"
|
||||||
|
"log/slog"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
_ "github.com/glebarez/go-sqlite"
|
||||||
|
"github.com/jmoiron/sqlx"
|
||||||
|
)
|
||||||
|
|
||||||
|
// mockEmbedder returns zero vectors of a fixed dimension.
|
||||||
|
type mockEmbedder struct {
|
||||||
|
dim int
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *mockEmbedder) Embed(text string) ([]float32, error) {
|
||||||
|
vec := make([]float32, m.dim)
|
||||||
|
return vec, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (m *mockEmbedder) EmbedSlice(texts []string) ([][]float32, error) {
|
||||||
|
vecs := make([][]float32, len(texts))
|
||||||
|
for i := range vecs {
|
||||||
|
vecs[i] = make([]float32, m.dim)
|
||||||
|
}
|
||||||
|
return vecs, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// dummyStore implements storage.FullRepo with a minimal set of methods.
|
||||||
|
// Only DB() is used by VectorStorage; other methods return empty values.
|
||||||
|
type dummyStore struct {
|
||||||
|
db *sqlx.DB
|
||||||
|
}
|
||||||
|
|
||||||
|
func (d dummyStore) DB() *sqlx.DB { return d.db }
|
||||||
|
|
||||||
|
// ChatHistory methods
|
||||||
|
func (d dummyStore) ListChats() ([]models.Chat, error) { return nil, nil }
|
||||||
|
func (d dummyStore) GetChatByID(id uint32) (*models.Chat, error) { return nil, nil }
|
||||||
|
func (d dummyStore) GetChatByChar(char string) ([]models.Chat, error) { return nil, nil }
|
||||||
|
func (d dummyStore) GetLastChat() (*models.Chat, error) { return nil, nil }
|
||||||
|
func (d dummyStore) GetLastChatByAgent(agent string) (*models.Chat, error) { return nil, nil }
|
||||||
|
func (d dummyStore) UpsertChat(chat *models.Chat) (*models.Chat, error) { return chat, nil }
|
||||||
|
func (d dummyStore) RemoveChat(id uint32) error { return nil }
|
||||||
|
func (d dummyStore) ChatGetMaxID() (uint32, error) { return 0, nil }
|
||||||
|
|
||||||
|
// Memories methods
|
||||||
|
func (d dummyStore) Memorise(m *models.Memory) (*models.Memory, error) { return m, nil }
|
||||||
|
func (d dummyStore) Recall(agent, topic string) (string, error) { return "", nil }
|
||||||
|
func (d dummyStore) RecallTopics(agent string) ([]string, error) { return nil, nil }
|
||||||
|
func (d dummyStore) Forget(agent, topic string) error { return nil }
|
||||||
|
|
||||||
|
// VectorRepo methods (not used but required by interface)
|
||||||
|
func (d dummyStore) WriteVector(row *models.VectorRow) error { return nil }
|
||||||
|
func (d dummyStore) SearchClosest(q []float32, limit int) ([]models.VectorRow, error) {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
func (d dummyStore) ListFiles() ([]string, error) { return nil, nil }
|
||||||
|
func (d dummyStore) RemoveEmbByFileName(filename string) error { return nil }
|
||||||
|
|
||||||
|
var _ storage.FullRepo = dummyStore{}
|
||||||
|
|
||||||
|
// setupTestRAG creates an in‑memory SQLite database, creates the necessary tables,
|
||||||
|
// inserts the provided chunks, and returns a RAG instance with a mock embedder.
|
||||||
|
func setupTestRAG(t *testing.T, chunks []*models.VectorRow) (*RAG, error) {
|
||||||
|
t.Helper()
|
||||||
|
db, err := sqlx.Open("sqlite", ":memory:")
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("open in‑memory db: %w", err)
|
||||||
|
}
|
||||||
|
// Create the required tables (embeddings_768 and fts_embeddings).
|
||||||
|
// Use the same schema as production.
|
||||||
|
_, err = db.Exec(`
|
||||||
|
CREATE TABLE embeddings_768 (
|
||||||
|
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||||
|
embeddings BLOB NOT NULL,
|
||||||
|
slug TEXT NOT NULL,
|
||||||
|
raw_text TEXT NOT NULL,
|
||||||
|
filename TEXT NOT NULL DEFAULT ''
|
||||||
|
);
|
||||||
|
`)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("create embeddings table: %w", err)
|
||||||
|
}
|
||||||
|
_, err = db.Exec(`
|
||||||
|
CREATE VIRTUAL TABLE fts_embeddings USING fts5(
|
||||||
|
slug UNINDEXED,
|
||||||
|
raw_text,
|
||||||
|
filename UNINDEXED,
|
||||||
|
embedding_size UNINDEXED,
|
||||||
|
tokenize='porter unicode61'
|
||||||
|
);
|
||||||
|
`)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("create FTS table: %w", err)
|
||||||
|
}
|
||||||
|
// Create a logger that discards output.
|
||||||
|
logger := slog.New(slog.NewTextHandler(nil, &slog.HandlerOptions{Level: slog.LevelError}))
|
||||||
|
store := dummyStore{db: db}
|
||||||
|
// Create config with embedding dimension 768.
|
||||||
|
cfg := &config.Config{
|
||||||
|
EmbedDims: 768,
|
||||||
|
RAGWordLimit: 250,
|
||||||
|
RAGOverlapWords: 25,
|
||||||
|
RAGBatchSize: 1,
|
||||||
|
}
|
||||||
|
// Create a RAG instance using New, which will create an embedder based on config.
|
||||||
|
// We'll override the embedder afterwards via reflection.
|
||||||
|
rag, err := New(logger, store, cfg)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("create RAG: %w", err)
|
||||||
|
}
|
||||||
|
// Replace the embedder with our mock.
|
||||||
|
rag.SetEmbedderForTesting(&mockEmbedder{dim: cfg.EmbedDims})
|
||||||
|
// Insert the provided chunks using the storage directly.
|
||||||
|
if len(chunks) > 0 {
|
||||||
|
// Ensure each chunk has embeddings of correct dimension (zero vector).
|
||||||
|
for _, chunk := range chunks {
|
||||||
|
if len(chunk.Embeddings) != cfg.EmbedDims {
|
||||||
|
chunk.Embeddings = make([]float32, cfg.EmbedDims)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
err = rag.storage.WriteVectors(chunks)
|
||||||
|
if err != nil {
|
||||||
|
return nil, fmt.Errorf("write test chunks: %w", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return rag, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// createTestChunks returns a slice of VectorRow representing the target chunk
|
||||||
|
// (kjv_bible.epub_1786_0), several bald‑related noise chunks, and unrelated chunks.
|
||||||
|
func createTestChunks() []*models.VectorRow {
|
||||||
|
// Target chunk: 2 Kings 2:23‑24 containing "bald head" and "two she bears".
|
||||||
|
targetRaw := `And he said, Ye shall not send.
|
||||||
|
|
||||||
|
|
||||||
|
2:17 And when they urged him till he was ashamed, he said, Send. They sent
|
||||||
|
therefore fifty men; and they sought three days, but found him not.
|
||||||
|
|
||||||
|
|
||||||
|
2:18 And when they came again to him, (for he tarried at Jericho,) he said unto
|
||||||
|
them, Did I not say unto you, Go not? 2:19 And the men of the city said unto
|
||||||
|
Elisha, Behold, I pray thee, the situation of this city is pleasant, as my lord
|
||||||
|
seeth: but the water is naught, and the ground barren.
|
||||||
|
|
||||||
|
|
||||||
|
2:20 And he said, Bring me a new cruse, and put salt therein. And they brought
|
||||||
|
it to him.
|
||||||
|
|
||||||
|
|
||||||
|
2:21 And he went forth unto the spring of the waters, and cast the salt in
|
||||||
|
there, and said, Thus saith the LORD, I have healed these waters; there shall
|
||||||
|
not be from thence any more death or barren land.
|
||||||
|
|
||||||
|
|
||||||
|
2:22 So the waters were healed unto this day, according to the saying of Elisha
|
||||||
|
which he spake.
|
||||||
|
|
||||||
|
|
||||||
|
2:23 And he went up from thence unto Bethel: and as he was going up by the way,
|
||||||
|
there came forth little children out of the city, and mocked him, and said unto
|
||||||
|
him, Go up, thou bald head; go up, thou bald head.
|
||||||
|
|
||||||
|
|
||||||
|
2:24 And he turned back, and looked on them, and cursed them in the name of the
|
||||||
|
LORD. And there came forth two she bears out of the wood, and tare forty and
|
||||||
|
two children of them.`
|
||||||
|
// Noise chunk 1: Leviticus containing "bald locust"
|
||||||
|
noise1Raw := `11:12 Whatsoever hath no fins nor scales in the waters, that shall be an
|
||||||
|
abomination unto you.
|
||||||
|
|
||||||
|
|
||||||
|
11:13 And these are they which ye shall have in abomination among the fowls;
|
||||||
|
they shall not be eaten, they are an abomination: the eagle, and the ossifrage,
|
||||||
|
and the ospray, 11:14 And the vulture, and the kite after his kind; 11:15 Every
|
||||||
|
raven after his kind; 11:16 And the owl, and the night hawk, and the cuckow,
|
||||||
|
and the hawk after his kind, 11:17 And the little owl, and the cormorant, and
|
||||||
|
the great owl, 11:18 And the swan, and the pelican, and the gier eagle, 11:19
|
||||||
|
And the stork, the heron after her kind, and the lapwing, and the bat.
|
||||||
|
|
||||||
|
|
||||||
|
11:20 All fowls that creep, going upon all four, shall be an abomination unto
|
||||||
|
you.
|
||||||
|
|
||||||
|
|
||||||
|
11:21 Yet these may ye eat of every flying creeping thing that goeth upon all
|
||||||
|
four, which have legs above their feet, to leap withal upon the earth; 11:22
|
||||||
|
Even these of them ye may eat; the locust after his kind, and the bald locust
|
||||||
|
after his kind, and the beetle after his kind, and the grasshopper after his
|
||||||
|
kind.
|
||||||
|
|
||||||
|
|
||||||
|
11:23 But all other flying creeping things, which have four feet, shall be an
|
||||||
|
abomination unto you.
|
||||||
|
|
||||||
|
|
||||||
|
11:24 And for these ye shall be unclean: whosoever toucheth the carcase of them
|
||||||
|
shall be unclean until the even.`
|
||||||
|
// Noise chunk 2: Leviticus containing "bald"
|
||||||
|
noise2Raw := `11:13 And these are they which ye shall have in abomination among the fowls;
|
||||||
|
they shall not be eaten, they are an abomination: the eagle, and the ossifrage,
|
||||||
|
and the ospray, 11:14 And the vulture, and the kite after his kind; 11:15 Every
|
||||||
|
raven after his kind; 11:16 And the owl, and the night hawk, and the cuckow,
|
||||||
|
and the hawk after his kind, 11:17 And the little owl, and the cormorant, and
|
||||||
|
the great owl, 11:18 And the swan, and the pelican, and the gier eagle, 11:19
|
||||||
|
And the stork, the heron after her kind, and the lapwing, and the bat.
|
||||||
|
|
||||||
|
|
||||||
|
11:20 All fowls that creep, going upon all four, shall be an abomination unto
|
||||||
|
you.
|
||||||
|
|
||||||
|
|
||||||
|
11:21 Yet these may ye eat of every flying creeping thing that goeth upon all
|
||||||
|
four, which have legs above their feet, to leap withal upon the earth; 11:22
|
||||||
|
Even these of them ye may eat; the locust after his kind, and the bald locust
|
||||||
|
after his kind, and the beetle after his kind, and the grasshopper after his
|
||||||
|
kind.
|
||||||
|
|
||||||
|
|
||||||
|
11:23 But all other flying creeping things, which have four feet, shall be an
|
||||||
|
abomination unto you.
|
||||||
|
|
||||||
|
|
||||||
|
11:24 And for these ye shall be unclean: whosoever toucheth the carcase of them
|
||||||
|
shall be unclean until the even.`
|
||||||
|
// Additional Leviticus noise chunks (simulating 28 bald-related chunks)
|
||||||
|
// Using variations of the same text with different slugs
|
||||||
|
leviticusSlugs := []string{
|
||||||
|
"kjv_bible.epub_564_0",
|
||||||
|
"kjv_bible.epub_565_0",
|
||||||
|
"kjv_bible.epub_579_0",
|
||||||
|
"kjv_bible.epub_580_0",
|
||||||
|
"kjv_bible.epub_581_0",
|
||||||
|
"kjv_bible.epub_582_0",
|
||||||
|
"kjv_bible.epub_583_0",
|
||||||
|
"kjv_bible.epub_584_0",
|
||||||
|
"kjv_bible.epub_585_0",
|
||||||
|
"kjv_bible.epub_586_0",
|
||||||
|
"kjv_bible.epub_587_0",
|
||||||
|
"kjv_bible.epub_588_0",
|
||||||
|
"kjv_bible.epub_589_0",
|
||||||
|
"kjv_bible.epub_590_0",
|
||||||
|
}
|
||||||
|
leviticusTexts := []string{
|
||||||
|
noise1Raw,
|
||||||
|
noise2Raw,
|
||||||
|
`13:40 And the man whose hair is fallen off his head, he is bald; yet is he
|
||||||
|
clean.
|
||||||
|
|
||||||
|
|
||||||
|
13:41 And he that hath his hair fallen off from the part of his head toward his
|
||||||
|
face, he is forehead bald; yet is he clean.`,
|
||||||
|
`13:42 And if there be in the bald head, or bald forehead, a white reddish sore;
|
||||||
|
it is a leprosy sprung up in his bald head, or his bald forehead.`,
|
||||||
|
`13:43 Then the priest shall look upon it: and, behold, if the rising of the
|
||||||
|
sore be white reddish in his bald head, or in his bald forehead, as the leprosy
|
||||||
|
appearedh in the skin of the flesh;`,
|
||||||
|
`13:44 He is a leprous man, he is unclean: the priest shall pronounce him utterly
|
||||||
|
unclean; his plague is in his head.`,
|
||||||
|
`13:45 And the leper in whom the plague is, his clothes shall be rent, and his
|
||||||
|
head bare, and he shall put a covering upon his upper lip, and shall cry,
|
||||||
|
Unclean, unclean.`,
|
||||||
|
`13:46 All the days wherein the plague shall be in him he shall be defiled; he
|
||||||
|
is unclean: he shall dwell alone; without the camp shall his habitation be.`,
|
||||||
|
`13:47 The garment also that the plague of leprosy is in, whether it be a woollen
|
||||||
|
garment, or a linen garment;`,
|
||||||
|
`13:48 Whether it be in the warp, or woof; of linen, or of woollen; whether in a
|
||||||
|
skin, or in any thing made of skin;`,
|
||||||
|
`13:49 And if the plague be greenish or reddish in the garment, or in the skin,
|
||||||
|
either in the warp, or in the woof, or in any thing of skin; it is a plague of
|
||||||
|
leprosy, and shall be shewed unto the priest:`,
|
||||||
|
`13:50 And the priest shall look upon the plague, and shut up it that hath the
|
||||||
|
plague seven days:`,
|
||||||
|
`13:51 And he shall look on the plague on the seventh day: if the plague be spread
|
||||||
|
in the garment, either in the warp, or in the woof, or in a skin, or in any work
|
||||||
|
that is made of skin; the plague is a fretting leprosy; it is unclean.`,
|
||||||
|
`13:52 He shall therefore burn that garment, whether warp or woof, in woollen or
|
||||||
|
in linen, or any thing of skin, wherein the plague is: for it is a fretting
|
||||||
|
leprosy; it shall be burnt in the fire.`,
|
||||||
|
}
|
||||||
|
// Unrelated chunk 1: ghost_7.txt_777_0
|
||||||
|
unrelated1Raw := `Doesn’t he have any pride as a hunter?!
|
||||||
|
|
||||||
|
I didn’t see what other choice I had. I would just have to grovel and be ready to flee at any given moment.
|
||||||
|
The Hidden Curse clan house was in the central region of the imperial capital. It was a high-class area with extraordinary property values that hosted the residences of people like Lord Gladis. This district was near the Imperial Castle, though “near” was a
|
||||||
|
relative term as it was still a few kilometers away.
|
||||||
|
|
||||||
|
The clan house was made of brick and conformed to an older style of architecture.`
|
||||||
|
// Unrelated chunk 2: ghost_7.txt_778_0
|
||||||
|
unrelated2Raw := `I would just have to grovel and be ready to flee at any given moment.
|
||||||
|
The Hidden Curse clan house was in the central region of the imperial capital. It was a high-class area with extraordinary property values that hosted the residences of people like Lord Gladis. This district was near the Imperial Castle, though “near” was a
|
||||||
|
relative term as it was still a few kilometers away.
|
||||||
|
|
||||||
|
The clan house was made of brick and conformed to an older style of architecture. Nearly everyone knew about this mansion and its clock tower. It stood tall over the neighboring mansions and rumor had it that you could see the whole capital from the top. It
|
||||||
|
spoke to this clan’s renown and history that they were able to get away with building something that dwarfed the mansions of the nobility.`
|
||||||
|
chunks := []*models.VectorRow{
|
||||||
|
{
|
||||||
|
Slug: "kjv_bible.epub_1786_0",
|
||||||
|
RawText: targetRaw,
|
||||||
|
FileName: "kjv_bible.epub",
|
||||||
|
Embeddings: nil, // will be filled with zero vector later
|
||||||
|
},
|
||||||
|
}
|
||||||
|
// Add Leviticus noise chunks
|
||||||
|
for i, slug := range leviticusSlugs {
|
||||||
|
text := leviticusTexts[i%len(leviticusTexts)]
|
||||||
|
chunks = append(chunks, &models.VectorRow{
|
||||||
|
Slug: slug,
|
||||||
|
RawText: text,
|
||||||
|
FileName: "kjv_bible.epub",
|
||||||
|
Embeddings: nil,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
// Add unrelated chunks
|
||||||
|
chunks = append(chunks,
|
||||||
|
&models.VectorRow{
|
||||||
|
Slug: "ghost_7.txt_777_0",
|
||||||
|
RawText: unrelated1Raw,
|
||||||
|
FileName: "ghost_7.txt",
|
||||||
|
Embeddings: nil,
|
||||||
|
},
|
||||||
|
&models.VectorRow{
|
||||||
|
Slug: "ghost_7.txt_778_0",
|
||||||
|
RawText: unrelated2Raw,
|
||||||
|
FileName: "ghost_7.txt",
|
||||||
|
Embeddings: nil,
|
||||||
|
},
|
||||||
|
)
|
||||||
|
return chunks
|
||||||
|
}
|
||||||
|
func assertTargetInTopN(t *testing.T, results []models.VectorRow, topN int) bool {
|
||||||
|
t.Helper()
|
||||||
|
for i, row := range results {
|
||||||
|
if i >= topN {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
if row.Slug == "kjv_bible.epub_1786_0" {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestBiblicalQuery(t *testing.T) {
|
||||||
|
chunks := createTestChunks()
|
||||||
|
rag, err := setupTestRAG(t, chunks)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("setup failed: %v", err)
|
||||||
|
}
|
||||||
|
query := "bald prophet and two she bears"
|
||||||
|
results, err := rag.Search(query, 10)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("search failed: %v", err)
|
||||||
|
}
|
||||||
|
// The target chunk should be in the top results.
|
||||||
|
if !assertTargetInTopN(t, results, 5) {
|
||||||
|
t.Errorf("target chunk not found in top 5 results for query %q", query)
|
||||||
|
t.Logf("results slugs: %v", func() []string {
|
||||||
|
slugs := make([]string, len(results))
|
||||||
|
for i, r := range results {
|
||||||
|
slugs[i] = r.Slug
|
||||||
|
}
|
||||||
|
return slugs
|
||||||
|
}())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestQueryVariations(t *testing.T) {
|
||||||
|
chunks := createTestChunks()
|
||||||
|
rag, err := setupTestRAG(t, chunks)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("setup failed: %v", err)
|
||||||
|
}
|
||||||
|
tests := []struct {
|
||||||
|
name string
|
||||||
|
query string
|
||||||
|
topN int
|
||||||
|
}{
|
||||||
|
{"she bears", "she bears", 5},
|
||||||
|
{"bald head", "bald head", 5},
|
||||||
|
{"two she bears out of the wood", "two she bears out of the wood", 5},
|
||||||
|
{"bald prophet", "bald prophet", 10},
|
||||||
|
{"go up thou bald head", "\"go up thou bald head\"", 5},
|
||||||
|
{"two she bears", "\"two she bears\"", 5},
|
||||||
|
}
|
||||||
|
for _, tt := range tests {
|
||||||
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
|
results, err := rag.Search(tt.query, 10)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("search failed: %v", err)
|
||||||
|
}
|
||||||
|
if !assertTargetInTopN(t, results, tt.topN) {
|
||||||
|
t.Errorf("target chunk not found in top %d results for query %q", tt.topN, tt.query)
|
||||||
|
t.Logf("results slugs: %v", func() []string {
|
||||||
|
slugs := make([]string, len(results))
|
||||||
|
for i, r := range results {
|
||||||
|
slugs[i] = r.Slug
|
||||||
|
}
|
||||||
|
return slugs
|
||||||
|
}())
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
129
rag/rag_real_test.go
Normal file
129
rag/rag_real_test.go
Normal file
@@ -0,0 +1,129 @@
|
|||||||
|
package rag
|
||||||
|
|
||||||
|
import (
|
||||||
|
"gf-lt/config"
|
||||||
|
"gf-lt/storage"
|
||||||
|
"log/slog"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
|
"testing"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestRealBiblicalQuery(t *testing.T) {
|
||||||
|
if testing.Short() {
|
||||||
|
t.Skip("skipping real embedder test in short mode")
|
||||||
|
}
|
||||||
|
// Check if the embedder model exists
|
||||||
|
modelPath := filepath.Join("..", "onnx", "embedgemma", "model_q4.onnx")
|
||||||
|
if _, err := os.Stat(modelPath); os.IsNotExist(err) {
|
||||||
|
t.Skipf("embedder model not found at %s; skipping real embedder test", modelPath)
|
||||||
|
}
|
||||||
|
tokenizerPath := filepath.Join("..", "onnx", "embedgemma", "tokenizer.json")
|
||||||
|
dbPath := filepath.Join("..", "gflt.db")
|
||||||
|
if _, err := os.Stat(dbPath); os.IsNotExist(err) {
|
||||||
|
t.Skipf("database not found at %s; skipping real embedder test", dbPath)
|
||||||
|
}
|
||||||
|
cfg := &config.Config{
|
||||||
|
EmbedModelPath: modelPath,
|
||||||
|
EmbedTokenizerPath: tokenizerPath,
|
||||||
|
EmbedDims: 768,
|
||||||
|
RAGWordLimit: 250,
|
||||||
|
RAGOverlapWords: 25,
|
||||||
|
RAGBatchSize: 1,
|
||||||
|
}
|
||||||
|
logger := slog.New(slog.NewTextHandler(nil, &slog.HandlerOptions{Level: slog.LevelError}))
|
||||||
|
store := storage.NewProviderSQL(dbPath, logger)
|
||||||
|
if store == nil {
|
||||||
|
t.Fatal("failed to create storage provider")
|
||||||
|
}
|
||||||
|
rag, err := New(logger, store, cfg)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("failed to create RAG instance: %v", err)
|
||||||
|
}
|
||||||
|
t.Cleanup(func() { rag.Destroy() })
|
||||||
|
query := "bald prophet and two she bears"
|
||||||
|
results, err := rag.Search(query, 30)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("search failed: %v", err)
|
||||||
|
}
|
||||||
|
found := false
|
||||||
|
for i, row := range results {
|
||||||
|
if row.Slug == "kjv_bible.epub_1786_0" {
|
||||||
|
found = true
|
||||||
|
t.Logf("target chunk found at rank %d", i+1)
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if !found {
|
||||||
|
t.Errorf("target chunk not found in search results for query %q", query)
|
||||||
|
t.Logf("results slugs:")
|
||||||
|
for i, r := range results {
|
||||||
|
t.Logf("%d: %s", i+1, r.Slug)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestRealQueryVariations(t *testing.T) {
|
||||||
|
if testing.Short() {
|
||||||
|
t.Skip("skipping real embedder test in short mode")
|
||||||
|
}
|
||||||
|
modelPath := filepath.Join("..", "onnx", "embedgemma", "model_q4.onnx")
|
||||||
|
if _, err := os.Stat(modelPath); os.IsNotExist(err) {
|
||||||
|
t.Skipf("embedder model not found at %s; skipping real embedder test", modelPath)
|
||||||
|
}
|
||||||
|
tokenizerPath := filepath.Join("..", "onnx", "embedgemma", "tokenizer.json")
|
||||||
|
dbPath := filepath.Join("..", "gflt.db")
|
||||||
|
if _, err := os.Stat(dbPath); os.IsNotExist(err) {
|
||||||
|
t.Skipf("database not found at %s; skipping real embedder test", dbPath)
|
||||||
|
}
|
||||||
|
cfg := &config.Config{
|
||||||
|
EmbedModelPath: modelPath,
|
||||||
|
EmbedTokenizerPath: tokenizerPath,
|
||||||
|
EmbedDims: 768,
|
||||||
|
RAGWordLimit: 250,
|
||||||
|
RAGOverlapWords: 25,
|
||||||
|
RAGBatchSize: 1,
|
||||||
|
}
|
||||||
|
logger := slog.New(slog.NewTextHandler(nil, &slog.HandlerOptions{Level: slog.LevelError}))
|
||||||
|
store := storage.NewProviderSQL(dbPath, logger)
|
||||||
|
if store == nil {
|
||||||
|
t.Fatal("failed to create storage provider")
|
||||||
|
}
|
||||||
|
rag, err := New(logger, store, cfg)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("failed to create RAG instance: %v", err)
|
||||||
|
}
|
||||||
|
t.Cleanup(func() { rag.Destroy() })
|
||||||
|
tests := []struct {
|
||||||
|
name string
|
||||||
|
query string
|
||||||
|
}{
|
||||||
|
{"she bears", "she bears"},
|
||||||
|
{"bald head", "bald head"},
|
||||||
|
{"two she bears out of the wood", "two she bears out of the wood"},
|
||||||
|
{"bald prophet", "bald prophet"},
|
||||||
|
{"go up thou bald head", "\"go up thou bald head\""},
|
||||||
|
{"two she bears", "\"two she bears\""},
|
||||||
|
}
|
||||||
|
for _, tt := range tests {
|
||||||
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
|
results, err := rag.Search(tt.query, 10)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("search failed: %v", err)
|
||||||
|
}
|
||||||
|
found := false
|
||||||
|
for _, row := range results {
|
||||||
|
if row.Slug == "kjv_bible.epub_1786_0" {
|
||||||
|
found = true
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if !found {
|
||||||
|
t.Errorf("target chunk not found for query %q", tt.query)
|
||||||
|
for i, r := range results {
|
||||||
|
t.Logf("%d: %s", i+1, r.Slug)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
151
rag/rag_test.go
Normal file
151
rag/rag_test.go
Normal file
@@ -0,0 +1,151 @@
|
|||||||
|
package rag
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestDetectPhrases(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
query string
|
||||||
|
expect []string
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
query: "bald prophet and two she bears",
|
||||||
|
expect: []string{"bald prophet", "two she", "two she bears", "she bears"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
query: "she bears",
|
||||||
|
expect: []string{"she bears"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
query: "the quick brown fox",
|
||||||
|
expect: []string{"quick brown", "quick brown fox", "brown fox"},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
query: "in the house", // stop words
|
||||||
|
expect: []string{}, // "in" and "the" are stop words
|
||||||
|
},
|
||||||
|
{
|
||||||
|
query: "a", // short
|
||||||
|
expect: []string{},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
for _, tt := range tests {
|
||||||
|
got := detectPhrases(tt.query)
|
||||||
|
if len(got) != len(tt.expect) {
|
||||||
|
t.Errorf("detectPhrases(%q) = %v, want %v", tt.query, got, tt.expect)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
for i := range got {
|
||||||
|
if got[i] != tt.expect[i] {
|
||||||
|
t.Errorf("detectPhrases(%q) = %v, want %v", tt.query, got, tt.expect)
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestCountPhraseMatches(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
text string
|
||||||
|
query string
|
||||||
|
expect int
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
text: "two she bears came out of the wood",
|
||||||
|
query: "she bears",
|
||||||
|
expect: 1,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
text: "bald head and she bears",
|
||||||
|
query: "bald prophet and two she bears",
|
||||||
|
expect: 1, // only "she bears" matches
|
||||||
|
},
|
||||||
|
{
|
||||||
|
text: "no match here",
|
||||||
|
query: "she bears",
|
||||||
|
expect: 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
text: "she bears and bald prophet",
|
||||||
|
query: "bald prophet she bears",
|
||||||
|
expect: 2, // "she bears" and "bald prophet"
|
||||||
|
},
|
||||||
|
}
|
||||||
|
for _, tt := range tests {
|
||||||
|
got := countPhraseMatches(tt.text, tt.query)
|
||||||
|
if got != tt.expect {
|
||||||
|
t.Errorf("countPhraseMatches(%q, %q) = %d, want %d", tt.text, tt.query, got, tt.expect)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestAreSlugsAdjacent(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
slug1 string
|
||||||
|
slug2 string
|
||||||
|
expect bool
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
slug1: "kjv_bible.epub_1786_0",
|
||||||
|
slug2: "kjv_bible.epub_1787_0",
|
||||||
|
expect: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
slug1: "kjv_bible.epub_1787_0",
|
||||||
|
slug2: "kjv_bible.epub_1786_0",
|
||||||
|
expect: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
slug1: "kjv_bible.epub_1786_0",
|
||||||
|
slug2: "kjv_bible.epub_1788_0",
|
||||||
|
expect: false,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
slug1: "otherfile.txt_1_0",
|
||||||
|
slug2: "kjv_bible.epub_1786_0",
|
||||||
|
expect: false,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
slug1: "file_1_0",
|
||||||
|
slug2: "file_1_1",
|
||||||
|
expect: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
slug1: "file_1_0",
|
||||||
|
slug2: "file_2_0", // different batch
|
||||||
|
expect: true, // sequential batches with same chunk index are adjacent
|
||||||
|
},
|
||||||
|
}
|
||||||
|
for _, tt := range tests {
|
||||||
|
got := areSlugsAdjacent(tt.slug1, tt.slug2)
|
||||||
|
if got != tt.expect {
|
||||||
|
t.Errorf("areSlugsAdjacent(%q, %q) = %v, want %v", tt.slug1, tt.slug2, got, tt.expect)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestParseSlugIndices(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
slug string
|
||||||
|
wantBatch int
|
||||||
|
wantChunk int
|
||||||
|
wantOk bool
|
||||||
|
}{
|
||||||
|
{"kjv_bible.epub_1786_0", 1786, 0, true},
|
||||||
|
{"file_1_5", 1, 5, true},
|
||||||
|
{"no_underscore", 0, 0, false},
|
||||||
|
{"file_abc_def", 0, 0, false},
|
||||||
|
{"file_123_456_extra", 456, 0, false}, // regex matches last two numbers
|
||||||
|
}
|
||||||
|
for _, tt := range tests {
|
||||||
|
batch, chunk, ok := parseSlugIndices(tt.slug)
|
||||||
|
if ok != tt.wantOk {
|
||||||
|
t.Errorf("parseSlugIndices(%q) ok = %v, want %v", tt.slug, ok, tt.wantOk)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if ok && (batch != tt.wantBatch || chunk != tt.wantChunk) {
|
||||||
|
t.Errorf("parseSlugIndices(%q) = (%d, %d), want (%d, %d)", tt.slug, batch, chunk, tt.wantBatch, tt.wantChunk)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -340,11 +340,9 @@ func (vs *VectorStorage) scanRows(rows *sql.Rows) ([]models.VectorRow, error) {
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
// Convert BM25 score to distance-like metric (lower is better)
|
// Convert BM25 score to distance-like metric (lower is better)
|
||||||
// BM25 is negative, more negative is better. We'll normalize to positive distance.
|
// BM25 is negative, more negative is better. Keep as negative.
|
||||||
distance := float32(-score) // Make positive (since score is negative)
|
distance := float32(score) // Keep negative, more negative is better
|
||||||
if distance < 0 {
|
// No clamping needed; negative distances are fine
|
||||||
distance = 0
|
|
||||||
}
|
|
||||||
results = append(results, models.VectorRow{
|
results = append(results, models.VectorRow{
|
||||||
Slug: slug,
|
Slug: slug,
|
||||||
RawText: rawText,
|
RawText: rawText,
|
||||||
|
|||||||
28
session.go
28
session.go
@@ -1,6 +1,7 @@
|
|||||||
package main
|
package main
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"bytes"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
@@ -161,10 +162,31 @@ func loadOldChatOrGetNew() []models.RoleMsg {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func copyToClipboard(text string) error {
|
func copyToClipboard(text string) error {
|
||||||
cmd := exec.Command("xclip", "-selection", "clipboard")
|
var cmd *exec.Cmd
|
||||||
cmd.Stdin = nil
|
if _, err := exec.LookPath("xclip"); err == nil {
|
||||||
|
cmd = exec.Command("xclip", "-selection", "clipboard")
|
||||||
|
} else if _, err := exec.LookPath("wl-copy"); err == nil {
|
||||||
|
cmd = exec.Command("wl-copy")
|
||||||
|
} else {
|
||||||
|
return errors.New("no clipboard tool found (install xclip or wl-clipboard)")
|
||||||
|
}
|
||||||
|
cmd.Stdin = strings.NewReader(text)
|
||||||
cmd.Stdout = nil
|
cmd.Stdout = nil
|
||||||
cmd.Stderr = nil
|
cmd.Stderr = nil
|
||||||
cmd.Stdin = strings.NewReader(text)
|
|
||||||
return cmd.Run()
|
return cmd.Run()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func readFromClipboard() (string, error) {
|
||||||
|
var cmd *exec.Cmd
|
||||||
|
if _, err := exec.LookPath("xclip"); err == nil {
|
||||||
|
cmd = exec.Command("xclip", "-selection", "clipboard", "-out")
|
||||||
|
} else if _, err := exec.LookPath("wl-paste"); err == nil {
|
||||||
|
cmd = exec.Command("wl-paste")
|
||||||
|
} else {
|
||||||
|
return "", errors.New("no clipboard tool found (install xclip or wl-clipboard)")
|
||||||
|
}
|
||||||
|
var out bytes.Buffer
|
||||||
|
cmd.Stdout = &out
|
||||||
|
err := cmd.Run()
|
||||||
|
return out.String(), err
|
||||||
|
}
|
||||||
|
|||||||
@@ -6,6 +6,7 @@ type Memories interface {
|
|||||||
Memorise(m *models.Memory) (*models.Memory, error)
|
Memorise(m *models.Memory) (*models.Memory, error)
|
||||||
Recall(agent, topic string) (string, error)
|
Recall(agent, topic string) (string, error)
|
||||||
RecallTopics(agent string) ([]string, error)
|
RecallTopics(agent string) ([]string, error)
|
||||||
|
Forget(agent, topic string) error
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p ProviderSQL) Memorise(m *models.Memory) (*models.Memory, error) {
|
func (p ProviderSQL) Memorise(m *models.Memory) (*models.Memory, error) {
|
||||||
@@ -52,3 +53,13 @@ func (p ProviderSQL) RecallTopics(agent string) ([]string, error) {
|
|||||||
}
|
}
|
||||||
return topics, nil
|
return topics, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (p ProviderSQL) Forget(agent, topic string) error {
|
||||||
|
query := "DELETE FROM memories WHERE agent = $1 AND topic = $2"
|
||||||
|
_, err := p.db.Exec(query, agent, topic)
|
||||||
|
if err != nil {
|
||||||
|
p.logger.Error("failed to delete memory", "query", query, "error", err)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|||||||
@@ -23,9 +23,20 @@ func (p *ProviderSQL) Migrate() error {
|
|||||||
p.logger.Error("Failed to read migrations directory;", "error", err)
|
p.logger.Error("Failed to read migrations directory;", "error", err)
|
||||||
return fmt.Errorf("failed to read migrations directory: %w", err)
|
return fmt.Errorf("failed to read migrations directory: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Check if FTS already has data - skip populate migration if so
|
||||||
|
var ftsCount int
|
||||||
|
_ = p.db.QueryRow("SELECT COUNT(*) FROM fts_embeddings").Scan(&ftsCount)
|
||||||
|
skipFTSMigration := ftsCount > 0
|
||||||
|
|
||||||
// Execute each .up.sql file
|
// Execute each .up.sql file
|
||||||
for _, file := range files {
|
for _, file := range files {
|
||||||
if strings.HasSuffix(file.Name(), ".up.sql") {
|
if strings.HasSuffix(file.Name(), ".up.sql") {
|
||||||
|
// Skip FTS populate migration if already populated
|
||||||
|
if skipFTSMigration && strings.Contains(file.Name(), "004_populate_fts") {
|
||||||
|
p.logger.Debug("Skipping FTS migration - already populated", "file", file.Name())
|
||||||
|
continue
|
||||||
|
}
|
||||||
err := p.executeMigration(migrationsDir, file.Name())
|
err := p.executeMigration(migrationsDir, file.Name())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
p.logger.Error("Failed to execute migration %s: %v", file.Name(), err)
|
p.logger.Error("Failed to execute migration %s: %v", file.Name(), err)
|
||||||
|
|||||||
@@ -1,26 +1,4 @@
|
|||||||
-- Populate FTS table with existing embeddings
|
-- Populate FTS table with existing embeddings (incremental - only inserts missing rows)
|
||||||
DELETE FROM fts_embeddings;
|
-- Only use 768 embeddings as that's what we use
|
||||||
|
INSERT OR IGNORE INTO fts_embeddings (slug, raw_text, filename, embedding_size)
|
||||||
INSERT INTO fts_embeddings (slug, raw_text, filename, embedding_size)
|
|
||||||
SELECT slug, raw_text, filename, 384 FROM embeddings_384;
|
|
||||||
|
|
||||||
INSERT INTO fts_embeddings (slug, raw_text, filename, embedding_size)
|
|
||||||
SELECT slug, raw_text, filename, 768 FROM embeddings_768;
|
SELECT slug, raw_text, filename, 768 FROM embeddings_768;
|
||||||
|
|
||||||
INSERT INTO fts_embeddings (slug, raw_text, filename, embedding_size)
|
|
||||||
SELECT slug, raw_text, filename, 1024 FROM embeddings_1024;
|
|
||||||
|
|
||||||
INSERT INTO fts_embeddings (slug, raw_text, filename, embedding_size)
|
|
||||||
SELECT slug, raw_text, filename, 1536 FROM embeddings_1536;
|
|
||||||
|
|
||||||
INSERT INTO fts_embeddings (slug, raw_text, filename, embedding_size)
|
|
||||||
SELECT slug, raw_text, filename, 2048 FROM embeddings_2048;
|
|
||||||
|
|
||||||
INSERT INTO fts_embeddings (slug, raw_text, filename, embedding_size)
|
|
||||||
SELECT slug, raw_text, filename, 3072 FROM embeddings_3072;
|
|
||||||
|
|
||||||
INSERT INTO fts_embeddings (slug, raw_text, filename, embedding_size)
|
|
||||||
SELECT slug, raw_text, filename, 4096 FROM embeddings_4096;
|
|
||||||
|
|
||||||
INSERT INTO fts_embeddings (slug, raw_text, filename, embedding_size)
|
|
||||||
SELECT slug, raw_text, filename, 5120 FROM embeddings_5120;
|
|
||||||
87
storage/migrations/005_drop_unused_embeddings.down.sql
Normal file
87
storage/migrations/005_drop_unused_embeddings.down.sql
Normal file
@@ -0,0 +1,87 @@
|
|||||||
|
-- Recreate unused embedding tables (for rollback)
|
||||||
|
CREATE TABLE IF NOT EXISTS embeddings_384 (
|
||||||
|
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||||
|
embeddings BLOB NOT NULL,
|
||||||
|
slug TEXT NOT NULL,
|
||||||
|
raw_text TEXT NOT NULL,
|
||||||
|
filename TEXT NOT NULL,
|
||||||
|
created_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS embeddings_1024 (
|
||||||
|
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||||
|
embeddings BLOB NOT NULL,
|
||||||
|
slug TEXT NOT NULL,
|
||||||
|
raw_text TEXT NOT NULL,
|
||||||
|
filename TEXT NOT NULL,
|
||||||
|
created_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS embeddings_1536 (
|
||||||
|
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||||
|
embeddings BLOB NOT NULL,
|
||||||
|
slug TEXT NOT NULL,
|
||||||
|
raw_text TEXT NOT NULL,
|
||||||
|
filename TEXT NOT NULL,
|
||||||
|
created_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS embeddings_2048 (
|
||||||
|
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||||
|
embeddings BLOB NOT NULL,
|
||||||
|
slug TEXT NOT NULL,
|
||||||
|
raw_text TEXT NOT NULL,
|
||||||
|
filename TEXT NOT NULL,
|
||||||
|
created_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS embeddings_3072 (
|
||||||
|
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||||
|
embeddings BLOB NOT NULL,
|
||||||
|
slug TEXT NOT NULL,
|
||||||
|
raw_text TEXT NOT NULL,
|
||||||
|
filename TEXT NOT NULL,
|
||||||
|
created_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS embeddings_4096 (
|
||||||
|
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||||
|
embeddings BLOB NOT NULL,
|
||||||
|
slug TEXT NOT NULL,
|
||||||
|
raw_text TEXT NOT NULL,
|
||||||
|
filename TEXT NOT NULL,
|
||||||
|
created_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE TABLE IF NOT EXISTS embeddings_5120 (
|
||||||
|
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||||
|
embeddings BLOB NOT NULL,
|
||||||
|
slug TEXT NOT NULL,
|
||||||
|
raw_text TEXT NOT NULL,
|
||||||
|
filename TEXT NOT NULL,
|
||||||
|
created_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP
|
||||||
|
);
|
||||||
|
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_384_filename ON embeddings_384(filename);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_1024_filename ON embeddings_1024(filename);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_1536_filename ON embeddings_1536(filename);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_2048_filename ON embeddings_2048(filename);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_3072_filename ON embeddings_3072(filename);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_4096_filename ON embeddings_4096(filename);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_5120_filename ON embeddings_5120(filename);
|
||||||
|
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_384_slug ON embeddings_384(slug);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_1024_slug ON embeddings_1024(slug);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_1536_slug ON embeddings_1536(slug);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_2048_slug ON embeddings_2048(slug);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_3072_slug ON embeddings_3072(slug);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_4096_slug ON embeddings_4096(slug);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_5120_slug ON embeddings_5120(slug);
|
||||||
|
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_384_created_at ON embeddings_384(created_at);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_1024_created_at ON embeddings_1024(created_at);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_1536_created_at ON embeddings_1536(created_at);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_2048_created_at ON embeddings_2048(created_at);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_3072_created_at ON embeddings_3072(created_at);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_4096_created_at ON embeddings_4096(created_at);
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_embeddings_5120_created_at ON embeddings_5120(created_at);
|
||||||
32
storage/migrations/005_drop_unused_embeddings.up.sql
Normal file
32
storage/migrations/005_drop_unused_embeddings.up.sql
Normal file
@@ -0,0 +1,32 @@
|
|||||||
|
-- Drop unused embedding tables (we only use 768)
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_384_filename;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_1024_filename;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_1536_filename;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_2048_filename;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_3072_filename;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_4096_filename;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_5120_filename;
|
||||||
|
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_384_slug;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_1024_slug;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_1536_slug;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_2048_slug;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_3072_slug;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_4096_slug;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_5120_slug;
|
||||||
|
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_384_created_at;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_1024_created_at;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_1536_created_at;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_2048_created_at;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_3072_created_at;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_4096_created_at;
|
||||||
|
DROP INDEX IF EXISTS idx_embeddings_5120_created_at;
|
||||||
|
|
||||||
|
DROP TABLE IF EXISTS embeddings_384;
|
||||||
|
DROP TABLE IF EXISTS embeddings_1024;
|
||||||
|
DROP TABLE IF EXISTS embeddings_1536;
|
||||||
|
DROP TABLE IF EXISTS embeddings_2048;
|
||||||
|
DROP TABLE IF EXISTS embeddings_3072;
|
||||||
|
DROP TABLE IF EXISTS embeddings_4096;
|
||||||
|
DROP TABLE IF EXISTS embeddings_5120;
|
||||||
@@ -48,22 +48,8 @@ func mathBitsToFloat32(b uint32) float32 {
|
|||||||
|
|
||||||
func fetchTableName(emb []float32) (string, error) {
|
func fetchTableName(emb []float32) (string, error) {
|
||||||
switch len(emb) {
|
switch len(emb) {
|
||||||
case 384:
|
|
||||||
return "embeddings_384", nil
|
|
||||||
case 768:
|
case 768:
|
||||||
return "embeddings_768", nil
|
return "embeddings_768", nil
|
||||||
case 1024:
|
|
||||||
return "embeddings_1024", nil
|
|
||||||
case 1536:
|
|
||||||
return "embeddings_1536", nil
|
|
||||||
case 2048:
|
|
||||||
return "embeddings_2048", nil
|
|
||||||
case 3072:
|
|
||||||
return "embeddings_3072", nil
|
|
||||||
case 4096:
|
|
||||||
return "embeddings_4096", nil
|
|
||||||
case 5120:
|
|
||||||
return "embeddings_5120", nil
|
|
||||||
default:
|
default:
|
||||||
return "", fmt.Errorf("no table for the size of %d", len(emb))
|
return "", fmt.Errorf("no table for the size of %d", len(emb))
|
||||||
}
|
}
|
||||||
@@ -170,62 +156,25 @@ func sqrt(f float32) float32 {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (p ProviderSQL) ListFiles() ([]string, error) {
|
func (p ProviderSQL) ListFiles() ([]string, error) {
|
||||||
fileLists := make([][]string, 0)
|
query := "SELECT DISTINCT filename FROM embeddings_768"
|
||||||
|
|
||||||
// Query all supported tables and combine results
|
|
||||||
tableNames := []string{
|
|
||||||
"embeddings_384", "embeddings_768", "embeddings_1024", "embeddings_1536",
|
|
||||||
"embeddings_2048", "embeddings_3072", "embeddings_4096", "embeddings_5120",
|
|
||||||
}
|
|
||||||
for _, table := range tableNames {
|
|
||||||
query := "SELECT DISTINCT filename FROM " + table
|
|
||||||
rows, err := p.db.Query(query)
|
rows, err := p.db.Query(query)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
// Continue if one table doesn't exist
|
return nil, err
|
||||||
continue
|
|
||||||
}
|
}
|
||||||
|
defer rows.Close()
|
||||||
var files []string
|
var allFiles []string
|
||||||
for rows.Next() {
|
for rows.Next() {
|
||||||
var filename string
|
var filename string
|
||||||
if err := rows.Scan(&filename); err != nil {
|
if err := rows.Scan(&filename); err != nil {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
files = append(files, filename)
|
allFiles = append(allFiles, filename)
|
||||||
}
|
|
||||||
rows.Close()
|
|
||||||
|
|
||||||
fileLists = append(fileLists, files)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Combine and deduplicate
|
|
||||||
fileSet := make(map[string]bool)
|
|
||||||
var allFiles []string
|
|
||||||
for _, files := range fileLists {
|
|
||||||
for _, file := range files {
|
|
||||||
if !fileSet[file] {
|
|
||||||
fileSet[file] = true
|
|
||||||
allFiles = append(allFiles, file)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
return allFiles, nil
|
return allFiles, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p ProviderSQL) RemoveEmbByFileName(filename string) error {
|
func (p ProviderSQL) RemoveEmbByFileName(filename string) error {
|
||||||
var errors []string
|
query := "DELETE FROM embeddings_768 WHERE filename = ?"
|
||||||
tableNames := []string{
|
_, err := p.db.Exec(query, filename)
|
||||||
"embeddings_384", "embeddings_768", "embeddings_1024", "embeddings_1536",
|
return err
|
||||||
"embeddings_2048", "embeddings_3072", "embeddings_4096", "embeddings_5120",
|
|
||||||
}
|
|
||||||
for _, table := range tableNames {
|
|
||||||
query := fmt.Sprintf("DELETE FROM %s WHERE filename = ?", table)
|
|
||||||
if _, err := p.db.Exec(query, filename); err != nil {
|
|
||||||
errors = append(errors, err.Error())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if len(errors) > 0 {
|
|
||||||
return fmt.Errorf("errors occurred: %v", errors)
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
}
|
||||||
|
|||||||
38
tables.go
38
tables.go
@@ -2,6 +2,7 @@ package main
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"gf-lt/tools"
|
||||||
"image"
|
"image"
|
||||||
"os"
|
"os"
|
||||||
"path"
|
"path"
|
||||||
@@ -171,7 +172,7 @@ func makeChatTable(chatMap map[string]models.Chat) *tview.Table {
|
|||||||
return
|
return
|
||||||
case "move sysprompt onto 1st msg":
|
case "move sysprompt onto 1st msg":
|
||||||
chatBody.Messages[1].Content = chatBody.Messages[0].Content + chatBody.Messages[1].Content
|
chatBody.Messages[1].Content = chatBody.Messages[0].Content + chatBody.Messages[1].Content
|
||||||
chatBody.Messages[0].Content = rpDefenitionSysMsg
|
chatBody.Messages[0].Content = tools.RpDefenitionSysMsg
|
||||||
textView.SetText(chatToText(chatBody.Messages, cfg.ShowSys))
|
textView.SetText(chatToText(chatBody.Messages, cfg.ShowSys))
|
||||||
activeChatName = selectedChat
|
activeChatName = selectedChat
|
||||||
pages.RemovePage(historyPage)
|
pages.RemovePage(historyPage)
|
||||||
@@ -243,11 +244,9 @@ func makeRAGTable(fileList []string, loadedFiles []string) *tview.Flex {
|
|||||||
for _, f := range loadedFiles {
|
for _, f := range loadedFiles {
|
||||||
loadedSet[f] = true
|
loadedSet[f] = true
|
||||||
}
|
}
|
||||||
|
|
||||||
// Build merged list: files from ragdir + orphaned files from DB
|
// Build merged list: files from ragdir + orphaned files from DB
|
||||||
ragFiles := make([]ragFileInfo, 0, len(fileList)+len(loadedFiles))
|
ragFiles := make([]ragFileInfo, 0, len(fileList)+len(loadedFiles))
|
||||||
seen := make(map[string]bool)
|
seen := make(map[string]bool)
|
||||||
|
|
||||||
// Add files from ragdir
|
// Add files from ragdir
|
||||||
for _, f := range fileList {
|
for _, f := range fileList {
|
||||||
ragFiles = append(ragFiles, ragFileInfo{
|
ragFiles = append(ragFiles, ragFileInfo{
|
||||||
@@ -258,7 +257,6 @@ func makeRAGTable(fileList []string, loadedFiles []string) *tview.Flex {
|
|||||||
})
|
})
|
||||||
seen[f] = true
|
seen[f] = true
|
||||||
}
|
}
|
||||||
|
|
||||||
// Add orphaned files (in DB but not in ragdir)
|
// Add orphaned files (in DB but not in ragdir)
|
||||||
for _, f := range loadedFiles {
|
for _, f := range loadedFiles {
|
||||||
if !seen[f] {
|
if !seen[f] {
|
||||||
@@ -275,7 +273,7 @@ func makeRAGTable(fileList []string, loadedFiles []string) *tview.Flex {
|
|||||||
fileTable := tview.NewTable().
|
fileTable := tview.NewTable().
|
||||||
SetBorders(true)
|
SetBorders(true)
|
||||||
longStatusView := tview.NewTextView()
|
longStatusView := tview.NewTextView()
|
||||||
longStatusView.SetText("status text")
|
longStatusView.SetText("press x to exit")
|
||||||
longStatusView.SetBorder(true).SetTitle("status")
|
longStatusView.SetBorder(true).SetTitle("status")
|
||||||
longStatusView.SetChangedFunc(func() {
|
longStatusView.SetChangedFunc(func() {
|
||||||
app.Draw()
|
app.Draw()
|
||||||
@@ -376,7 +374,6 @@ func makeRAGTable(fileList []string, loadedFiles []string) *tview.Flex {
|
|||||||
}
|
}
|
||||||
errCh := make(chan error, 1) // why?
|
errCh := make(chan error, 1) // why?
|
||||||
go func() {
|
go func() {
|
||||||
defer pages.RemovePage(RAGPage)
|
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case err := <-errCh:
|
case err := <-errCh:
|
||||||
@@ -417,7 +414,6 @@ func makeRAGTable(fileList []string, loadedFiles []string) *tview.Flex {
|
|||||||
}
|
}
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
// defer pages.RemovePage(RAGPage)
|
|
||||||
tc := fileTable.GetCell(row, column)
|
tc := fileTable.GetCell(row, column)
|
||||||
tc.SetTextColor(tcell.ColorRed)
|
tc.SetTextColor(tcell.ColorRed)
|
||||||
fileTable.SetSelectable(false, false)
|
fileTable.SetSelectable(false, false)
|
||||||
@@ -430,7 +426,6 @@ func makeRAGTable(fileList []string, loadedFiles []string) *tview.Flex {
|
|||||||
f := ragFiles[row-1]
|
f := ragFiles[row-1]
|
||||||
// Handle "-" case (orphaned file with no delete option)
|
// Handle "-" case (orphaned file with no delete option)
|
||||||
if tc.Text == "-" {
|
if tc.Text == "-" {
|
||||||
pages.RemovePage(RAGPage)
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
switch tc.Text {
|
switch tc.Text {
|
||||||
@@ -441,14 +436,14 @@ func makeRAGTable(fileList []string, loadedFiles []string) *tview.Flex {
|
|||||||
if err := ragger.LoadRAG(fpath); err != nil {
|
if err := ragger.LoadRAG(fpath); err != nil {
|
||||||
logger.Error("failed to embed file", "chat", fpath, "error", err)
|
logger.Error("failed to embed file", "chat", fpath, "error", err)
|
||||||
showToast("RAG", "failed to embed file; error: "+err.Error())
|
showToast("RAG", "failed to embed file; error: "+err.Error())
|
||||||
app.QueueUpdate(func() {
|
|
||||||
pages.RemovePage(RAGPage)
|
|
||||||
})
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
showToast("RAG", "file loaded successfully")
|
showToast("RAG", "file loaded successfully")
|
||||||
app.QueueUpdate(func() {
|
app.QueueUpdate(func() {
|
||||||
pages.RemovePage(RAGPage)
|
pages.RemovePage(RAGPage)
|
||||||
|
loadedFiles, _ := ragger.ListLoaded()
|
||||||
|
chatRAGTable := makeRAGTable(fileList, loadedFiles)
|
||||||
|
pages.AddPage(RAGPage, chatRAGTable, true, true)
|
||||||
})
|
})
|
||||||
}()
|
}()
|
||||||
return
|
return
|
||||||
@@ -458,14 +453,14 @@ func makeRAGTable(fileList []string, loadedFiles []string) *tview.Flex {
|
|||||||
if err := ragger.RemoveFile(f.name); err != nil {
|
if err := ragger.RemoveFile(f.name); err != nil {
|
||||||
logger.Error("failed to unload file from RAG", "filename", f.name, "error", err)
|
logger.Error("failed to unload file from RAG", "filename", f.name, "error", err)
|
||||||
showToast("RAG", "failed to unload file; error: "+err.Error())
|
showToast("RAG", "failed to unload file; error: "+err.Error())
|
||||||
app.QueueUpdate(func() {
|
|
||||||
pages.RemovePage(RAGPage)
|
|
||||||
})
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
showToast("RAG", "file unloaded successfully")
|
showToast("RAG", "file unloaded successfully")
|
||||||
app.QueueUpdate(func() {
|
app.QueueUpdate(func() {
|
||||||
pages.RemovePage(RAGPage)
|
pages.RemovePage(RAGPage)
|
||||||
|
loadedFiles, _ := ragger.ListLoaded()
|
||||||
|
chatRAGTable := makeRAGTable(fileList, loadedFiles)
|
||||||
|
pages.AddPage(RAGPage, chatRAGTable, true, true)
|
||||||
})
|
})
|
||||||
}()
|
}()
|
||||||
return
|
return
|
||||||
@@ -476,6 +471,21 @@ func makeRAGTable(fileList []string, loadedFiles []string) *tview.Flex {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
showToast("chat deleted", fpath+" was deleted")
|
showToast("chat deleted", fpath+" was deleted")
|
||||||
|
go func() {
|
||||||
|
app.QueueUpdate(func() {
|
||||||
|
pages.RemovePage(RAGPage)
|
||||||
|
newFileList, _ := os.ReadDir(cfg.RAGDir)
|
||||||
|
loadedFiles, _ := ragger.ListLoaded()
|
||||||
|
var newFiles []string
|
||||||
|
for _, f := range newFileList {
|
||||||
|
if !f.IsDir() {
|
||||||
|
newFiles = append(newFiles, f.Name())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
chatRAGTable := makeRAGTable(newFiles, loadedFiles)
|
||||||
|
pages.AddPage(RAGPage, chatRAGTable, true, true)
|
||||||
|
})
|
||||||
|
}()
|
||||||
return
|
return
|
||||||
default:
|
default:
|
||||||
pages.RemovePage(RAGPage)
|
pages.RemovePage(RAGPage)
|
||||||
|
|||||||
416
tools/chain.go
Normal file
416
tools/chain.go
Normal file
@@ -0,0 +1,416 @@
|
|||||||
|
package tools
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"os"
|
||||||
|
"os/exec"
|
||||||
|
"path/filepath"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
// Operator represents a chain operator between commands.
|
||||||
|
type Operator int
|
||||||
|
|
||||||
|
const (
|
||||||
|
OpNone Operator = iota
|
||||||
|
OpAnd // &&
|
||||||
|
OpOr // ||
|
||||||
|
OpSeq // ;
|
||||||
|
OpPipe // |
|
||||||
|
)
|
||||||
|
|
||||||
|
// Segment is a single command in a chain.
|
||||||
|
type Segment struct {
|
||||||
|
Raw string
|
||||||
|
Op Operator // operator AFTER this segment
|
||||||
|
}
|
||||||
|
|
||||||
|
// ParseChain splits a command string into segments by &&, ;, and |.
|
||||||
|
// Respects quoted strings (single and double quotes).
|
||||||
|
func ParseChain(input string) []Segment {
|
||||||
|
var segments []Segment
|
||||||
|
var current strings.Builder
|
||||||
|
runes := []rune(input)
|
||||||
|
n := len(runes)
|
||||||
|
for i := 0; i < n; i++ {
|
||||||
|
ch := runes[i]
|
||||||
|
// handle quotes
|
||||||
|
if ch == '\'' || ch == '"' {
|
||||||
|
quote := ch
|
||||||
|
current.WriteRune(ch)
|
||||||
|
i++
|
||||||
|
for i < n && runes[i] != quote {
|
||||||
|
current.WriteRune(runes[i])
|
||||||
|
i++
|
||||||
|
}
|
||||||
|
if i < n {
|
||||||
|
current.WriteRune(runes[i])
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// &&
|
||||||
|
if ch == '&' && i+1 < n && runes[i+1] == '&' {
|
||||||
|
segments = append(segments, Segment{
|
||||||
|
Raw: strings.TrimSpace(current.String()),
|
||||||
|
Op: OpAnd,
|
||||||
|
})
|
||||||
|
current.Reset()
|
||||||
|
i++ // skip second &
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// ;
|
||||||
|
if ch == ';' {
|
||||||
|
segments = append(segments, Segment{
|
||||||
|
Raw: strings.TrimSpace(current.String()),
|
||||||
|
Op: OpSeq,
|
||||||
|
})
|
||||||
|
current.Reset()
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// ||
|
||||||
|
if ch == '|' && i+1 < n && runes[i+1] == '|' {
|
||||||
|
segments = append(segments, Segment{
|
||||||
|
Raw: strings.TrimSpace(current.String()),
|
||||||
|
Op: OpOr,
|
||||||
|
})
|
||||||
|
current.Reset()
|
||||||
|
i++ // skip second |
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// | (single pipe)
|
||||||
|
if ch == '|' {
|
||||||
|
segments = append(segments, Segment{
|
||||||
|
Raw: strings.TrimSpace(current.String()),
|
||||||
|
Op: OpPipe,
|
||||||
|
})
|
||||||
|
current.Reset()
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
current.WriteRune(ch)
|
||||||
|
}
|
||||||
|
// last segment
|
||||||
|
last := strings.TrimSpace(current.String())
|
||||||
|
if last != "" {
|
||||||
|
segments = append(segments, Segment{Raw: last, Op: OpNone})
|
||||||
|
}
|
||||||
|
return segments
|
||||||
|
}
|
||||||
|
|
||||||
|
// ExecChain executes a command string with pipe/chaining support.
|
||||||
|
// Returns the combined output of all commands.
|
||||||
|
func ExecChain(command string) string {
|
||||||
|
segments := ParseChain(command)
|
||||||
|
if len(segments) == 0 {
|
||||||
|
return "[error] empty command"
|
||||||
|
}
|
||||||
|
var collected []string
|
||||||
|
var lastOutput string
|
||||||
|
var lastErr error
|
||||||
|
pipeInput := ""
|
||||||
|
for i, seg := range segments {
|
||||||
|
if i > 0 {
|
||||||
|
prevOp := segments[i-1].Op
|
||||||
|
// && semantics: skip if previous failed
|
||||||
|
if prevOp == OpAnd && lastErr != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// || semantics: skip if previous succeeded
|
||||||
|
if prevOp == OpOr && lastErr == nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// determine stdin for this segment
|
||||||
|
segStdin := ""
|
||||||
|
if i == 0 {
|
||||||
|
segStdin = pipeInput
|
||||||
|
} else if segments[i-1].Op == OpPipe {
|
||||||
|
segStdin = lastOutput
|
||||||
|
}
|
||||||
|
lastOutput, lastErr = execSingle(seg.Raw, segStdin)
|
||||||
|
// pipe: output flows to next command's stdin
|
||||||
|
// && or ;: collect output
|
||||||
|
if i < len(segments)-1 && seg.Op == OpPipe {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if lastOutput != "" {
|
||||||
|
collected = append(collected, lastOutput)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return strings.Join(collected, "\n")
|
||||||
|
}
|
||||||
|
|
||||||
|
// execSingle executes a single command (with arguments) and returns output and error.
|
||||||
|
func execSingle(command, stdin string) (string, error) {
|
||||||
|
parts := tokenize(command)
|
||||||
|
if len(parts) == 0 {
|
||||||
|
return "", errors.New("empty command")
|
||||||
|
}
|
||||||
|
name := parts[0]
|
||||||
|
args := parts[1:]
|
||||||
|
// Check if it's a built-in Go command
|
||||||
|
if result, isBuiltin := execBuiltin(name, args, stdin); isBuiltin {
|
||||||
|
return result, nil
|
||||||
|
}
|
||||||
|
// Otherwise execute as system command
|
||||||
|
cmd := exec.Command(name, args...)
|
||||||
|
if stdin != "" {
|
||||||
|
cmd.Stdin = strings.NewReader(stdin)
|
||||||
|
}
|
||||||
|
output, err := cmd.CombinedOutput()
|
||||||
|
if err != nil {
|
||||||
|
return string(output), err
|
||||||
|
}
|
||||||
|
return string(output), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// tokenize splits a command string by whitespace, respecting quotes.
|
||||||
|
func tokenize(input string) []string {
|
||||||
|
var tokens []string
|
||||||
|
var current strings.Builder
|
||||||
|
inQuote := false
|
||||||
|
var quoteChar rune
|
||||||
|
for _, ch := range input {
|
||||||
|
if inQuote {
|
||||||
|
if ch == quoteChar {
|
||||||
|
inQuote = false
|
||||||
|
} else {
|
||||||
|
current.WriteRune(ch)
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if ch == '\'' || ch == '"' {
|
||||||
|
inQuote = true
|
||||||
|
quoteChar = ch
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if ch == ' ' || ch == '\t' {
|
||||||
|
if current.Len() > 0 {
|
||||||
|
tokens = append(tokens, current.String())
|
||||||
|
current.Reset()
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
current.WriteRune(ch)
|
||||||
|
}
|
||||||
|
if current.Len() > 0 {
|
||||||
|
tokens = append(tokens, current.String())
|
||||||
|
}
|
||||||
|
return tokens
|
||||||
|
}
|
||||||
|
|
||||||
|
// execBuiltin executes a built-in command if it exists.
|
||||||
|
// Returns (result, true) if it was a built-in (even if result is empty).
|
||||||
|
// Returns ("", false) if it's not a built-in command.
|
||||||
|
func execBuiltin(name string, args []string, stdin string) (string, bool) {
|
||||||
|
switch name {
|
||||||
|
case "echo":
|
||||||
|
if stdin != "" {
|
||||||
|
return stdin, true
|
||||||
|
}
|
||||||
|
return strings.Join(args, " "), true
|
||||||
|
case "time":
|
||||||
|
return "2006-01-02 15:04:05 MST", true
|
||||||
|
case "cat":
|
||||||
|
if len(args) == 0 {
|
||||||
|
if stdin != "" {
|
||||||
|
return stdin, true
|
||||||
|
}
|
||||||
|
return "", true
|
||||||
|
}
|
||||||
|
path := args[0]
|
||||||
|
abs := path
|
||||||
|
if !filepath.IsAbs(path) {
|
||||||
|
abs = filepath.Join(cfg.FilePickerDir, path)
|
||||||
|
}
|
||||||
|
data, err := os.ReadFile(abs)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] cat: %v", err), true
|
||||||
|
}
|
||||||
|
return string(data), true
|
||||||
|
case "pwd":
|
||||||
|
return cfg.FilePickerDir, true
|
||||||
|
case "cd":
|
||||||
|
if len(args) == 0 {
|
||||||
|
return "[error] usage: cd <dir>", true
|
||||||
|
}
|
||||||
|
dir := args[0]
|
||||||
|
// Resolve relative to cfg.FilePickerDir
|
||||||
|
abs := dir
|
||||||
|
if !filepath.IsAbs(dir) {
|
||||||
|
abs = filepath.Join(cfg.FilePickerDir, dir)
|
||||||
|
}
|
||||||
|
abs = filepath.Clean(abs)
|
||||||
|
info, err := os.Stat(abs)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] cd: %v", err), true
|
||||||
|
}
|
||||||
|
if !info.IsDir() {
|
||||||
|
return "[error] cd: not a directory: " + dir, true
|
||||||
|
}
|
||||||
|
cfg.FilePickerDir = abs
|
||||||
|
return "Changed directory to: " + cfg.FilePickerDir, true
|
||||||
|
case "mkdir":
|
||||||
|
if len(args) == 0 {
|
||||||
|
return "[error] usage: mkdir [-p] <dir>", true
|
||||||
|
}
|
||||||
|
createParents := false
|
||||||
|
var dirPath string
|
||||||
|
for _, a := range args {
|
||||||
|
if a == "-p" || a == "--parents" {
|
||||||
|
createParents = true
|
||||||
|
} else if dirPath == "" {
|
||||||
|
dirPath = a
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if dirPath == "" {
|
||||||
|
return "[error] usage: mkdir [-p] <dir>", true
|
||||||
|
}
|
||||||
|
abs := dirPath
|
||||||
|
if !filepath.IsAbs(dirPath) {
|
||||||
|
abs = filepath.Join(cfg.FilePickerDir, dirPath)
|
||||||
|
}
|
||||||
|
abs = filepath.Clean(abs)
|
||||||
|
var mkdirFunc func(string, os.FileMode) error
|
||||||
|
if createParents {
|
||||||
|
mkdirFunc = os.MkdirAll
|
||||||
|
} else {
|
||||||
|
mkdirFunc = os.Mkdir
|
||||||
|
}
|
||||||
|
if err := mkdirFunc(abs, 0o755); err != nil {
|
||||||
|
return fmt.Sprintf("[error] mkdir: %v", err), true
|
||||||
|
}
|
||||||
|
if createParents {
|
||||||
|
return "Created " + dirPath + " (with parents)", true
|
||||||
|
}
|
||||||
|
return "Created " + dirPath, true
|
||||||
|
case "ls":
|
||||||
|
dir := "."
|
||||||
|
for _, a := range args {
|
||||||
|
if !strings.HasPrefix(a, "-") {
|
||||||
|
dir = a
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
abs := dir
|
||||||
|
if !filepath.IsAbs(dir) {
|
||||||
|
abs = filepath.Join(cfg.FilePickerDir, dir)
|
||||||
|
}
|
||||||
|
entries, err := os.ReadDir(abs)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] ls: %v", err), true
|
||||||
|
}
|
||||||
|
var out strings.Builder
|
||||||
|
for _, e := range entries {
|
||||||
|
info, _ := e.Info()
|
||||||
|
switch {
|
||||||
|
case e.IsDir():
|
||||||
|
fmt.Fprintf(&out, "d %-8s %s/\n", "-", e.Name())
|
||||||
|
case info != nil:
|
||||||
|
size := info.Size()
|
||||||
|
sizeStr := strconv.FormatInt(size, 10)
|
||||||
|
if size > 1024 {
|
||||||
|
sizeStr = fmt.Sprintf("%.1fKB", float64(size)/1024)
|
||||||
|
}
|
||||||
|
fmt.Fprintf(&out, "f %-8s %s\n", sizeStr, e.Name())
|
||||||
|
default:
|
||||||
|
fmt.Fprintf(&out, "f %-8s %s\n", "?", e.Name())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if out.Len() == 0 {
|
||||||
|
return "(empty directory)", true
|
||||||
|
}
|
||||||
|
return strings.TrimRight(out.String(), "\n"), true
|
||||||
|
case "go":
|
||||||
|
// Allow all go subcommands
|
||||||
|
if len(args) == 0 {
|
||||||
|
return "[error] usage: go <subcommand> [options]", true
|
||||||
|
}
|
||||||
|
cmd := exec.Command("go", args...)
|
||||||
|
cmd.Dir = cfg.FilePickerDir
|
||||||
|
output, err := cmd.CombinedOutput()
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] go %s: %v\n%s", args[0], err, string(output)), true
|
||||||
|
}
|
||||||
|
return string(output), true
|
||||||
|
case "cp":
|
||||||
|
if len(args) < 2 {
|
||||||
|
return "[error] usage: cp <source> <dest>", true
|
||||||
|
}
|
||||||
|
src := args[0]
|
||||||
|
dst := args[1]
|
||||||
|
if !filepath.IsAbs(src) {
|
||||||
|
src = filepath.Join(cfg.FilePickerDir, src)
|
||||||
|
}
|
||||||
|
if !filepath.IsAbs(dst) {
|
||||||
|
dst = filepath.Join(cfg.FilePickerDir, dst)
|
||||||
|
}
|
||||||
|
data, err := os.ReadFile(src)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] cp: %v", err), true
|
||||||
|
}
|
||||||
|
err = os.WriteFile(dst, data, 0644)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] cp: %v", err), true
|
||||||
|
}
|
||||||
|
return "Copied " + src + " to " + dst, true
|
||||||
|
case "mv":
|
||||||
|
if len(args) < 2 {
|
||||||
|
return "[error] usage: mv <source> <dest>", true
|
||||||
|
}
|
||||||
|
src := args[0]
|
||||||
|
dst := args[1]
|
||||||
|
if !filepath.IsAbs(src) {
|
||||||
|
src = filepath.Join(cfg.FilePickerDir, src)
|
||||||
|
}
|
||||||
|
if !filepath.IsAbs(dst) {
|
||||||
|
dst = filepath.Join(cfg.FilePickerDir, dst)
|
||||||
|
}
|
||||||
|
err := os.Rename(src, dst)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] mv: %v", err), true
|
||||||
|
}
|
||||||
|
return "Moved " + src + " to " + dst, true
|
||||||
|
case "rm":
|
||||||
|
if len(args) == 0 {
|
||||||
|
return "[error] usage: rm [-r] <file>", true
|
||||||
|
}
|
||||||
|
recursive := false
|
||||||
|
var target string
|
||||||
|
for _, a := range args {
|
||||||
|
if a == "-r" || a == "-rf" || a == "-fr" || a == "-recursive" {
|
||||||
|
recursive = true
|
||||||
|
} else if target == "" {
|
||||||
|
target = a
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if target == "" {
|
||||||
|
return "[error] usage: rm [-r] <file>", true
|
||||||
|
}
|
||||||
|
abs := target
|
||||||
|
if !filepath.IsAbs(target) {
|
||||||
|
abs = filepath.Join(cfg.FilePickerDir, target)
|
||||||
|
}
|
||||||
|
info, err := os.Stat(abs)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] rm: %v", err), true
|
||||||
|
}
|
||||||
|
if info.IsDir() {
|
||||||
|
if recursive {
|
||||||
|
err = os.RemoveAll(abs)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] rm: %v", err), true
|
||||||
|
}
|
||||||
|
return "Removed " + abs, true
|
||||||
|
}
|
||||||
|
return "[error] rm: is a directory (use -r)", true
|
||||||
|
}
|
||||||
|
err = os.Remove(abs)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] rm: %v", err), true
|
||||||
|
}
|
||||||
|
return "Removed " + abs, true
|
||||||
|
}
|
||||||
|
return "", false
|
||||||
|
}
|
||||||
755
tools/fs.go
Normal file
755
tools/fs.go
Normal file
@@ -0,0 +1,755 @@
|
|||||||
|
package tools
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/base64"
|
||||||
|
"encoding/json"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"gf-lt/models"
|
||||||
|
"os"
|
||||||
|
"os/exec"
|
||||||
|
"path/filepath"
|
||||||
|
"sort"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
var memoryStore MemoryStore
|
||||||
|
var agentRole string
|
||||||
|
|
||||||
|
type MemoryStore interface {
|
||||||
|
Memorise(agent, topic, data string) (string, error)
|
||||||
|
Recall(agent, topic string) (string, error)
|
||||||
|
RecallTopics(agent string) ([]string, error)
|
||||||
|
Forget(agent, topic string) error
|
||||||
|
}
|
||||||
|
|
||||||
|
func SetMemoryStore(store MemoryStore, role string) {
|
||||||
|
memoryStore = store
|
||||||
|
agentRole = role
|
||||||
|
}
|
||||||
|
|
||||||
|
func SetFSRoot(dir string) {
|
||||||
|
if cfg == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
cfg.FilePickerDir = dir
|
||||||
|
}
|
||||||
|
|
||||||
|
func GetFSRoot() string {
|
||||||
|
return cfg.FilePickerDir
|
||||||
|
}
|
||||||
|
|
||||||
|
func SetFSCwd(dir string) error {
|
||||||
|
abs, err := filepath.Abs(dir)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
info, err := os.Stat(abs)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if !info.IsDir() {
|
||||||
|
return fmt.Errorf("not a directory: %s", dir)
|
||||||
|
}
|
||||||
|
cfg.FilePickerDir = abs
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func resolvePath(rel string) (string, error) {
|
||||||
|
if cfg.FilePickerDir == "" {
|
||||||
|
return "", errors.New("fs root not set")
|
||||||
|
}
|
||||||
|
if filepath.IsAbs(rel) {
|
||||||
|
abs := filepath.Clean(rel)
|
||||||
|
if !strings.HasPrefix(abs, cfg.FilePickerDir+string(os.PathSeparator)) && abs != cfg.FilePickerDir {
|
||||||
|
return "", fmt.Errorf("path escapes fs root: %s", rel)
|
||||||
|
}
|
||||||
|
return abs, nil
|
||||||
|
}
|
||||||
|
abs := filepath.Join(cfg.FilePickerDir, rel)
|
||||||
|
abs = filepath.Clean(abs)
|
||||||
|
if !strings.HasPrefix(abs, cfg.FilePickerDir+string(os.PathSeparator)) && abs != cfg.FilePickerDir {
|
||||||
|
return "", fmt.Errorf("path escapes fs root: %s", rel)
|
||||||
|
}
|
||||||
|
return abs, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func humanSize(n int64) string {
|
||||||
|
switch {
|
||||||
|
case n >= 1<<20:
|
||||||
|
return fmt.Sprintf("%.1fMB", float64(n)/float64(1<<20))
|
||||||
|
case n >= 1<<10:
|
||||||
|
return fmt.Sprintf("%.1fKB", float64(n)/float64(1<<10))
|
||||||
|
default:
|
||||||
|
return fmt.Sprintf("%dB", n)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func IsImageFile(path string) bool {
|
||||||
|
ext := strings.ToLower(filepath.Ext(path))
|
||||||
|
return ext == ".png" || ext == ".jpg" || ext == ".jpeg" || ext == ".gif" || ext == ".webp" || ext == ".svg"
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsLs(args []string, stdin string) string {
|
||||||
|
dir := ""
|
||||||
|
if len(args) > 0 {
|
||||||
|
dir = args[0]
|
||||||
|
}
|
||||||
|
abs, err := resolvePath(dir)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] %v", err)
|
||||||
|
}
|
||||||
|
entries, err := os.ReadDir(abs)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] ls: %v", err)
|
||||||
|
}
|
||||||
|
var out strings.Builder
|
||||||
|
for _, e := range entries {
|
||||||
|
info, _ := e.Info()
|
||||||
|
switch {
|
||||||
|
case e.IsDir():
|
||||||
|
fmt.Fprintf(&out, "d %-8s %s/\n", "-", e.Name())
|
||||||
|
case info != nil:
|
||||||
|
fmt.Fprintf(&out, "f %-8s %s\n", humanSize(info.Size()), e.Name())
|
||||||
|
default:
|
||||||
|
fmt.Fprintf(&out, "f %-8s %s\n", "?", e.Name())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if out.Len() == 0 {
|
||||||
|
return "(empty directory)"
|
||||||
|
}
|
||||||
|
return strings.TrimRight(out.String(), "\n")
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsCat(args []string, stdin string) string {
|
||||||
|
b64 := false
|
||||||
|
var path string
|
||||||
|
for _, a := range args {
|
||||||
|
if a == "-b" || a == "--base64" {
|
||||||
|
b64 = true
|
||||||
|
} else if path == "" {
|
||||||
|
path = a
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if path == "" {
|
||||||
|
return "[error] usage: cat <path>"
|
||||||
|
}
|
||||||
|
abs, err := resolvePath(path)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] %v", err)
|
||||||
|
}
|
||||||
|
data, err := os.ReadFile(abs)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] cat: %v", err)
|
||||||
|
}
|
||||||
|
if b64 {
|
||||||
|
result := base64.StdEncoding.EncodeToString(data)
|
||||||
|
if IsImageFile(path) {
|
||||||
|
result += fmt.Sprintf("\n", abs)
|
||||||
|
}
|
||||||
|
return result
|
||||||
|
}
|
||||||
|
return string(data)
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsViewImg(args []string, stdin string) string {
|
||||||
|
if len(args) == 0 {
|
||||||
|
return "[error] usage: view_img <image-path>"
|
||||||
|
}
|
||||||
|
path := args[0]
|
||||||
|
var abs string
|
||||||
|
if filepath.IsAbs(path) {
|
||||||
|
abs = path
|
||||||
|
} else {
|
||||||
|
var err error
|
||||||
|
abs, err = resolvePath(path)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if _, err := os.Stat(abs); err != nil {
|
||||||
|
return fmt.Sprintf("[error] view_img: %v", err)
|
||||||
|
}
|
||||||
|
if !IsImageFile(path) {
|
||||||
|
return fmt.Sprintf("[error] not an image file: %s (use cat to read text files)", path)
|
||||||
|
}
|
||||||
|
dataURL, err := models.CreateImageURLFromPath(abs)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] view_img: %v", err)
|
||||||
|
}
|
||||||
|
result := models.MultimodalToolResp{
|
||||||
|
Type: "multimodal_content",
|
||||||
|
Parts: []map[string]string{
|
||||||
|
{"type": "text", "text": "Image: " + path},
|
||||||
|
{"type": "image_url", "url": dataURL},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
jsonResult, err := json.Marshal(result)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] view_img: %v", err)
|
||||||
|
}
|
||||||
|
return string(jsonResult)
|
||||||
|
}
|
||||||
|
|
||||||
|
// FsSee is deprecated, use FsViewImg
|
||||||
|
func FsSee(args []string, stdin string) string {
|
||||||
|
return FsViewImg(args, stdin)
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsWrite(args []string, stdin string) string {
|
||||||
|
b64 := false
|
||||||
|
var path string
|
||||||
|
var contentParts []string
|
||||||
|
for _, a := range args {
|
||||||
|
switch a {
|
||||||
|
case "-b", "--base64":
|
||||||
|
b64 = true
|
||||||
|
default:
|
||||||
|
if path == "" {
|
||||||
|
path = a
|
||||||
|
} else {
|
||||||
|
contentParts = append(contentParts, a)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if path == "" {
|
||||||
|
return "[error] usage: write <path> [content] or pipe stdin"
|
||||||
|
}
|
||||||
|
abs, err := resolvePath(path)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] %v", err)
|
||||||
|
}
|
||||||
|
if err := os.MkdirAll(filepath.Dir(abs), 0o755); err != nil {
|
||||||
|
return fmt.Sprintf("[error] mkdir: %v", err)
|
||||||
|
}
|
||||||
|
var data []byte
|
||||||
|
if b64 {
|
||||||
|
src := stdin
|
||||||
|
if src == "" && len(contentParts) > 0 {
|
||||||
|
src = strings.Join(contentParts, " ")
|
||||||
|
}
|
||||||
|
src = strings.TrimSpace(src)
|
||||||
|
var err error
|
||||||
|
data, err = base64.StdEncoding.DecodeString(src)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] base64 decode: %v", err)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if len(contentParts) > 0 {
|
||||||
|
data = []byte(strings.Join(contentParts, " "))
|
||||||
|
} else {
|
||||||
|
data = []byte(stdin)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if err := os.WriteFile(abs, data, 0o644); err != nil {
|
||||||
|
return fmt.Sprintf("[error] write: %v", err)
|
||||||
|
}
|
||||||
|
size := humanSize(int64(len(data)))
|
||||||
|
result := fmt.Sprintf("Written %s → %s", size, path)
|
||||||
|
if IsImageFile(path) {
|
||||||
|
result += fmt.Sprintf("\n", abs)
|
||||||
|
}
|
||||||
|
return result
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsStat(args []string, stdin string) string {
|
||||||
|
if len(args) == 0 {
|
||||||
|
return "[error] usage: stat <path>"
|
||||||
|
}
|
||||||
|
abs, err := resolvePath(args[0])
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] %v", err)
|
||||||
|
}
|
||||||
|
info, err := os.Stat(abs)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] stat: %v", err)
|
||||||
|
}
|
||||||
|
mime := "application/octet-stream"
|
||||||
|
if IsImageFile(args[0]) {
|
||||||
|
ext := strings.ToLower(filepath.Ext(args[0]))
|
||||||
|
switch ext {
|
||||||
|
case ".png":
|
||||||
|
mime = "image/png"
|
||||||
|
case ".jpg", ".jpeg":
|
||||||
|
mime = "image/jpeg"
|
||||||
|
case ".gif":
|
||||||
|
mime = "image/gif"
|
||||||
|
case ".webp":
|
||||||
|
mime = "image/webp"
|
||||||
|
case ".svg":
|
||||||
|
mime = "image/svg+xml"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
var out strings.Builder
|
||||||
|
fmt.Fprintf(&out, "File: %s\n", args[0])
|
||||||
|
fmt.Fprintf(&out, "Size: %s (%d bytes)\n", humanSize(info.Size()), info.Size())
|
||||||
|
fmt.Fprintf(&out, "Type: %s\n", mime)
|
||||||
|
fmt.Fprintf(&out, "Modified: %s\n", info.ModTime().Format(time.RFC3339))
|
||||||
|
if info.IsDir() {
|
||||||
|
fmt.Fprintf(&out, "Kind: directory\n")
|
||||||
|
}
|
||||||
|
return strings.TrimRight(out.String(), "\n")
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsRm(args []string, stdin string) string {
|
||||||
|
if len(args) == 0 {
|
||||||
|
return "[error] usage: rm <path>"
|
||||||
|
}
|
||||||
|
abs, err := resolvePath(args[0])
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] %v", err)
|
||||||
|
}
|
||||||
|
if err := os.RemoveAll(abs); err != nil {
|
||||||
|
return fmt.Sprintf("[error] rm: %v", err)
|
||||||
|
}
|
||||||
|
return "Removed " + args[0]
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsCp(args []string, stdin string) string {
|
||||||
|
if len(args) < 2 {
|
||||||
|
return "[error] usage: cp <src> <dst>"
|
||||||
|
}
|
||||||
|
srcAbs, err := resolvePath(args[0])
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] %v", err)
|
||||||
|
}
|
||||||
|
dstAbs, err := resolvePath(args[1])
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] %v", err)
|
||||||
|
}
|
||||||
|
data, err := os.ReadFile(srcAbs)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] cp read: %v", err)
|
||||||
|
}
|
||||||
|
if err := os.MkdirAll(filepath.Dir(dstAbs), 0o755); err != nil {
|
||||||
|
return fmt.Sprintf("[error] cp mkdir: %v", err)
|
||||||
|
}
|
||||||
|
if err := os.WriteFile(dstAbs, data, 0o644); err != nil {
|
||||||
|
return fmt.Sprintf("[error] cp write: %v", err)
|
||||||
|
}
|
||||||
|
return fmt.Sprintf("Copied %s → %s (%s)", args[0], args[1], humanSize(int64(len(data))))
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsMv(args []string, stdin string) string {
|
||||||
|
if len(args) < 2 {
|
||||||
|
return "[error] usage: mv <src> <dst>"
|
||||||
|
}
|
||||||
|
srcAbs, err := resolvePath(args[0])
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] %v", err)
|
||||||
|
}
|
||||||
|
dstAbs, err := resolvePath(args[1])
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] %v", err)
|
||||||
|
}
|
||||||
|
if err := os.MkdirAll(filepath.Dir(dstAbs), 0o755); err != nil {
|
||||||
|
return fmt.Sprintf("[error] mv mkdir: %v", err)
|
||||||
|
}
|
||||||
|
if err := os.Rename(srcAbs, dstAbs); err != nil {
|
||||||
|
return fmt.Sprintf("[error] mv: %v", err)
|
||||||
|
}
|
||||||
|
return fmt.Sprintf("Moved %s → %s", args[0], args[1])
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsMkdir(args []string, stdin string) string {
|
||||||
|
if len(args) == 0 {
|
||||||
|
return "[error] usage: mkdir [-p] <dir>"
|
||||||
|
}
|
||||||
|
createParents := false
|
||||||
|
var dirPath string
|
||||||
|
for _, a := range args {
|
||||||
|
if a == "-p" || a == "--parents" {
|
||||||
|
createParents = true
|
||||||
|
} else if dirPath == "" {
|
||||||
|
dirPath = a
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if dirPath == "" {
|
||||||
|
return "[error] usage: mkdir [-p] <dir>"
|
||||||
|
}
|
||||||
|
abs, err := resolvePath(dirPath)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] %v", err)
|
||||||
|
}
|
||||||
|
var mkdirFunc func(string, os.FileMode) error
|
||||||
|
if createParents {
|
||||||
|
mkdirFunc = os.MkdirAll
|
||||||
|
} else {
|
||||||
|
mkdirFunc = os.Mkdir
|
||||||
|
}
|
||||||
|
if err := mkdirFunc(abs, 0o755); err != nil {
|
||||||
|
return fmt.Sprintf("[error] mkdir: %v", err)
|
||||||
|
}
|
||||||
|
if createParents {
|
||||||
|
return "Created " + dirPath + " (with parents)"
|
||||||
|
}
|
||||||
|
return "Created " + dirPath
|
||||||
|
}
|
||||||
|
|
||||||
|
// Text processing commands
|
||||||
|
|
||||||
|
func FsEcho(args []string, stdin string) string {
|
||||||
|
if stdin != "" {
|
||||||
|
return stdin
|
||||||
|
}
|
||||||
|
return strings.Join(args, " ")
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsTime(args []string, stdin string) string {
|
||||||
|
return time.Now().Format("2006-01-02 15:04:05 MST")
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsGrep(args []string, stdin string) string {
|
||||||
|
if len(args) == 0 {
|
||||||
|
return "[error] usage: grep [-i] [-v] [-c] <pattern>"
|
||||||
|
}
|
||||||
|
ignoreCase := false
|
||||||
|
invert := false
|
||||||
|
countOnly := false
|
||||||
|
var pattern string
|
||||||
|
for _, a := range args {
|
||||||
|
switch a {
|
||||||
|
case "-i":
|
||||||
|
ignoreCase = true
|
||||||
|
case "-v":
|
||||||
|
invert = true
|
||||||
|
case "-c":
|
||||||
|
countOnly = true
|
||||||
|
default:
|
||||||
|
pattern = a
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if pattern == "" {
|
||||||
|
return "[error] pattern required"
|
||||||
|
}
|
||||||
|
if ignoreCase {
|
||||||
|
pattern = strings.ToLower(pattern)
|
||||||
|
}
|
||||||
|
lines := strings.Split(stdin, "\n")
|
||||||
|
var matched []string
|
||||||
|
for _, line := range lines {
|
||||||
|
haystack := line
|
||||||
|
if ignoreCase {
|
||||||
|
haystack = strings.ToLower(line)
|
||||||
|
}
|
||||||
|
match := strings.Contains(haystack, pattern)
|
||||||
|
if invert {
|
||||||
|
match = !match
|
||||||
|
}
|
||||||
|
if match {
|
||||||
|
matched = append(matched, line)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if countOnly {
|
||||||
|
return strconv.Itoa(len(matched))
|
||||||
|
}
|
||||||
|
return strings.Join(matched, "\n")
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsHead(args []string, stdin string) string {
|
||||||
|
n := 10
|
||||||
|
for i, a := range args {
|
||||||
|
if a == "-n" && i+1 < len(args) {
|
||||||
|
if parsed, err := strconv.Atoi(args[i+1]); err == nil {
|
||||||
|
n = parsed
|
||||||
|
}
|
||||||
|
} else if strings.HasPrefix(a, "-") {
|
||||||
|
continue
|
||||||
|
} else if parsed, err := strconv.Atoi(a); err == nil {
|
||||||
|
n = parsed
|
||||||
|
}
|
||||||
|
}
|
||||||
|
lines := strings.Split(stdin, "\n")
|
||||||
|
if n > 0 && len(lines) > n {
|
||||||
|
lines = lines[:n]
|
||||||
|
}
|
||||||
|
return strings.Join(lines, "\n")
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsTail(args []string, stdin string) string {
|
||||||
|
n := 10
|
||||||
|
for i, a := range args {
|
||||||
|
if a == "-n" && i+1 < len(args) {
|
||||||
|
if parsed, err := strconv.Atoi(args[i+1]); err == nil {
|
||||||
|
n = parsed
|
||||||
|
}
|
||||||
|
} else if strings.HasPrefix(a, "-") {
|
||||||
|
continue
|
||||||
|
} else if parsed, err := strconv.Atoi(a); err == nil {
|
||||||
|
n = parsed
|
||||||
|
}
|
||||||
|
}
|
||||||
|
lines := strings.Split(stdin, "\n")
|
||||||
|
if n > 0 && len(lines) > n {
|
||||||
|
lines = lines[len(lines)-n:]
|
||||||
|
}
|
||||||
|
return strings.Join(lines, "\n")
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsWc(args []string, stdin string) string {
|
||||||
|
lines := len(strings.Split(stdin, "\n"))
|
||||||
|
words := len(strings.Fields(stdin))
|
||||||
|
chars := len(stdin)
|
||||||
|
if len(args) > 0 {
|
||||||
|
switch args[0] {
|
||||||
|
case "-l":
|
||||||
|
return strconv.Itoa(lines)
|
||||||
|
case "-w":
|
||||||
|
return strconv.Itoa(words)
|
||||||
|
case "-c":
|
||||||
|
return strconv.Itoa(chars)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return fmt.Sprintf("%d lines, %d words, %d chars", lines, words, chars)
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsSort(args []string, stdin string) string {
|
||||||
|
lines := strings.Split(stdin, "\n")
|
||||||
|
reverse := false
|
||||||
|
numeric := false
|
||||||
|
for _, a := range args {
|
||||||
|
switch a {
|
||||||
|
case "-r":
|
||||||
|
reverse = true
|
||||||
|
case "-n":
|
||||||
|
numeric = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sortFunc := func(i, j int) bool {
|
||||||
|
if numeric {
|
||||||
|
ni, _ := strconv.Atoi(lines[i])
|
||||||
|
nj, _ := strconv.Atoi(lines[j])
|
||||||
|
if reverse {
|
||||||
|
return ni > nj
|
||||||
|
}
|
||||||
|
return ni < nj
|
||||||
|
}
|
||||||
|
if reverse {
|
||||||
|
return lines[i] > lines[j]
|
||||||
|
}
|
||||||
|
return lines[i] < lines[j]
|
||||||
|
}
|
||||||
|
sort.Slice(lines, sortFunc)
|
||||||
|
return strings.Join(lines, "\n")
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsUniq(args []string, stdin string) string {
|
||||||
|
lines := strings.Split(stdin, "\n")
|
||||||
|
showCount := false
|
||||||
|
for _, a := range args {
|
||||||
|
if a == "-c" {
|
||||||
|
showCount = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
var result []string
|
||||||
|
var prev string
|
||||||
|
first := true
|
||||||
|
count := 0
|
||||||
|
for _, line := range lines {
|
||||||
|
if first || line != prev {
|
||||||
|
if !first && showCount {
|
||||||
|
result = append(result, fmt.Sprintf("%d %s", count, prev))
|
||||||
|
} else if !first {
|
||||||
|
result = append(result, prev)
|
||||||
|
}
|
||||||
|
count = 1
|
||||||
|
prev = line
|
||||||
|
first = false
|
||||||
|
} else {
|
||||||
|
count++
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if !first {
|
||||||
|
if showCount {
|
||||||
|
result = append(result, fmt.Sprintf("%d %s", count, prev))
|
||||||
|
} else {
|
||||||
|
result = append(result, prev)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return strings.Join(result, "\n")
|
||||||
|
}
|
||||||
|
|
||||||
|
var allowedGitSubcommands = map[string]bool{
|
||||||
|
"status": true,
|
||||||
|
"log": true,
|
||||||
|
"diff": true,
|
||||||
|
"show": true,
|
||||||
|
"branch": true,
|
||||||
|
"reflog": true,
|
||||||
|
"rev-parse": true,
|
||||||
|
"shortlog": true,
|
||||||
|
"describe": true,
|
||||||
|
"rev-list": true,
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsGit(args []string, stdin string) string {
|
||||||
|
if len(args) == 0 {
|
||||||
|
return "[error] usage: git <subcommand> [options]"
|
||||||
|
}
|
||||||
|
subcmd := args[0]
|
||||||
|
if !allowedGitSubcommands[subcmd] {
|
||||||
|
return fmt.Sprintf("[error] git: '%s' is not an allowed git command. Allowed: status, log, diff, show, branch, reflog, rev-parse, shortlog, describe, rev-list", subcmd)
|
||||||
|
}
|
||||||
|
abs, err := resolvePath(".")
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] git: %v", err)
|
||||||
|
}
|
||||||
|
// Pass all args to git (first arg is subcommand, rest are options)
|
||||||
|
cmd := exec.Command("git", args...)
|
||||||
|
cmd.Dir = abs
|
||||||
|
output, err := cmd.CombinedOutput()
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] git %s: %v\n%s", subcmd, err, string(output))
|
||||||
|
}
|
||||||
|
return string(output)
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsPwd(args []string, stdin string) string {
|
||||||
|
return cfg.FilePickerDir
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsCd(args []string, stdin string) string {
|
||||||
|
if len(args) == 0 {
|
||||||
|
return "[error] usage: cd <dir>"
|
||||||
|
}
|
||||||
|
dir := args[0]
|
||||||
|
abs, err := resolvePath(dir)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] cd: %v", err)
|
||||||
|
}
|
||||||
|
info, err := os.Stat(abs)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] cd: %v", err)
|
||||||
|
}
|
||||||
|
if !info.IsDir() {
|
||||||
|
return "[error] cd: not a directory: " + dir
|
||||||
|
}
|
||||||
|
cfg.FilePickerDir = abs
|
||||||
|
return "Changed directory to: " + cfg.FilePickerDir
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsSed(args []string, stdin string) string {
|
||||||
|
if len(args) == 0 {
|
||||||
|
return "[error] usage: sed 's/old/new/[g]' [file]"
|
||||||
|
}
|
||||||
|
inPlace := false
|
||||||
|
var filePath string
|
||||||
|
var pattern string
|
||||||
|
for _, a := range args {
|
||||||
|
switch a {
|
||||||
|
case "-i", "--in-place":
|
||||||
|
inPlace = true
|
||||||
|
default:
|
||||||
|
if strings.HasPrefix(a, "s") && len(a) > 1 {
|
||||||
|
pattern = a
|
||||||
|
} else if filePath == "" && !strings.HasPrefix(a, "-") {
|
||||||
|
filePath = a
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if pattern == "" {
|
||||||
|
return "[error] usage: sed 's/old/new/[g]' [file]"
|
||||||
|
}
|
||||||
|
// Parse pattern: s/old/new/flags
|
||||||
|
parts := strings.Split(pattern[1:], "/")
|
||||||
|
if len(parts) < 2 {
|
||||||
|
return "[error] invalid sed pattern. Use: s/old/new/[g]"
|
||||||
|
}
|
||||||
|
oldStr := parts[0]
|
||||||
|
newStr := parts[1]
|
||||||
|
global := len(parts) >= 3 && strings.Contains(parts[2], "g")
|
||||||
|
var content string
|
||||||
|
switch {
|
||||||
|
case filePath != "" && stdin == "":
|
||||||
|
abs, err := resolvePath(filePath)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] sed: %v", err)
|
||||||
|
}
|
||||||
|
data, err := os.ReadFile(abs)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] sed: %v", err)
|
||||||
|
}
|
||||||
|
content = string(data)
|
||||||
|
case stdin != "":
|
||||||
|
content = stdin
|
||||||
|
default:
|
||||||
|
return "[error] sed: no input (use file path or pipe from stdin)"
|
||||||
|
}
|
||||||
|
// Apply sed replacement
|
||||||
|
if global {
|
||||||
|
content = strings.ReplaceAll(content, oldStr, newStr)
|
||||||
|
} else {
|
||||||
|
content = strings.Replace(content, oldStr, newStr, 1)
|
||||||
|
}
|
||||||
|
if inPlace && filePath != "" {
|
||||||
|
abs, err := resolvePath(filePath)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] sed: %v", err)
|
||||||
|
}
|
||||||
|
if err := os.WriteFile(abs, []byte(content), 0644); err != nil {
|
||||||
|
return fmt.Sprintf("[error] sed: %v", err)
|
||||||
|
}
|
||||||
|
return "Modified " + filePath
|
||||||
|
}
|
||||||
|
return content
|
||||||
|
}
|
||||||
|
|
||||||
|
func FsMemory(args []string, stdin string) string {
|
||||||
|
if len(args) == 0 {
|
||||||
|
return "[error] usage: memory store <topic> <data> | memory get <topic> | memory list | memory forget <topic>"
|
||||||
|
}
|
||||||
|
if memoryStore == nil {
|
||||||
|
return "[error] memory store not initialized"
|
||||||
|
}
|
||||||
|
switch args[0] {
|
||||||
|
case "store":
|
||||||
|
if len(args) < 3 && stdin == "" {
|
||||||
|
return "[error] usage: memory store <topic> <data>"
|
||||||
|
}
|
||||||
|
topic := args[1]
|
||||||
|
var data string
|
||||||
|
if len(args) >= 3 {
|
||||||
|
data = strings.Join(args[2:], " ")
|
||||||
|
} else {
|
||||||
|
data = stdin
|
||||||
|
}
|
||||||
|
_, err := memoryStore.Memorise(agentRole, topic, data)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] failed to store: %v", err)
|
||||||
|
}
|
||||||
|
return "Stored under topic: " + topic
|
||||||
|
case "get":
|
||||||
|
if len(args) < 2 {
|
||||||
|
return "[error] usage: memory get <topic>"
|
||||||
|
}
|
||||||
|
topic := args[1]
|
||||||
|
data, err := memoryStore.Recall(agentRole, topic)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] failed to recall: %v", err)
|
||||||
|
}
|
||||||
|
return fmt.Sprintf("Topic: %s\n%s", topic, data)
|
||||||
|
case "list", "topics":
|
||||||
|
topics, err := memoryStore.RecallTopics(agentRole)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] failed to list topics: %v", err)
|
||||||
|
}
|
||||||
|
if len(topics) == 0 {
|
||||||
|
return "No topics stored."
|
||||||
|
}
|
||||||
|
return "Topics: " + strings.Join(topics, ", ")
|
||||||
|
case "forget", "delete":
|
||||||
|
if len(args) < 2 {
|
||||||
|
return "[error] usage: memory forget <topic>"
|
||||||
|
}
|
||||||
|
topic := args[1]
|
||||||
|
err := memoryStore.Forget(agentRole, topic)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Sprintf("[error] failed to forget: %v", err)
|
||||||
|
}
|
||||||
|
return "Deleted topic: " + topic
|
||||||
|
default:
|
||||||
|
return fmt.Sprintf("[error] unknown subcommand: %s. Use: store, get, list, topics, forget, delete", args[0])
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
package main
|
package tools
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
@@ -12,87 +12,6 @@ import (
|
|||||||
"github.com/playwright-community/playwright-go"
|
"github.com/playwright-community/playwright-go"
|
||||||
)
|
)
|
||||||
|
|
||||||
var browserToolSysMsg = `
|
|
||||||
Additional browser automation tools (Playwright):
|
|
||||||
[
|
|
||||||
{
|
|
||||||
"name": "pw_start",
|
|
||||||
"args": [],
|
|
||||||
"when_to_use": "start a browser instance before doing any browser automation. Must be called first."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_stop",
|
|
||||||
"args": [],
|
|
||||||
"when_to_use": "stop the browser instance when done with automation."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_is_running",
|
|
||||||
"args": [],
|
|
||||||
"when_to_use": "check if browser is currently running."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_navigate",
|
|
||||||
"args": ["url"],
|
|
||||||
"when_to_use": "open a specific URL in the web browser."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_click",
|
|
||||||
"args": ["selector", "index"],
|
|
||||||
"when_to_use": "click on an element on the current webpage. Use 'index' for multiple matches (default 0)."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_fill",
|
|
||||||
"args": ["selector", "text", "index"],
|
|
||||||
"when_to_use": "type text into an input field. Use 'index' for multiple matches (default 0)."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_extract_text",
|
|
||||||
"args": ["selector"],
|
|
||||||
"when_to_use": "extract text content from the page or specific elements. Use selector 'body' for all page text."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_screenshot",
|
|
||||||
"args": ["selector", "full_page"],
|
|
||||||
"when_to_use": "take a screenshot of the page or a specific element. Returns a file path to the image. Use to verify actions or inspect visual state."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_screenshot_and_view",
|
|
||||||
"args": ["selector", "full_page"],
|
|
||||||
"when_to_use": "take a screenshot and return the image for viewing. Use to visually verify page state."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_wait_for_selector",
|
|
||||||
"args": ["selector", "timeout"],
|
|
||||||
"when_to_use": "wait for an element to appear on the page before proceeding with further actions."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_drag",
|
|
||||||
"args": ["x1", "y1", "x2", "y2"],
|
|
||||||
"when_to_use": "drag the mouse from point (x1,y1) to (x2,y2)."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_click_at",
|
|
||||||
"args": ["x", "y"],
|
|
||||||
"when_to_use": "click at specific X,Y coordinates on the page. Use when you know the exact position."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_get_html",
|
|
||||||
"args": ["selector"],
|
|
||||||
"when_to_use": "get the HTML content of the page or a specific element. Use to understand page structure or extract raw HTML."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_get_dom",
|
|
||||||
"args": ["selector"],
|
|
||||||
"when_to_use": "get a structured DOM representation with tag, attributes, text, and children. Use to inspect element hierarchy and properties."
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"name": "pw_search_elements",
|
|
||||||
"args": ["text", "selector"],
|
|
||||||
"when_to_use": "search for elements by text content or CSS selector. Returns matching elements with their tags, text, and HTML."
|
|
||||||
}
|
|
||||||
]
|
|
||||||
`
|
|
||||||
|
|
||||||
var (
|
var (
|
||||||
pw *playwright.Playwright
|
pw *playwright.Playwright
|
||||||
browser playwright.Browser
|
browser playwright.Browser
|
||||||
@@ -101,7 +20,7 @@ var (
|
|||||||
page playwright.Page
|
page playwright.Page
|
||||||
)
|
)
|
||||||
|
|
||||||
func pwShutDown() error {
|
func PwShutDown() error {
|
||||||
if pw == nil {
|
if pw == nil {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@@ -109,7 +28,7 @@ func pwShutDown() error {
|
|||||||
return pw.Stop()
|
return pw.Stop()
|
||||||
}
|
}
|
||||||
|
|
||||||
func installPW() error {
|
func InstallPW() error {
|
||||||
err := playwright.Install(&playwright.RunOptions{Verbose: false})
|
err := playwright.Install(&playwright.RunOptions{Verbose: false})
|
||||||
if err != nil {
|
if err != nil {
|
||||||
logger.Warn("playwright not available", "error", err)
|
logger.Warn("playwright not available", "error", err)
|
||||||
@@ -118,7 +37,7 @@ func installPW() error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func checkPlaywright() error {
|
func CheckPlaywright() error {
|
||||||
var err error
|
var err error
|
||||||
pw, err = playwright.Run()
|
pw, err = playwright.Run()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -455,6 +374,78 @@ func pwDrag(args map[string]string) []byte {
|
|||||||
return []byte(fmt.Sprintf(`{"success": true, "message": "Dragged from (%s,%s) to (%s,%s)"}`, x1, y1, x2, y2))
|
return []byte(fmt.Sprintf(`{"success": true, "message": "Dragged from (%s,%s) to (%s,%s)"}`, x1, y1, x2, y2))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func pwDragBySelector(args map[string]string) []byte {
|
||||||
|
fromSelector, ok := args["fromSelector"]
|
||||||
|
if !ok || fromSelector == "" {
|
||||||
|
return []byte(`{"error": "fromSelector not provided"}`)
|
||||||
|
}
|
||||||
|
toSelector, ok := args["toSelector"]
|
||||||
|
if !ok || toSelector == "" {
|
||||||
|
return []byte(`{"error": "toSelector not provided"}`)
|
||||||
|
}
|
||||||
|
if !browserStarted || page == nil {
|
||||||
|
return []byte(`{"error": "Browser not started. Call pw_start first."}`)
|
||||||
|
}
|
||||||
|
fromJS := fmt.Sprintf(`
|
||||||
|
function getCenter(selector) {
|
||||||
|
const el = document.querySelector(selector);
|
||||||
|
if (!el) return null;
|
||||||
|
const rect = el.getBoundingClientRect();
|
||||||
|
return { x: rect.left + rect.width / 2, y: rect.top + rect.height / 2 };
|
||||||
|
}
|
||||||
|
getCenter(%q)
|
||||||
|
`, fromSelector)
|
||||||
|
toJS := fmt.Sprintf(`
|
||||||
|
function getCenter(selector) {
|
||||||
|
const el = document.querySelector(selector);
|
||||||
|
if (!el) return null;
|
||||||
|
const rect = el.getBoundingClientRect();
|
||||||
|
return { x: rect.left + rect.width / 2, y: rect.top + rect.height / 2 };
|
||||||
|
}
|
||||||
|
getCenter(%q)
|
||||||
|
`, toSelector)
|
||||||
|
fromResult, err := page.Evaluate(fromJS)
|
||||||
|
if err != nil {
|
||||||
|
return []byte(fmt.Sprintf(`{"error": "failed to get from element: %s"}`, err.Error()))
|
||||||
|
}
|
||||||
|
fromMap, ok := fromResult.(map[string]interface{})
|
||||||
|
if !ok || fromMap == nil {
|
||||||
|
return []byte(fmt.Sprintf(`{"error": "from selector '%s' not found"}`, fromSelector))
|
||||||
|
}
|
||||||
|
fromX := fromMap["x"].(float64)
|
||||||
|
fromY := fromMap["y"].(float64)
|
||||||
|
toResult, err := page.Evaluate(toJS)
|
||||||
|
if err != nil {
|
||||||
|
return []byte(fmt.Sprintf(`{"error": "failed to get to element: %s"}`, err.Error()))
|
||||||
|
}
|
||||||
|
toMap, ok := toResult.(map[string]interface{})
|
||||||
|
if !ok || toMap == nil {
|
||||||
|
return []byte(fmt.Sprintf(`{"error": "to selector '%s' not found"}`, toSelector))
|
||||||
|
}
|
||||||
|
toX := toMap["x"].(float64)
|
||||||
|
toY := toMap["y"].(float64)
|
||||||
|
mouse := page.Mouse()
|
||||||
|
err = mouse.Move(fromX, fromY)
|
||||||
|
if err != nil {
|
||||||
|
return []byte(fmt.Sprintf(`{"error": "failed to move mouse: %s"}`, err.Error()))
|
||||||
|
}
|
||||||
|
err = mouse.Down()
|
||||||
|
if err != nil {
|
||||||
|
return []byte(fmt.Sprintf(`{"error": "failed to mouse down: %s"}`, err.Error()))
|
||||||
|
}
|
||||||
|
err = mouse.Move(toX, toY)
|
||||||
|
if err != nil {
|
||||||
|
return []byte(fmt.Sprintf(`{"error": "failed to move mouse: %s"}`, err.Error()))
|
||||||
|
}
|
||||||
|
err = mouse.Up()
|
||||||
|
if err != nil {
|
||||||
|
return []byte(fmt.Sprintf(`{"error": "failed to mouse up: %s"}`, err.Error()))
|
||||||
|
}
|
||||||
|
msg := fmt.Sprintf("Dragged from %s (%.0f,%.0f) to %s (%.0f,%.0f)", fromSelector, fromX, fromY, toSelector, toX, toY)
|
||||||
|
return []byte(fmt.Sprintf(`{"success": true, "message": "%s"}`, msg))
|
||||||
|
}
|
||||||
|
|
||||||
|
// nolint:unused
|
||||||
func pwClickAt(args map[string]string) []byte {
|
func pwClickAt(args map[string]string) []byte {
|
||||||
x, ok := args["x"]
|
x, ok := args["x"]
|
||||||
if !ok {
|
if !ok {
|
||||||
@@ -605,6 +596,7 @@ func pwGetDOM(args map[string]string) []byte {
|
|||||||
return []byte(fmt.Sprintf(`{"dom": %s}`, string(data)))
|
return []byte(fmt.Sprintf(`{"dom": %s}`, string(data)))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// nolint:unused
|
||||||
func pwSearchElements(args map[string]string) []byte {
|
func pwSearchElements(args map[string]string) []byte {
|
||||||
text := args["text"]
|
text := args["text"]
|
||||||
selector := args["selector"]
|
selector := args["selector"]
|
||||||
1914
tools/tools.go
Normal file
1914
tools/tools.go
Normal file
File diff suppressed because it is too large
Load Diff
60
tui.go
60
tui.go
@@ -3,6 +3,7 @@ package main
|
|||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"gf-lt/models"
|
"gf-lt/models"
|
||||||
|
"gf-lt/tools"
|
||||||
"image"
|
"image"
|
||||||
_ "image/jpeg"
|
_ "image/jpeg"
|
||||||
_ "image/png"
|
_ "image/png"
|
||||||
@@ -42,7 +43,6 @@ var (
|
|||||||
confirmPageName = "confirm"
|
confirmPageName = "confirm"
|
||||||
fullscreenMode bool
|
fullscreenMode bool
|
||||||
positionVisible bool = true
|
positionVisible bool = true
|
||||||
scrollToEndEnabled bool = true
|
|
||||||
// pages
|
// pages
|
||||||
historyPage = "historyPage"
|
historyPage = "historyPage"
|
||||||
agentPage = "agentPage"
|
agentPage = "agentPage"
|
||||||
@@ -51,7 +51,6 @@ var (
|
|||||||
helpPage = "helpPage"
|
helpPage = "helpPage"
|
||||||
renamePage = "renamePage"
|
renamePage = "renamePage"
|
||||||
RAGPage = "RAGPage"
|
RAGPage = "RAGPage"
|
||||||
RAGLoadedPage = "RAGLoadedPage"
|
|
||||||
propsPage = "propsPage"
|
propsPage = "propsPage"
|
||||||
codeBlockPage = "codeBlockPage"
|
codeBlockPage = "codeBlockPage"
|
||||||
imgPage = "imgPage"
|
imgPage = "imgPage"
|
||||||
@@ -70,8 +69,8 @@ var (
|
|||||||
[yellow]F4[white]: edit msg
|
[yellow]F4[white]: edit msg
|
||||||
[yellow]F5[white]: toggle fullscreen for input/chat window
|
[yellow]F5[white]: toggle fullscreen for input/chat window
|
||||||
[yellow]F6[white]: interrupt bot resp
|
[yellow]F6[white]: interrupt bot resp
|
||||||
[yellow]F7[white]: copy last msg to clipboard (linux xclip)
|
[yellow]F7[white]: copy last msg to clipboard (linux xclip or wl-copy)
|
||||||
[yellow]F8[white]: copy n msg to clipboard (linux xclip)
|
[yellow]F8[white]: copy n msg to clipboard (linux xclip or wl-copy)
|
||||||
[yellow]F9[white]: table to copy from; with all code blocks
|
[yellow]F9[white]: table to copy from; with all code blocks
|
||||||
[yellow]F10[white]: switch if LLM will respond on this message (for user to write multiple messages in a row)
|
[yellow]F10[white]: switch if LLM will respond on this message (for user to write multiple messages in a row)
|
||||||
[yellow]F11[white]: import json chat file
|
[yellow]F11[white]: import json chat file
|
||||||
@@ -106,6 +105,7 @@ var (
|
|||||||
[yellow]Alt+t[white]: toggle thinking blocks visibility (collapse/expand <think> blocks)
|
[yellow]Alt+t[white]: toggle thinking blocks visibility (collapse/expand <think> blocks)
|
||||||
[yellow]Ctrl+t[white]: toggle tool call/response visibility (collapse/expand tool calls and non-shell tool responses)
|
[yellow]Ctrl+t[white]: toggle tool call/response visibility (collapse/expand tool calls and non-shell tool responses)
|
||||||
[yellow]Alt+i[white]: show colorscheme selection popup
|
[yellow]Alt+i[white]: show colorscheme selection popup
|
||||||
|
[yellow]Insert[white]: paste from clipboard to the text area (use it instead shift+insert)
|
||||||
|
|
||||||
=== scrolling chat window (some keys similar to vim) ===
|
=== scrolling chat window (some keys similar to vim) ===
|
||||||
[yellow]arrows up/down and j/k[white]: scroll up and down
|
[yellow]arrows up/down and j/k[white]: scroll up and down
|
||||||
@@ -224,12 +224,13 @@ func showToast(title, message string) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func init() {
|
func initTUI() {
|
||||||
// Start background goroutine to update model color cache
|
// Start background goroutine to update model color cache
|
||||||
startModelColorUpdater()
|
startModelColorUpdater()
|
||||||
tview.Styles = colorschemes["default"]
|
tview.Styles = colorschemes["default"]
|
||||||
app = tview.NewApplication()
|
app = tview.NewApplication()
|
||||||
pages = tview.NewPages()
|
pages = tview.NewPages()
|
||||||
|
outputHandler = &TUIOutputHandler{tv: textView}
|
||||||
shellInput = tview.NewInputField().
|
shellInput = tview.NewInputField().
|
||||||
SetLabel(fmt.Sprintf("[%s]$ ", cfg.FilePickerDir)). // dynamic prompt
|
SetLabel(fmt.Sprintf("[%s]$ ", cfg.FilePickerDir)). // dynamic prompt
|
||||||
SetFieldWidth(0).
|
SetFieldWidth(0).
|
||||||
@@ -320,6 +321,24 @@ func init() {
|
|||||||
textArea = tview.NewTextArea().
|
textArea = tview.NewTextArea().
|
||||||
SetPlaceholder("input is multiline; press <Enter> to start the next line;\npress <Esc> to send the message.")
|
SetPlaceholder("input is multiline; press <Enter> to start the next line;\npress <Esc> to send the message.")
|
||||||
textArea.SetBorder(true).SetTitle("input")
|
textArea.SetBorder(true).SetTitle("input")
|
||||||
|
textArea.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
||||||
|
if event.Key() == tcell.KeyInsert {
|
||||||
|
text, err := readFromClipboard()
|
||||||
|
if err != nil {
|
||||||
|
logger.Error("failed to read clipboard", "error", err)
|
||||||
|
return event
|
||||||
|
}
|
||||||
|
maxPaste := 100000
|
||||||
|
if len(text) > maxPaste {
|
||||||
|
text = text[:maxPaste]
|
||||||
|
showToast("paste truncated", "pasted text exceeded 100KB limit")
|
||||||
|
}
|
||||||
|
current := textArea.GetText()
|
||||||
|
textArea.SetText(current+text, true)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return event
|
||||||
|
})
|
||||||
textView = tview.NewTextView().
|
textView = tview.NewTextView().
|
||||||
SetDynamicColors(true).
|
SetDynamicColors(true).
|
||||||
SetRegions(true).
|
SetRegions(true).
|
||||||
@@ -634,7 +653,7 @@ func init() {
|
|||||||
updateStatusLine()
|
updateStatusLine()
|
||||||
textView.SetText(chatToText(chatBody.Messages, cfg.ShowSys))
|
textView.SetText(chatToText(chatBody.Messages, cfg.ShowSys))
|
||||||
colorText()
|
colorText()
|
||||||
if scrollToEndEnabled {
|
if cfg.AutoScrollEnabled {
|
||||||
textView.ScrollToEnd()
|
textView.ScrollToEnd()
|
||||||
}
|
}
|
||||||
// init sysmap
|
// init sysmap
|
||||||
@@ -663,9 +682,9 @@ func init() {
|
|||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyRune && event.Rune() == '2' && event.Modifiers()&tcell.ModAlt != 0 {
|
if event.Key() == tcell.KeyRune && event.Rune() == '2' && event.Modifiers()&tcell.ModAlt != 0 {
|
||||||
// toggle auto-scrolling
|
// toggle auto-scrolling
|
||||||
scrollToEndEnabled = !scrollToEndEnabled
|
cfg.AutoScrollEnabled = !cfg.AutoScrollEnabled
|
||||||
status := "disabled"
|
status := "disabled"
|
||||||
if scrollToEndEnabled {
|
if cfg.AutoScrollEnabled {
|
||||||
status = "enabled"
|
status = "enabled"
|
||||||
}
|
}
|
||||||
showToast("autoscroll", "Auto-scrolling "+status)
|
showToast("autoscroll", "Auto-scrolling "+status)
|
||||||
@@ -732,7 +751,7 @@ func init() {
|
|||||||
updateStatusLine()
|
updateStatusLine()
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyF2 && !botRespMode {
|
if event.Key() == tcell.KeyF2 && !botRespMode.Load() {
|
||||||
// regen last msg
|
// regen last msg
|
||||||
if len(chatBody.Messages) == 0 {
|
if len(chatBody.Messages) == 0 {
|
||||||
showToast("info", "no messages to regenerate")
|
showToast("info", "no messages to regenerate")
|
||||||
@@ -749,7 +768,7 @@ func init() {
|
|||||||
chatRoundChan <- &models.ChatRoundReq{Role: cfg.UserRole, Regen: true}
|
chatRoundChan <- &models.ChatRoundReq{Role: cfg.UserRole, Regen: true}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyF3 && !botRespMode {
|
if event.Key() == tcell.KeyF3 && !botRespMode.Load() {
|
||||||
// delete last msg
|
// delete last msg
|
||||||
// check textarea text; if it ends with bot icon delete only icon:
|
// check textarea text; if it ends with bot icon delete only icon:
|
||||||
text := textView.GetText(true)
|
text := textView.GetText(true)
|
||||||
@@ -805,9 +824,9 @@ func init() {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyF6 {
|
if event.Key() == tcell.KeyF6 {
|
||||||
interruptResp = true
|
interruptResp.Store(true)
|
||||||
botRespMode = false
|
botRespMode.Store(false)
|
||||||
toolRunningMode = false
|
toolRunningMode.Store(false)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyF7 {
|
if event.Key() == tcell.KeyF7 {
|
||||||
@@ -832,7 +851,7 @@ func init() {
|
|||||||
if event.Key() == tcell.KeyF9 {
|
if event.Key() == tcell.KeyF9 {
|
||||||
// table of codeblocks to copy
|
// table of codeblocks to copy
|
||||||
text := textView.GetText(false)
|
text := textView.GetText(false)
|
||||||
cb := codeBlockRE.FindAllString(text, -1)
|
cb := models.CodeBlockRE.FindAllString(text, -1)
|
||||||
if len(cb) == 0 {
|
if len(cb) == 0 {
|
||||||
showToast("notify", "no code blocks in chat")
|
showToast("notify", "no code blocks in chat")
|
||||||
return nil
|
return nil
|
||||||
@@ -931,7 +950,7 @@ func init() {
|
|||||||
if event.Key() == tcell.KeyCtrlK {
|
if event.Key() == tcell.KeyCtrlK {
|
||||||
// add message from tools
|
// add message from tools
|
||||||
cfg.ToolUse = !cfg.ToolUse
|
cfg.ToolUse = !cfg.ToolUse
|
||||||
updateToolCapabilities()
|
UpdateToolCapabilities()
|
||||||
updateStatusLine()
|
updateStatusLine()
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@@ -1037,7 +1056,7 @@ func init() {
|
|||||||
if event.Key() == tcell.KeyCtrlC {
|
if event.Key() == tcell.KeyCtrlC {
|
||||||
logger.Info("caught Ctrl+C via tcell event")
|
logger.Info("caught Ctrl+C via tcell event")
|
||||||
go func() {
|
go func() {
|
||||||
if err := pwShutDown(); err != nil {
|
if err := tools.PwShutDown(); err != nil {
|
||||||
logger.Error("shutdown failed", "err", err)
|
logger.Error("shutdown failed", "err", err)
|
||||||
}
|
}
|
||||||
app.Stop()
|
app.Stop()
|
||||||
@@ -1102,7 +1121,7 @@ func init() {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
// cannot send msg in editMode or botRespMode
|
// cannot send msg in editMode or botRespMode
|
||||||
if event.Key() == tcell.KeyEscape && !editMode && !botRespMode {
|
if event.Key() == tcell.KeyEscape && !editMode && !botRespMode.Load() {
|
||||||
if shellMode {
|
if shellMode {
|
||||||
cmdText := shellInput.GetText()
|
cmdText := shellInput.GetText()
|
||||||
if cmdText != "" {
|
if cmdText != "" {
|
||||||
@@ -1129,7 +1148,7 @@ func init() {
|
|||||||
}
|
}
|
||||||
// check if plain text
|
// check if plain text
|
||||||
if !injectRole {
|
if !injectRole {
|
||||||
matches := roleRE.FindStringSubmatch(msgText)
|
matches := models.RoleRE.FindStringSubmatch(msgText)
|
||||||
if len(matches) > 1 {
|
if len(matches) > 1 {
|
||||||
persona = matches[1]
|
persona = matches[1]
|
||||||
msgText = strings.TrimLeft(msgText[len(matches[0]):], " ")
|
msgText = strings.TrimLeft(msgText[len(matches[0]):], " ")
|
||||||
@@ -1139,7 +1158,7 @@ func init() {
|
|||||||
fmt.Fprintf(textView, "%s[-:-:b](%d) <%s>: [-:-:-]\n%s\n",
|
fmt.Fprintf(textView, "%s[-:-:b](%d) <%s>: [-:-:-]\n%s\n",
|
||||||
nl, len(chatBody.Messages), persona, msgText)
|
nl, len(chatBody.Messages), persona, msgText)
|
||||||
textArea.SetText("", true)
|
textArea.SetText("", true)
|
||||||
if scrollToEndEnabled {
|
if cfg.AutoScrollEnabled {
|
||||||
textView.ScrollToEnd()
|
textView.ScrollToEnd()
|
||||||
}
|
}
|
||||||
colorText()
|
colorText()
|
||||||
@@ -1168,9 +1187,10 @@ func init() {
|
|||||||
app.SetFocus(focusSwitcher[currentF])
|
app.SetFocus(focusSwitcher[currentF])
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if isASCII(string(event.Rune())) && !botRespMode {
|
if isASCII(string(event.Rune())) && !botRespMode.Load() {
|
||||||
return event
|
return event
|
||||||
}
|
}
|
||||||
return event
|
return event
|
||||||
})
|
})
|
||||||
|
go updateModelLists()
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user