Compare commits
52 Commits
feat/tab-c
...
master
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
7d51c5d0f3 | ||
|
|
b97cd67d72 | ||
|
|
888c9fec65 | ||
|
|
4f07994bdc | ||
|
|
776fd7a2c4 | ||
|
|
9c6b0dc1fa | ||
|
|
9f51bd3853 | ||
|
|
b386c1181f | ||
|
|
b8e7649e69 | ||
|
|
6664c1a0fc | ||
|
|
e0c3fe554f | ||
|
|
40943ff4d3 | ||
|
|
6c03a1a277 | ||
|
|
27288e2aaa | ||
|
|
1c728ec7a7 | ||
|
|
78059083c2 | ||
|
|
34cd4ac141 | ||
|
|
343366b12d | ||
|
|
978369eeaa | ||
|
|
c39e1c267d | ||
|
|
9af21895c6 | ||
|
|
e3bd6f219f | ||
|
|
ae62c2c8d8 | ||
|
|
04db7c2f01 | ||
|
|
3d889e70b5 | ||
|
|
ef53e9bebe | ||
|
|
a546bfe596 | ||
|
|
23c21f87bb | ||
|
|
850ca103e5 | ||
|
|
b7b5fcbf79 | ||
|
|
1e13c7796d | ||
|
|
9a727b21ad | ||
|
|
beb944c390 | ||
|
|
5844dd1494 | ||
|
|
84c4010213 | ||
|
|
86260e218c | ||
|
|
2c694e2b2b | ||
|
|
66ccb7a732 | ||
|
|
deece322ef | ||
|
|
e7c8fef32d | ||
|
|
eedda0ec4b | ||
|
|
96ffbd5cf5 | ||
|
|
85b11fa9ff | ||
|
|
1675af98d4 | ||
|
|
61a0ddfdfd | ||
|
|
26ab5c59e3 | ||
|
|
35cc8c068f | ||
|
|
27fdec1361 | ||
|
|
76827a71cc | ||
|
|
3a9a7dbe99 | ||
|
|
d3361c13c5 | ||
|
|
7c1a8b0122 |
13
Makefile
13
Makefile
@@ -1,4 +1,4 @@
|
|||||||
.PHONY: setconfig run lint setup-whisper build-whisper download-whisper-model docker-up docker-down docker-logs noextra-run
|
.PHONY: setconfig run lint install-linters setup-whisper build-whisper download-whisper-model docker-up docker-down docker-logs noextra-run installdelve checkdelve
|
||||||
|
|
||||||
run: setconfig
|
run: setconfig
|
||||||
go build -tags extra -o gf-lt && ./gf-lt
|
go build -tags extra -o gf-lt && ./gf-lt
|
||||||
@@ -15,8 +15,17 @@ noextra-run: setconfig
|
|||||||
setconfig:
|
setconfig:
|
||||||
find config.toml &>/dev/null || cp config.example.toml config.toml
|
find config.toml &>/dev/null || cp config.example.toml config.toml
|
||||||
|
|
||||||
|
installdelve:
|
||||||
|
go install github.com/go-delve/delve/cmd/dlv@latest
|
||||||
|
|
||||||
|
checkdelve:
|
||||||
|
which dlv &>/dev/null || installdelve
|
||||||
|
|
||||||
|
install-linters: ## Install additional linters (noblanks)
|
||||||
|
go install github.com/GrailFinder/noblanks-linter/cmd/noblanks@latest
|
||||||
|
|
||||||
lint: ## Run linters. Use make install-linters first.
|
lint: ## Run linters. Use make install-linters first.
|
||||||
golangci-lint run -c .golangci.yml ./...
|
golangci-lint run -c .golangci.yml ./...; noblanks ./...
|
||||||
|
|
||||||
# Whisper STT Setup (in batteries directory)
|
# Whisper STT Setup (in batteries directory)
|
||||||
setup-whisper: build-whisper download-whisper-model
|
setup-whisper: build-whisper download-whisper-model
|
||||||
|
|||||||
@@ -8,7 +8,7 @@ made with use of [tview](https://github.com/rivo/tview)
|
|||||||
- tts/stt (run make commands to get deps);
|
- tts/stt (run make commands to get deps);
|
||||||
- image input;
|
- image input;
|
||||||
- function calls (function calls are implemented natively, to avoid calling outside sources);
|
- function calls (function calls are implemented natively, to avoid calling outside sources);
|
||||||
- [character specific context (unique feature)](char-specific-context.md)
|
- [character specific context (unique feature)](docs/char-specific-context.md)
|
||||||
|
|
||||||
#### how it looks
|
#### how it looks
|
||||||

|

|
||||||
|
|||||||
@@ -110,8 +110,8 @@ func (ag *AgentClient) buildRequest(sysprompt, msg string) ([]byte, error) {
|
|||||||
req := models.NewDSChatReq(*chatBody)
|
req := models.NewDSChatReq(*chatBody)
|
||||||
return json.Marshal(req)
|
return json.Marshal(req)
|
||||||
case isOpenRouter:
|
case isOpenRouter:
|
||||||
// OpenRouter chat
|
// OpenRouter chat - agents don't use reasoning by default
|
||||||
req := models.NewOpenRouterChatReq(*chatBody, defaultProps)
|
req := models.NewOpenRouterChatReq(*chatBody, defaultProps, "")
|
||||||
return json.Marshal(req)
|
return json.Marshal(req)
|
||||||
default:
|
default:
|
||||||
// Assume llama.cpp chat (OpenAI format)
|
// Assume llama.cpp chat (OpenAI format)
|
||||||
@@ -140,7 +140,6 @@ func (ag *AgentClient) LLMRequest(body io.Reader) ([]byte, error) {
|
|||||||
ag.log.Error("failed to read request body", "error", err)
|
ag.log.Error("failed to read request body", "error", err)
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
req, err := http.NewRequest("POST", ag.cfg.CurrentAPI, bytes.NewReader(bodyBytes))
|
req, err := http.NewRequest("POST", ag.cfg.CurrentAPI, bytes.NewReader(bodyBytes))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
ag.log.Error("failed to create request", "error", err)
|
ag.log.Error("failed to create request", "error", err)
|
||||||
@@ -150,22 +149,18 @@ func (ag *AgentClient) LLMRequest(body io.Reader) ([]byte, error) {
|
|||||||
req.Header.Add("Content-Type", "application/json")
|
req.Header.Add("Content-Type", "application/json")
|
||||||
req.Header.Add("Authorization", "Bearer "+ag.getToken())
|
req.Header.Add("Authorization", "Bearer "+ag.getToken())
|
||||||
req.Header.Set("Accept-Encoding", "gzip")
|
req.Header.Set("Accept-Encoding", "gzip")
|
||||||
|
|
||||||
ag.log.Debug("agent LLM request", "url", ag.cfg.CurrentAPI, "body_preview", string(bodyBytes[:min(len(bodyBytes), 500)]))
|
ag.log.Debug("agent LLM request", "url", ag.cfg.CurrentAPI, "body_preview", string(bodyBytes[:min(len(bodyBytes), 500)]))
|
||||||
|
|
||||||
resp, err := httpClient.Do(req)
|
resp, err := httpClient.Do(req)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
ag.log.Error("llamacpp api request failed", "error", err, "url", ag.cfg.CurrentAPI)
|
ag.log.Error("llamacpp api request failed", "error", err, "url", ag.cfg.CurrentAPI)
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
defer resp.Body.Close()
|
defer resp.Body.Close()
|
||||||
|
|
||||||
responseBytes, err := io.ReadAll(resp.Body)
|
responseBytes, err := io.ReadAll(resp.Body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
ag.log.Error("failed to read response", "error", err)
|
ag.log.Error("failed to read response", "error", err)
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
if resp.StatusCode >= 400 {
|
if resp.StatusCode >= 400 {
|
||||||
ag.log.Error("agent LLM request failed", "status", resp.StatusCode, "response", string(responseBytes[:min(len(responseBytes), 1000)]))
|
ag.log.Error("agent LLM request failed", "status", resp.StatusCode, "response", string(responseBytes[:min(len(responseBytes), 1000)]))
|
||||||
return responseBytes, fmt.Errorf("HTTP %d: %s", resp.StatusCode, string(responseBytes[:min(len(responseBytes), 200)]))
|
return responseBytes, fmt.Errorf("HTTP %d: %s", resp.StatusCode, string(responseBytes[:min(len(responseBytes), 200)]))
|
||||||
@@ -178,7 +173,6 @@ func (ag *AgentClient) LLMRequest(body io.Reader) ([]byte, error) {
|
|||||||
// Return raw response as fallback
|
// Return raw response as fallback
|
||||||
return responseBytes, nil
|
return responseBytes, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
return []byte(text), nil
|
return []byte(text), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
343
bot.go
343
bot.go
@@ -23,8 +23,6 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/neurosnap/sentences/english"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
@@ -46,6 +44,7 @@ var (
|
|||||||
ragger *rag.RAG
|
ragger *rag.RAG
|
||||||
chunkParser ChunkParser
|
chunkParser ChunkParser
|
||||||
lastToolCall *models.FuncCall
|
lastToolCall *models.FuncCall
|
||||||
|
lastRespStats *models.ResponseStats
|
||||||
//nolint:unused // TTS_ENABLED conditionally uses this
|
//nolint:unused // TTS_ENABLED conditionally uses this
|
||||||
orator Orator
|
orator Orator
|
||||||
asr STT
|
asr STT
|
||||||
@@ -118,7 +117,7 @@ func processMessageTag(msg *models.RoleMsg) *models.RoleMsg {
|
|||||||
}
|
}
|
||||||
// If KnownTo already set, assume tag already processed (content cleaned).
|
// If KnownTo already set, assume tag already processed (content cleaned).
|
||||||
// However, we still check for new tags (maybe added later).
|
// However, we still check for new tags (maybe added later).
|
||||||
knownTo := parseKnownToTag(msg.Content)
|
knownTo := parseKnownToTag(msg.GetText())
|
||||||
// If tag found, replace KnownTo with new list (merge with existing?)
|
// If tag found, replace KnownTo with new list (merge with existing?)
|
||||||
// For simplicity, if knownTo is not nil, replace.
|
// For simplicity, if knownTo is not nil, replace.
|
||||||
if knownTo == nil {
|
if knownTo == nil {
|
||||||
@@ -403,6 +402,30 @@ func fetchLCPModels() ([]string, error) {
|
|||||||
return localModels, nil
|
return localModels, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// fetchLCPModelsWithLoadStatus returns models with "(loaded)" indicator for loaded models
|
||||||
|
func fetchLCPModelsWithLoadStatus() ([]string, error) {
|
||||||
|
models, err := fetchLCPModelsWithStatus()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
result := make([]string, 0, len(models.Data))
|
||||||
|
li := 0 // loaded index
|
||||||
|
for i, m := range models.Data {
|
||||||
|
modelName := m.ID
|
||||||
|
if m.Status.Value == "loaded" {
|
||||||
|
modelName = "(loaded) " + modelName
|
||||||
|
li = i
|
||||||
|
}
|
||||||
|
result = append(result, modelName)
|
||||||
|
}
|
||||||
|
if li == 0 {
|
||||||
|
return result, nil // no loaded models
|
||||||
|
}
|
||||||
|
loadedModel := result[li]
|
||||||
|
result = append(result[:li], result[li+1:]...)
|
||||||
|
return slices.Concat([]string{loadedModel}, result), nil
|
||||||
|
}
|
||||||
|
|
||||||
// fetchLCPModelsWithStatus returns the full LCPModels struct including status information.
|
// fetchLCPModelsWithStatus returns the full LCPModels struct including status information.
|
||||||
func fetchLCPModelsWithStatus() (*models.LCPModels, error) {
|
func fetchLCPModelsWithStatus() (*models.LCPModels, error) {
|
||||||
resp, err := http.Get(cfg.FetchModelNameAPI)
|
resp, err := http.Get(cfg.FetchModelNameAPI)
|
||||||
@@ -456,6 +479,7 @@ func monitorModelLoad(modelID string) {
|
|||||||
if err := notifyUser("model loaded", "Model "+modelID+" is now loaded and ready."); err != nil {
|
if err := notifyUser("model loaded", "Model "+modelID+" is now loaded and ready."); err != nil {
|
||||||
logger.Debug("failed to notify user", "error", err)
|
logger.Debug("failed to notify user", "error", err)
|
||||||
}
|
}
|
||||||
|
refreshChatDisplay()
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -466,30 +490,28 @@ func monitorModelLoad(modelID string) {
|
|||||||
// extractDetailedErrorFromBytes extracts detailed error information from response body bytes
|
// extractDetailedErrorFromBytes extracts detailed error information from response body bytes
|
||||||
func extractDetailedErrorFromBytes(body []byte, statusCode int) string {
|
func extractDetailedErrorFromBytes(body []byte, statusCode int) string {
|
||||||
// Try to parse as JSON to extract detailed error information
|
// Try to parse as JSON to extract detailed error information
|
||||||
var errorResponse map[string]interface{}
|
var errorResponse map[string]any
|
||||||
if err := json.Unmarshal(body, &errorResponse); err == nil {
|
if err := json.Unmarshal(body, &errorResponse); err == nil {
|
||||||
// Check if it's an error response with detailed information
|
// Check if it's an error response with detailed information
|
||||||
if errorData, ok := errorResponse["error"]; ok {
|
if errorData, ok := errorResponse["error"]; ok {
|
||||||
if errorMap, ok := errorData.(map[string]interface{}); ok {
|
if errorMap, ok := errorData.(map[string]any); ok {
|
||||||
var errorMsg string
|
var errorMsg string
|
||||||
if msg, ok := errorMap["message"]; ok {
|
if msg, ok := errorMap["message"]; ok {
|
||||||
errorMsg = fmt.Sprintf("%v", msg)
|
errorMsg = fmt.Sprintf("%v", msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
var details []string
|
var details []string
|
||||||
if code, ok := errorMap["code"]; ok {
|
if code, ok := errorMap["code"]; ok {
|
||||||
details = append(details, fmt.Sprintf("Code: %v", code))
|
details = append(details, fmt.Sprintf("Code: %v", code))
|
||||||
}
|
}
|
||||||
|
|
||||||
if metadata, ok := errorMap["metadata"]; ok {
|
if metadata, ok := errorMap["metadata"]; ok {
|
||||||
// Handle metadata which might contain raw error details
|
// Handle metadata which might contain raw error details
|
||||||
if metadataMap, ok := metadata.(map[string]interface{}); ok {
|
if metadataMap, ok := metadata.(map[string]any); ok {
|
||||||
if raw, ok := metadataMap["raw"]; ok {
|
if raw, ok := metadataMap["raw"]; ok {
|
||||||
// Parse the raw error string if it's JSON
|
// Parse the raw error string if it's JSON
|
||||||
var rawError map[string]interface{}
|
var rawError map[string]any
|
||||||
if rawStr, ok := raw.(string); ok && json.Unmarshal([]byte(rawStr), &rawError) == nil {
|
if rawStr, ok := raw.(string); ok && json.Unmarshal([]byte(rawStr), &rawError) == nil {
|
||||||
if rawErrorData, ok := rawError["error"]; ok {
|
if rawErrorData, ok := rawError["error"]; ok {
|
||||||
if rawErrorMap, ok := rawErrorData.(map[string]interface{}); ok {
|
if rawErrorMap, ok := rawErrorData.(map[string]any); ok {
|
||||||
if rawMsg, ok := rawErrorMap["message"]; ok {
|
if rawMsg, ok := rawErrorMap["message"]; ok {
|
||||||
return fmt.Sprintf("API Error: %s", rawMsg)
|
return fmt.Sprintf("API Error: %s", rawMsg)
|
||||||
}
|
}
|
||||||
@@ -500,20 +522,30 @@ func extractDetailedErrorFromBytes(body []byte, statusCode int) string {
|
|||||||
}
|
}
|
||||||
details = append(details, fmt.Sprintf("Metadata: %v", metadata))
|
details = append(details, fmt.Sprintf("Metadata: %v", metadata))
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(details) > 0 {
|
if len(details) > 0 {
|
||||||
return fmt.Sprintf("API Error: %s (%s)", errorMsg, strings.Join(details, ", "))
|
return fmt.Sprintf("API Error: %s (%s)", errorMsg, strings.Join(details, ", "))
|
||||||
}
|
}
|
||||||
|
|
||||||
return "API Error: " + errorMsg
|
return "API Error: " + errorMsg
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// If not a structured error response, return the raw body with status
|
// If not a structured error response, return the raw body with status
|
||||||
return fmt.Sprintf("HTTP Status: %d, Response Body: %s", statusCode, string(body))
|
return fmt.Sprintf("HTTP Status: %d, Response Body: %s", statusCode, string(body))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func finalizeRespStats(tokenCount int, startTime time.Time) {
|
||||||
|
duration := time.Since(startTime).Seconds()
|
||||||
|
var tps float64
|
||||||
|
if duration > 0 {
|
||||||
|
tps = float64(tokenCount) / duration
|
||||||
|
}
|
||||||
|
lastRespStats = &models.ResponseStats{
|
||||||
|
Tokens: tokenCount,
|
||||||
|
Duration: duration,
|
||||||
|
TokensPerSec: tps,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// sendMsgToLLM expects streaming resp
|
// sendMsgToLLM expects streaming resp
|
||||||
func sendMsgToLLM(body io.Reader) {
|
func sendMsgToLLM(body io.Reader) {
|
||||||
choseChunkParser()
|
choseChunkParser()
|
||||||
@@ -542,7 +574,6 @@ func sendMsgToLLM(body io.Reader) {
|
|||||||
streamDone <- true
|
streamDone <- true
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check if the initial response is an error before starting to stream
|
// Check if the initial response is an error before starting to stream
|
||||||
if resp.StatusCode >= 400 {
|
if resp.StatusCode >= 400 {
|
||||||
// Read the response body to get detailed error information
|
// Read the response body to get detailed error information
|
||||||
@@ -557,7 +588,6 @@ func sendMsgToLLM(body io.Reader) {
|
|||||||
streamDone <- true
|
streamDone <- true
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
// Parse the error response for detailed information
|
// Parse the error response for detailed information
|
||||||
detailedError := extractDetailedErrorFromBytes(bodyBytes, resp.StatusCode)
|
detailedError := extractDetailedErrorFromBytes(bodyBytes, resp.StatusCode)
|
||||||
logger.Error("API returned error status", "status_code", resp.StatusCode, "detailed_error", detailedError)
|
logger.Error("API returned error status", "status_code", resp.StatusCode, "detailed_error", detailedError)
|
||||||
@@ -568,10 +598,17 @@ func sendMsgToLLM(body io.Reader) {
|
|||||||
streamDone <- true
|
streamDone <- true
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
//
|
||||||
defer resp.Body.Close()
|
defer resp.Body.Close()
|
||||||
reader := bufio.NewReader(resp.Body)
|
reader := bufio.NewReader(resp.Body)
|
||||||
counter := uint32(0)
|
counter := uint32(0)
|
||||||
|
tokenCount := 0
|
||||||
|
startTime := time.Now()
|
||||||
|
hasReasoning := false
|
||||||
|
reasoningSent := false
|
||||||
|
defer func() {
|
||||||
|
finalizeRespStats(tokenCount, startTime)
|
||||||
|
}()
|
||||||
for {
|
for {
|
||||||
var (
|
var (
|
||||||
answerText string
|
answerText string
|
||||||
@@ -644,10 +681,16 @@ func sendMsgToLLM(body io.Reader) {
|
|||||||
// break
|
// break
|
||||||
// }
|
// }
|
||||||
if chunk.Finished {
|
if chunk.Finished {
|
||||||
|
// Close the thinking block if we were streaming reasoning and haven't closed it yet
|
||||||
|
if hasReasoning && !reasoningSent {
|
||||||
|
chunkChan <- "</think>"
|
||||||
|
tokenCount++
|
||||||
|
}
|
||||||
if chunk.Chunk != "" {
|
if chunk.Chunk != "" {
|
||||||
logger.Warn("text inside of finish llmchunk", "chunk", chunk, "counter", counter)
|
logger.Warn("text inside of finish llmchunk", "chunk", chunk, "counter", counter)
|
||||||
answerText = strings.ReplaceAll(chunk.Chunk, "\n\n", "\n")
|
answerText = strings.ReplaceAll(chunk.Chunk, "\n\n", "\n")
|
||||||
chunkChan <- answerText
|
chunkChan <- answerText
|
||||||
|
tokenCount++
|
||||||
}
|
}
|
||||||
streamDone <- true
|
streamDone <- true
|
||||||
break
|
break
|
||||||
@@ -655,6 +698,28 @@ func sendMsgToLLM(body io.Reader) {
|
|||||||
if counter == 0 {
|
if counter == 0 {
|
||||||
chunk.Chunk = strings.TrimPrefix(chunk.Chunk, " ")
|
chunk.Chunk = strings.TrimPrefix(chunk.Chunk, " ")
|
||||||
}
|
}
|
||||||
|
// Handle reasoning chunks - stream them immediately as they arrive
|
||||||
|
if chunk.Reasoning != "" && !reasoningSent {
|
||||||
|
if !hasReasoning {
|
||||||
|
// First reasoning chunk - send opening tag
|
||||||
|
chunkChan <- "<think>"
|
||||||
|
tokenCount++
|
||||||
|
hasReasoning = true
|
||||||
|
}
|
||||||
|
// Stream reasoning content immediately
|
||||||
|
answerText = strings.ReplaceAll(chunk.Reasoning, "\n\n", "\n")
|
||||||
|
if answerText != "" {
|
||||||
|
chunkChan <- answerText
|
||||||
|
tokenCount++
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// When we get content and have been streaming reasoning, close the thinking block
|
||||||
|
if chunk.Chunk != "" && hasReasoning && !reasoningSent {
|
||||||
|
// Close the thinking block before sending actual content
|
||||||
|
chunkChan <- "</think>"
|
||||||
|
tokenCount++
|
||||||
|
reasoningSent = true
|
||||||
|
}
|
||||||
// bot sends way too many \n
|
// bot sends way too many \n
|
||||||
answerText = strings.ReplaceAll(chunk.Chunk, "\n\n", "\n")
|
answerText = strings.ReplaceAll(chunk.Chunk, "\n\n", "\n")
|
||||||
// Accumulate text to check for stop strings that might span across chunks
|
// Accumulate text to check for stop strings that might span across chunks
|
||||||
@@ -664,8 +729,12 @@ func sendMsgToLLM(body io.Reader) {
|
|||||||
slices.Contains(stopStrings, answerText) {
|
slices.Contains(stopStrings, answerText) {
|
||||||
logger.Debug("stop string detected on client side for completion endpoint", "stop_string", answerText)
|
logger.Debug("stop string detected on client side for completion endpoint", "stop_string", answerText)
|
||||||
streamDone <- true
|
streamDone <- true
|
||||||
|
break
|
||||||
}
|
}
|
||||||
|
if answerText != "" {
|
||||||
chunkChan <- answerText
|
chunkChan <- answerText
|
||||||
|
tokenCount++
|
||||||
|
}
|
||||||
openAIToolChan <- chunk.ToolChunk
|
openAIToolChan <- chunk.ToolChunk
|
||||||
if chunk.FuncName != "" {
|
if chunk.FuncName != "" {
|
||||||
lastToolCall.Name = chunk.FuncName
|
lastToolCall.Name = chunk.FuncName
|
||||||
@@ -682,68 +751,6 @@ func sendMsgToLLM(body io.Reader) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func chatRagUse(qText string) (string, error) {
|
|
||||||
logger.Debug("Starting RAG query", "original_query", qText)
|
|
||||||
tokenizer, err := english.NewSentenceTokenizer(nil)
|
|
||||||
if err != nil {
|
|
||||||
logger.Error("failed to create sentence tokenizer", "error", err)
|
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
// this where llm should find the questions in text and ask them
|
|
||||||
questionsS := tokenizer.Tokenize(qText)
|
|
||||||
questions := make([]string, len(questionsS))
|
|
||||||
for i, q := range questionsS {
|
|
||||||
questions[i] = q.Text
|
|
||||||
logger.Debug("RAG question extracted", "index", i, "question", q.Text)
|
|
||||||
}
|
|
||||||
|
|
||||||
if len(questions) == 0 {
|
|
||||||
logger.Warn("No questions extracted from query text", "query", qText)
|
|
||||||
return "No related results from RAG vector storage.", nil
|
|
||||||
}
|
|
||||||
|
|
||||||
respVecs := []models.VectorRow{}
|
|
||||||
for i, q := range questions {
|
|
||||||
logger.Debug("Processing RAG question", "index", i, "question", q)
|
|
||||||
emb, err := ragger.LineToVector(q)
|
|
||||||
if err != nil {
|
|
||||||
logger.Error("failed to get embeddings for RAG", "error", err, "index", i, "question", q)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
logger.Debug("Got embeddings for question", "index", i, "question_len", len(q), "embedding_len", len(emb))
|
|
||||||
|
|
||||||
// Create EmbeddingResp struct for the search
|
|
||||||
embeddingResp := &models.EmbeddingResp{
|
|
||||||
Embedding: emb,
|
|
||||||
Index: 0, // Not used in search but required for the struct
|
|
||||||
}
|
|
||||||
vecs, err := ragger.SearchEmb(embeddingResp)
|
|
||||||
if err != nil {
|
|
||||||
logger.Error("failed to query embeddings in RAG", "error", err, "index", i, "question", q)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
logger.Debug("RAG search returned vectors", "index", i, "question", q, "vector_count", len(vecs))
|
|
||||||
respVecs = append(respVecs, vecs...)
|
|
||||||
}
|
|
||||||
|
|
||||||
// get raw text
|
|
||||||
resps := []string{}
|
|
||||||
logger.Debug("RAG query final results", "total_vecs_found", len(respVecs))
|
|
||||||
for _, rv := range respVecs {
|
|
||||||
resps = append(resps, rv.RawText)
|
|
||||||
logger.Debug("RAG result", "slug", rv.Slug, "filename", rv.FileName, "raw_text_len", len(rv.RawText))
|
|
||||||
}
|
|
||||||
|
|
||||||
if len(resps) == 0 {
|
|
||||||
logger.Info("No RAG results found for query", "original_query", qText, "question_count", len(questions))
|
|
||||||
return "No related results from RAG vector storage.", nil
|
|
||||||
}
|
|
||||||
|
|
||||||
result := strings.Join(resps, "\n")
|
|
||||||
logger.Debug("RAG query completed", "result_len", len(result), "response_count", len(resps))
|
|
||||||
return result, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func roleToIcon(role string) string {
|
func roleToIcon(role string) string {
|
||||||
return "<" + role + ">: "
|
return "<" + role + ">: "
|
||||||
}
|
}
|
||||||
@@ -761,13 +768,46 @@ func chatWatcher(ctx context.Context) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func chatRound(r *models.ChatRoundReq) error {
|
// inpired by https://github.com/rivo/tview/issues/225
|
||||||
botRespMode = true
|
func showSpinner() {
|
||||||
|
spinners := []string{"⠋", "⠙", "⠹", "⠸", "⠼", "⠴", "⠦", "⠧", "⠇", "⠏"}
|
||||||
|
var i int
|
||||||
botPersona := cfg.AssistantRole
|
botPersona := cfg.AssistantRole
|
||||||
if cfg.WriteNextMsgAsCompletionAgent != "" {
|
if cfg.WriteNextMsgAsCompletionAgent != "" {
|
||||||
botPersona = cfg.WriteNextMsgAsCompletionAgent
|
botPersona = cfg.WriteNextMsgAsCompletionAgent
|
||||||
}
|
}
|
||||||
defer func() { botRespMode = false }()
|
for botRespMode || toolRunningMode {
|
||||||
|
time.Sleep(100 * time.Millisecond)
|
||||||
|
spin := i % len(spinners)
|
||||||
|
app.QueueUpdateDraw(func() {
|
||||||
|
switch {
|
||||||
|
case toolRunningMode:
|
||||||
|
textArea.SetTitle(spinners[spin] + " tool")
|
||||||
|
case botRespMode:
|
||||||
|
textArea.SetTitle(spinners[spin] + " " + botPersona)
|
||||||
|
default:
|
||||||
|
textArea.SetTitle(spinners[spin] + " input")
|
||||||
|
}
|
||||||
|
})
|
||||||
|
i++
|
||||||
|
}
|
||||||
|
app.QueueUpdateDraw(func() {
|
||||||
|
textArea.SetTitle("input")
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func chatRound(r *models.ChatRoundReq) error {
|
||||||
|
botRespMode = true
|
||||||
|
go showSpinner()
|
||||||
|
updateStatusLine()
|
||||||
|
botPersona := cfg.AssistantRole
|
||||||
|
if cfg.WriteNextMsgAsCompletionAgent != "" {
|
||||||
|
botPersona = cfg.WriteNextMsgAsCompletionAgent
|
||||||
|
}
|
||||||
|
defer func() {
|
||||||
|
botRespMode = false
|
||||||
|
ClearImageAttachment()
|
||||||
|
}()
|
||||||
// check that there is a model set to use if is not local
|
// check that there is a model set to use if is not local
|
||||||
choseChunkParser()
|
choseChunkParser()
|
||||||
reader, err := chunkParser.FormMsg(r.UserMsg, r.Role, r.Resume)
|
reader, err := chunkParser.FormMsg(r.UserMsg, r.Role, r.Resume)
|
||||||
@@ -780,23 +820,79 @@ func chatRound(r *models.ChatRoundReq) error {
|
|||||||
}
|
}
|
||||||
go sendMsgToLLM(reader)
|
go sendMsgToLLM(reader)
|
||||||
logger.Debug("looking at vars in chatRound", "msg", r.UserMsg, "regen", r.Regen, "resume", r.Resume)
|
logger.Debug("looking at vars in chatRound", "msg", r.UserMsg, "regen", r.Regen, "resume", r.Resume)
|
||||||
|
msgIdx := len(chatBody.Messages)
|
||||||
if !r.Resume {
|
if !r.Resume {
|
||||||
fmt.Fprintf(textView, "\n[-:-:b](%d) ", len(chatBody.Messages))
|
// Add empty message to chatBody immediately so it persists during Alt+T toggle
|
||||||
fmt.Fprint(textView, roleToIcon(botPersona))
|
chatBody.Messages = append(chatBody.Messages, models.RoleMsg{
|
||||||
fmt.Fprint(textView, "[-:-:-]\n")
|
Role: botPersona, Content: "",
|
||||||
if cfg.ThinkUse && !strings.Contains(cfg.CurrentAPI, "v1") {
|
})
|
||||||
// fmt.Fprint(textView, "<think>")
|
nl := "\n\n"
|
||||||
chunkChan <- "<think>"
|
prevText := textView.GetText(true)
|
||||||
|
if strings.HasSuffix(prevText, nl) {
|
||||||
|
nl = ""
|
||||||
|
} else if strings.HasSuffix(prevText, "\n") {
|
||||||
|
nl = "\n"
|
||||||
}
|
}
|
||||||
|
fmt.Fprintf(textView, "%s[-:-:b](%d) %s[-:-:-]\n", nl, msgIdx, roleToIcon(botPersona))
|
||||||
|
} else {
|
||||||
|
msgIdx = len(chatBody.Messages) - 1
|
||||||
}
|
}
|
||||||
respText := strings.Builder{}
|
respText := strings.Builder{}
|
||||||
toolResp := strings.Builder{}
|
toolResp := strings.Builder{}
|
||||||
|
// Variables for handling thinking blocks during streaming
|
||||||
|
inThinkingBlock := false
|
||||||
|
thinkingBuffer := strings.Builder{}
|
||||||
|
justExitedThinkingCollapsed := false
|
||||||
out:
|
out:
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case chunk := <-chunkChan:
|
case chunk := <-chunkChan:
|
||||||
|
// Handle thinking blocks during streaming
|
||||||
|
if strings.HasPrefix(chunk, "<think>") && !inThinkingBlock {
|
||||||
|
// Start of thinking block
|
||||||
|
inThinkingBlock = true
|
||||||
|
thinkingBuffer.Reset()
|
||||||
|
thinkingBuffer.WriteString(chunk)
|
||||||
|
if thinkingCollapsed {
|
||||||
|
// Show placeholder immediately when thinking starts in collapsed mode
|
||||||
|
fmt.Fprint(textView, "[yellow::i][thinking... (press Alt+T to expand)][-:-:-]")
|
||||||
|
if scrollToEndEnabled {
|
||||||
|
textView.ScrollToEnd()
|
||||||
|
}
|
||||||
|
respText.WriteString(chunk)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
} else if inThinkingBlock {
|
||||||
|
thinkingBuffer.WriteString(chunk)
|
||||||
|
if strings.Contains(chunk, "</think>") {
|
||||||
|
// End of thinking block
|
||||||
|
inThinkingBlock = false
|
||||||
|
if thinkingCollapsed {
|
||||||
|
// Thinking already displayed as placeholder, just update respText
|
||||||
|
respText.WriteString(chunk)
|
||||||
|
justExitedThinkingCollapsed = true
|
||||||
|
if scrollToEndEnabled {
|
||||||
|
textView.ScrollToEnd()
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// If not collapsed, fall through to normal display
|
||||||
|
} else if thinkingCollapsed {
|
||||||
|
// Still in thinking block and collapsed - just buffer, don't display
|
||||||
|
respText.WriteString(chunk)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// If not collapsed, fall through to normal display
|
||||||
|
}
|
||||||
|
// Add spacing after collapsed thinking block before real response
|
||||||
|
if justExitedThinkingCollapsed {
|
||||||
|
chunk = "\n\n" + chunk
|
||||||
|
justExitedThinkingCollapsed = false
|
||||||
|
}
|
||||||
fmt.Fprint(textView, chunk)
|
fmt.Fprint(textView, chunk)
|
||||||
respText.WriteString(chunk)
|
respText.WriteString(chunk)
|
||||||
|
// Update the message in chatBody.Messages so it persists during Alt+T
|
||||||
|
chatBody.Messages[msgIdx].Content = respText.String()
|
||||||
if scrollToEndEnabled {
|
if scrollToEndEnabled {
|
||||||
textView.ScrollToEnd()
|
textView.ScrollToEnd()
|
||||||
}
|
}
|
||||||
@@ -811,7 +907,6 @@ out:
|
|||||||
textView.ScrollToEnd()
|
textView.ScrollToEnd()
|
||||||
}
|
}
|
||||||
case <-streamDone:
|
case <-streamDone:
|
||||||
// drain any remaining chunks from chunkChan before exiting
|
|
||||||
for len(chunkChan) > 0 {
|
for len(chunkChan) > 0 {
|
||||||
chunk := <-chunkChan
|
chunk := <-chunkChan
|
||||||
fmt.Fprint(textView, chunk)
|
fmt.Fprint(textView, chunk)
|
||||||
@@ -820,34 +915,41 @@ out:
|
|||||||
textView.ScrollToEnd()
|
textView.ScrollToEnd()
|
||||||
}
|
}
|
||||||
if cfg.TTS_ENABLED {
|
if cfg.TTS_ENABLED {
|
||||||
// Send chunk to audio stream handler
|
|
||||||
TTSTextChan <- chunk
|
TTSTextChan <- chunk
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if cfg.TTS_ENABLED {
|
if cfg.TTS_ENABLED {
|
||||||
// msg is done; flush it down
|
|
||||||
TTSFlushChan <- true
|
TTSFlushChan <- true
|
||||||
}
|
}
|
||||||
break out
|
break out
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
var msgStats *models.ResponseStats
|
||||||
|
if lastRespStats != nil {
|
||||||
|
msgStats = &models.ResponseStats{
|
||||||
|
Tokens: lastRespStats.Tokens,
|
||||||
|
Duration: lastRespStats.Duration,
|
||||||
|
TokensPerSec: lastRespStats.TokensPerSec,
|
||||||
|
}
|
||||||
|
lastRespStats = nil
|
||||||
|
}
|
||||||
botRespMode = false
|
botRespMode = false
|
||||||
// numbers in chatbody and displayed must be the same
|
|
||||||
if r.Resume {
|
if r.Resume {
|
||||||
chatBody.Messages[len(chatBody.Messages)-1].Content += respText.String()
|
chatBody.Messages[len(chatBody.Messages)-1].Content += respText.String()
|
||||||
// lastM.Content = lastM.Content + respText.String()
|
|
||||||
// Process the updated message to check for known_to tags in resumed response
|
|
||||||
updatedMsg := chatBody.Messages[len(chatBody.Messages)-1]
|
updatedMsg := chatBody.Messages[len(chatBody.Messages)-1]
|
||||||
processedMsg := processMessageTag(&updatedMsg)
|
processedMsg := processMessageTag(&updatedMsg)
|
||||||
chatBody.Messages[len(chatBody.Messages)-1] = *processedMsg
|
chatBody.Messages[len(chatBody.Messages)-1] = *processedMsg
|
||||||
} else {
|
if msgStats != nil && chatBody.Messages[len(chatBody.Messages)-1].Role != cfg.ToolRole {
|
||||||
newMsg := models.RoleMsg{
|
chatBody.Messages[len(chatBody.Messages)-1].Stats = msgStats
|
||||||
Role: botPersona, Content: respText.String(),
|
|
||||||
}
|
}
|
||||||
// Process the new message to check for known_to tags in LLM response
|
} else {
|
||||||
newMsg = *processMessageTag(&newMsg)
|
chatBody.Messages[msgIdx].Content = respText.String()
|
||||||
chatBody.Messages = append(chatBody.Messages, newMsg)
|
processedMsg := processMessageTag(&chatBody.Messages[msgIdx])
|
||||||
stopTTSIfNotForUser(&newMsg)
|
chatBody.Messages[msgIdx] = *processedMsg
|
||||||
|
if msgStats != nil && chatBody.Messages[msgIdx].Role != cfg.ToolRole {
|
||||||
|
chatBody.Messages[msgIdx].Stats = msgStats
|
||||||
|
}
|
||||||
|
stopTTSIfNotForUser(&chatBody.Messages[msgIdx])
|
||||||
}
|
}
|
||||||
cleanChatBody()
|
cleanChatBody()
|
||||||
refreshChatDisplay()
|
refreshChatDisplay()
|
||||||
@@ -1068,9 +1170,13 @@ func findCall(msg, toolCall string) bool {
|
|||||||
chatRoundChan <- crr
|
chatRoundChan <- crr
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
// Show tool call progress indicator before execution
|
||||||
|
fmt.Fprintf(textView, "\n[yellow::i][tool: %s...][-:-:-]", fc.Name)
|
||||||
|
toolRunningMode = true
|
||||||
resp := callToolWithAgent(fc.Name, fc.Args)
|
resp := callToolWithAgent(fc.Name, fc.Args)
|
||||||
|
toolRunningMode = false
|
||||||
toolMsg := string(resp) // Remove the "tool response: " prefix and %+v formatting
|
toolMsg := string(resp) // Remove the "tool response: " prefix and %+v formatting
|
||||||
logger.Info("llm used tool call", "tool_resp", toolMsg, "tool_attrs", fc)
|
logger.Info("llm used a tool call", "tool_name", fc.Name, "too_args", fc.Args, "id", fc.ID, "tool_resp", toolMsg)
|
||||||
fmt.Fprintf(textView, "%s[-:-:b](%d) <%s>: [-:-:-]\n%s\n",
|
fmt.Fprintf(textView, "%s[-:-:b](%d) <%s>: [-:-:-]\n%s\n",
|
||||||
"\n\n", len(chatBody.Messages), cfg.ToolRole, toolMsg)
|
"\n\n", len(chatBody.Messages), cfg.ToolRole, toolMsg)
|
||||||
// Create tool response message with the proper tool_call_id
|
// Create tool response message with the proper tool_call_id
|
||||||
@@ -1107,13 +1213,29 @@ func chatToTextSlice(messages []models.RoleMsg, showSys bool) []string {
|
|||||||
func chatToText(messages []models.RoleMsg, showSys bool) string {
|
func chatToText(messages []models.RoleMsg, showSys bool) string {
|
||||||
s := chatToTextSlice(messages, showSys)
|
s := chatToTextSlice(messages, showSys)
|
||||||
text := strings.Join(s, "\n")
|
text := strings.Join(s, "\n")
|
||||||
|
|
||||||
// Collapse thinking blocks if enabled
|
// Collapse thinking blocks if enabled
|
||||||
if thinkingCollapsed {
|
if thinkingCollapsed {
|
||||||
placeholder := "[yellow::i][thinking... (press Alt+T to expand)][-:-:-]"
|
text = thinkRE.ReplaceAllStringFunc(text, func(match string) string {
|
||||||
text = thinkRE.ReplaceAllString(text, placeholder)
|
// Extract content between <think> and </think>
|
||||||
|
start := len("<think>")
|
||||||
|
end := len(match) - len("</think>")
|
||||||
|
if start < end && start < len(match) {
|
||||||
|
content := match[start:end]
|
||||||
|
return fmt.Sprintf("[yellow::i][thinking... (%d chars) (press Alt+T to expand)][-:-:-]", len(content))
|
||||||
|
}
|
||||||
|
return "[yellow::i][thinking... (press Alt+T to expand)][-:-:-]"
|
||||||
|
})
|
||||||
|
// Handle incomplete thinking blocks (during streaming when </think> hasn't arrived yet)
|
||||||
|
if strings.Contains(text, "<think>") && !strings.Contains(text, "</think>") {
|
||||||
|
// Find the incomplete thinking block and replace it
|
||||||
|
startIdx := strings.Index(text, "<think>")
|
||||||
|
if startIdx != -1 {
|
||||||
|
content := text[startIdx+len("<think>"):]
|
||||||
|
placeholder := fmt.Sprintf("[yellow::i][thinking... (%d chars) (press Alt+T to expand)][-:-:-]", len(content))
|
||||||
|
text = text[:startIdx] + placeholder
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return text
|
return text
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1124,12 +1246,9 @@ func removeThinking(chatBody *models.ChatBody) {
|
|||||||
if msg.Role == cfg.ToolRole {
|
if msg.Role == cfg.ToolRole {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
// find thinking and remove it
|
// find thinking and remove it - use SetText to preserve ContentParts
|
||||||
rm := models.RoleMsg{
|
msg.SetText(thinkRE.ReplaceAllString(msg.GetText(), ""))
|
||||||
Role: msg.Role,
|
msgs = append(msgs, msg)
|
||||||
Content: thinkRE.ReplaceAllString(msg.Content, ""),
|
|
||||||
}
|
|
||||||
msgs = append(msgs, rm)
|
|
||||||
}
|
}
|
||||||
chatBody.Messages = msgs
|
chatBody.Messages = msgs
|
||||||
}
|
}
|
||||||
@@ -1250,7 +1369,7 @@ func init() {
|
|||||||
var err error
|
var err error
|
||||||
cfg, err = config.LoadConfig("config.toml")
|
cfg, err = config.LoadConfig("config.toml")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
fmt.Println("failed to load config.toml")
|
fmt.Println("failed to load config.toml", err)
|
||||||
cancel()
|
cancel()
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
return
|
return
|
||||||
|
|||||||
34
bot_test.go
34
bot_test.go
@@ -1,12 +1,10 @@
|
|||||||
package main
|
package main
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"gf-lt/config"
|
"gf-lt/config"
|
||||||
"gf-lt/models"
|
"gf-lt/models"
|
||||||
"reflect"
|
"reflect"
|
||||||
"testing"
|
"testing"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestConsolidateConsecutiveAssistantMessages(t *testing.T) {
|
func TestConsolidateConsecutiveAssistantMessages(t *testing.T) {
|
||||||
// Mock config for testing
|
// Mock config for testing
|
||||||
testCfg := &config.Config{
|
testCfg := &config.Config{
|
||||||
@@ -14,7 +12,6 @@ func TestConsolidateConsecutiveAssistantMessages(t *testing.T) {
|
|||||||
WriteNextMsgAsCompletionAgent: "",
|
WriteNextMsgAsCompletionAgent: "",
|
||||||
}
|
}
|
||||||
cfg = testCfg
|
cfg = testCfg
|
||||||
|
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
name string
|
name string
|
||||||
input []models.RoleMsg
|
input []models.RoleMsg
|
||||||
@@ -114,38 +111,31 @@ func TestConsolidateConsecutiveAssistantMessages(t *testing.T) {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tt := range tests {
|
for _, tt := range tests {
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
result := consolidateAssistantMessages(tt.input)
|
result := consolidateAssistantMessages(tt.input)
|
||||||
|
|
||||||
if len(result) != len(tt.expected) {
|
if len(result) != len(tt.expected) {
|
||||||
t.Errorf("Expected %d messages, got %d", len(tt.expected), len(result))
|
t.Errorf("Expected %d messages, got %d", len(tt.expected), len(result))
|
||||||
t.Logf("Result: %+v", result)
|
t.Logf("Result: %+v", result)
|
||||||
t.Logf("Expected: %+v", tt.expected)
|
t.Logf("Expected: %+v", tt.expected)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
for i, expectedMsg := range tt.expected {
|
for i, expectedMsg := range tt.expected {
|
||||||
if i >= len(result) {
|
if i >= len(result) {
|
||||||
t.Errorf("Result has fewer messages than expected at index %d", i)
|
t.Errorf("Result has fewer messages than expected at index %d", i)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
actualMsg := result[i]
|
actualMsg := result[i]
|
||||||
if actualMsg.Role != expectedMsg.Role {
|
if actualMsg.Role != expectedMsg.Role {
|
||||||
t.Errorf("Message %d: expected role '%s', got '%s'", i, expectedMsg.Role, actualMsg.Role)
|
t.Errorf("Message %d: expected role '%s', got '%s'", i, expectedMsg.Role, actualMsg.Role)
|
||||||
}
|
}
|
||||||
|
|
||||||
if actualMsg.Content != expectedMsg.Content {
|
if actualMsg.Content != expectedMsg.Content {
|
||||||
t.Errorf("Message %d: expected content '%s', got '%s'", i, expectedMsg.Content, actualMsg.Content)
|
t.Errorf("Message %d: expected content '%s', got '%s'", i, expectedMsg.Content, actualMsg.Content)
|
||||||
}
|
}
|
||||||
|
|
||||||
if actualMsg.ToolCallID != expectedMsg.ToolCallID {
|
if actualMsg.ToolCallID != expectedMsg.ToolCallID {
|
||||||
t.Errorf("Message %d: expected ToolCallID '%s', got '%s'", i, expectedMsg.ToolCallID, actualMsg.ToolCallID)
|
t.Errorf("Message %d: expected ToolCallID '%s', got '%s'", i, expectedMsg.ToolCallID, actualMsg.ToolCallID)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Additional check: ensure no messages were lost
|
// Additional check: ensure no messages were lost
|
||||||
if !reflect.DeepEqual(result, tt.expected) {
|
if !reflect.DeepEqual(result, tt.expected) {
|
||||||
t.Errorf("Result does not match expected:\nResult: %+v\nExpected: %+v", result, tt.expected)
|
t.Errorf("Result does not match expected:\nResult: %+v\nExpected: %+v", result, tt.expected)
|
||||||
@@ -153,7 +143,6 @@ func TestConsolidateConsecutiveAssistantMessages(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestUnmarshalFuncCall(t *testing.T) {
|
func TestUnmarshalFuncCall(t *testing.T) {
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
name string
|
name string
|
||||||
@@ -213,7 +202,6 @@ func TestUnmarshalFuncCall(t *testing.T) {
|
|||||||
wantErr: true,
|
wantErr: true,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tt := range tests {
|
for _, tt := range tests {
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
got, err := unmarshalFuncCall(tt.jsonStr)
|
got, err := unmarshalFuncCall(tt.jsonStr)
|
||||||
@@ -238,7 +226,6 @@ func TestUnmarshalFuncCall(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestConvertJSONToMapStringString(t *testing.T) {
|
func TestConvertJSONToMapStringString(t *testing.T) {
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
name string
|
name string
|
||||||
@@ -265,7 +252,6 @@ func TestConvertJSONToMapStringString(t *testing.T) {
|
|||||||
wantErr: true,
|
wantErr: true,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tt := range tests {
|
for _, tt := range tests {
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
got, err := convertJSONToMapStringString(tt.jsonStr)
|
got, err := convertJSONToMapStringString(tt.jsonStr)
|
||||||
@@ -287,7 +273,6 @@ func TestConvertJSONToMapStringString(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestParseKnownToTag(t *testing.T) {
|
func TestParseKnownToTag(t *testing.T) {
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
name string
|
name string
|
||||||
@@ -378,7 +363,6 @@ func TestParseKnownToTag(t *testing.T) {
|
|||||||
wantKnownTo: []string{"Alice", "Bob", "Carl"},
|
wantKnownTo: []string{"Alice", "Bob", "Carl"},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tt := range tests {
|
for _, tt := range tests {
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
// Set up config
|
// Set up config
|
||||||
@@ -402,7 +386,6 @@ func TestParseKnownToTag(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestProcessMessageTag(t *testing.T) {
|
func TestProcessMessageTag(t *testing.T) {
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
name string
|
name string
|
||||||
@@ -498,7 +481,6 @@ func TestProcessMessageTag(t *testing.T) {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tt := range tests {
|
for _, tt := range tests {
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
testCfg := &config.Config{
|
testCfg := &config.Config{
|
||||||
@@ -529,7 +511,6 @@ func TestProcessMessageTag(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestFilterMessagesForCharacter(t *testing.T) {
|
func TestFilterMessagesForCharacter(t *testing.T) {
|
||||||
messages := []models.RoleMsg{
|
messages := []models.RoleMsg{
|
||||||
{Role: "system", Content: "System message", KnownTo: nil}, // visible to all
|
{Role: "system", Content: "System message", KnownTo: nil}, // visible to all
|
||||||
@@ -539,7 +520,6 @@ func TestFilterMessagesForCharacter(t *testing.T) {
|
|||||||
{Role: "Alice", Content: "Private to Carl", KnownTo: []string{"Alice", "Carl"}},
|
{Role: "Alice", Content: "Private to Carl", KnownTo: []string{"Alice", "Carl"}},
|
||||||
{Role: "Carl", Content: "Hi all", KnownTo: nil}, // visible to all
|
{Role: "Carl", Content: "Hi all", KnownTo: nil}, // visible to all
|
||||||
}
|
}
|
||||||
|
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
name string
|
name string
|
||||||
enabled bool
|
enabled bool
|
||||||
@@ -583,7 +563,6 @@ func TestFilterMessagesForCharacter(t *testing.T) {
|
|||||||
wantIndices: []int{0, 1, 5},
|
wantIndices: []int{0, 1, 5},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tt := range tests {
|
for _, tt := range tests {
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
testCfg := &config.Config{
|
testCfg := &config.Config{
|
||||||
@@ -591,15 +570,12 @@ func TestFilterMessagesForCharacter(t *testing.T) {
|
|||||||
CharSpecificContextTag: "@",
|
CharSpecificContextTag: "@",
|
||||||
}
|
}
|
||||||
cfg = testCfg
|
cfg = testCfg
|
||||||
|
|
||||||
got := filterMessagesForCharacter(messages, tt.character)
|
got := filterMessagesForCharacter(messages, tt.character)
|
||||||
|
|
||||||
if len(got) != len(tt.wantIndices) {
|
if len(got) != len(tt.wantIndices) {
|
||||||
t.Errorf("filterMessagesForCharacter() returned %d messages, want %d", len(got), len(tt.wantIndices))
|
t.Errorf("filterMessagesForCharacter() returned %d messages, want %d", len(got), len(tt.wantIndices))
|
||||||
t.Logf("got: %v", got)
|
t.Logf("got: %v", got)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
for i, idx := range tt.wantIndices {
|
for i, idx := range tt.wantIndices {
|
||||||
if got[i].Content != messages[idx].Content {
|
if got[i].Content != messages[idx].Content {
|
||||||
t.Errorf("filterMessagesForCharacter() message %d content = %q, want %q", i, got[i].Content, messages[idx].Content)
|
t.Errorf("filterMessagesForCharacter() message %d content = %q, want %q", i, got[i].Content, messages[idx].Content)
|
||||||
@@ -608,7 +584,6 @@ func TestFilterMessagesForCharacter(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestRoleMsgCopyPreservesKnownTo(t *testing.T) {
|
func TestRoleMsgCopyPreservesKnownTo(t *testing.T) {
|
||||||
// Test that the Copy() method preserves the KnownTo field
|
// Test that the Copy() method preserves the KnownTo field
|
||||||
originalMsg := models.RoleMsg{
|
originalMsg := models.RoleMsg{
|
||||||
@@ -616,9 +591,7 @@ func TestRoleMsgCopyPreservesKnownTo(t *testing.T) {
|
|||||||
Content: "Test message",
|
Content: "Test message",
|
||||||
KnownTo: []string{"Bob", "Charlie"},
|
KnownTo: []string{"Bob", "Charlie"},
|
||||||
}
|
}
|
||||||
|
|
||||||
copiedMsg := originalMsg.Copy()
|
copiedMsg := originalMsg.Copy()
|
||||||
|
|
||||||
if copiedMsg.Role != originalMsg.Role {
|
if copiedMsg.Role != originalMsg.Role {
|
||||||
t.Errorf("Copy() failed to preserve Role: got %q, want %q", copiedMsg.Role, originalMsg.Role)
|
t.Errorf("Copy() failed to preserve Role: got %q, want %q", copiedMsg.Role, originalMsg.Role)
|
||||||
}
|
}
|
||||||
@@ -635,7 +608,6 @@ func TestRoleMsgCopyPreservesKnownTo(t *testing.T) {
|
|||||||
t.Errorf("Copy() failed to preserve hasContentParts flag")
|
t.Errorf("Copy() failed to preserve hasContentParts flag")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestKnownToFieldPreservationScenario(t *testing.T) {
|
func TestKnownToFieldPreservationScenario(t *testing.T) {
|
||||||
// Test the specific scenario from the log where KnownTo field was getting lost
|
// Test the specific scenario from the log where KnownTo field was getting lost
|
||||||
originalMsg := models.RoleMsg{
|
originalMsg := models.RoleMsg{
|
||||||
@@ -643,28 +615,22 @@ func TestKnownToFieldPreservationScenario(t *testing.T) {
|
|||||||
Content: `Alice: "Okay, Bob. The word is... **'Ephemeral'**. (ooc: @Bob@)"`,
|
Content: `Alice: "Okay, Bob. The word is... **'Ephemeral'**. (ooc: @Bob@)"`,
|
||||||
KnownTo: []string{"Bob"}, // This was detected in the log
|
KnownTo: []string{"Bob"}, // This was detected in the log
|
||||||
}
|
}
|
||||||
|
|
||||||
t.Logf("Original message - Role: %s, Content: %s, KnownTo: %v",
|
t.Logf("Original message - Role: %s, Content: %s, KnownTo: %v",
|
||||||
originalMsg.Role, originalMsg.Content, originalMsg.KnownTo)
|
originalMsg.Role, originalMsg.Content, originalMsg.KnownTo)
|
||||||
|
|
||||||
// Simulate what happens when the message gets copied during processing
|
// Simulate what happens when the message gets copied during processing
|
||||||
copiedMsg := originalMsg.Copy()
|
copiedMsg := originalMsg.Copy()
|
||||||
|
|
||||||
t.Logf("Copied message - Role: %s, Content: %s, KnownTo: %v",
|
t.Logf("Copied message - Role: %s, Content: %s, KnownTo: %v",
|
||||||
copiedMsg.Role, copiedMsg.Content, copiedMsg.KnownTo)
|
copiedMsg.Role, copiedMsg.Content, copiedMsg.KnownTo)
|
||||||
|
|
||||||
// Check if KnownTo field survived the copy
|
// Check if KnownTo field survived the copy
|
||||||
if len(copiedMsg.KnownTo) == 0 {
|
if len(copiedMsg.KnownTo) == 0 {
|
||||||
t.Error("ERROR: KnownTo field was lost during copy!")
|
t.Error("ERROR: KnownTo field was lost during copy!")
|
||||||
} else {
|
} else {
|
||||||
t.Log("SUCCESS: KnownTo field was preserved during copy!")
|
t.Log("SUCCESS: KnownTo field was preserved during copy!")
|
||||||
}
|
}
|
||||||
|
|
||||||
// Verify the content is the same
|
// Verify the content is the same
|
||||||
if copiedMsg.Content != originalMsg.Content {
|
if copiedMsg.Content != originalMsg.Content {
|
||||||
t.Errorf("Content was changed during copy: got %s, want %s", copiedMsg.Content, originalMsg.Content)
|
t.Errorf("Content was changed during copy: got %s, want %s", copiedMsg.Content, originalMsg.Content)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Verify the KnownTo slice is properly copied
|
// Verify the KnownTo slice is properly copied
|
||||||
if !reflect.DeepEqual(copiedMsg.KnownTo, originalMsg.KnownTo) {
|
if !reflect.DeepEqual(copiedMsg.KnownTo, originalMsg.KnownTo) {
|
||||||
t.Errorf("KnownTo was not properly copied: got %v, want %v", copiedMsg.KnownTo, originalMsg.KnownTo)
|
t.Errorf("KnownTo was not properly copied: got %v, want %v", copiedMsg.KnownTo, originalMsg.KnownTo)
|
||||||
|
|||||||
63
colors.go
Normal file
63
colors.go
Normal file
@@ -0,0 +1,63 @@
|
|||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"github.com/gdamore/tcell/v2"
|
||||||
|
"github.com/rivo/tview"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
colorschemes = map[string]tview.Theme{
|
||||||
|
"default": tview.Theme{
|
||||||
|
PrimitiveBackgroundColor: tcell.ColorDefault,
|
||||||
|
ContrastBackgroundColor: tcell.ColorGray,
|
||||||
|
MoreContrastBackgroundColor: tcell.ColorSteelBlue,
|
||||||
|
BorderColor: tcell.ColorGray,
|
||||||
|
TitleColor: tcell.ColorRed,
|
||||||
|
GraphicsColor: tcell.ColorBlue,
|
||||||
|
PrimaryTextColor: tcell.ColorLightGray,
|
||||||
|
SecondaryTextColor: tcell.ColorYellow,
|
||||||
|
TertiaryTextColor: tcell.ColorOrange,
|
||||||
|
InverseTextColor: tcell.ColorPurple,
|
||||||
|
ContrastSecondaryTextColor: tcell.ColorLime,
|
||||||
|
},
|
||||||
|
"gruvbox": tview.Theme{
|
||||||
|
PrimitiveBackgroundColor: tcell.NewHexColor(0x282828), // Background: #282828 (dark gray)
|
||||||
|
ContrastBackgroundColor: tcell.ColorDarkGoldenrod, // Selected option: warm yellow (#b57614)
|
||||||
|
MoreContrastBackgroundColor: tcell.ColorDarkSlateGray, // Non-selected options: dark grayish-blue (#32302f)
|
||||||
|
BorderColor: tcell.ColorLightGray, // Light gray (#a89984)
|
||||||
|
TitleColor: tcell.ColorRed, // Red (#fb4934)
|
||||||
|
GraphicsColor: tcell.ColorDarkCyan, // Cyan (#689d6a)
|
||||||
|
PrimaryTextColor: tcell.ColorLightGray, // Light gray (#d5c4a1)
|
||||||
|
SecondaryTextColor: tcell.ColorYellow, // Yellow (#fabd2f)
|
||||||
|
TertiaryTextColor: tcell.ColorOrange, // Orange (#fe8019)
|
||||||
|
InverseTextColor: tcell.ColorWhite, // White (#f9f5d7) for selected text
|
||||||
|
ContrastSecondaryTextColor: tcell.ColorLightGreen, // Light green (#b8bb26)
|
||||||
|
},
|
||||||
|
"solarized": tview.Theme{
|
||||||
|
PrimitiveBackgroundColor: tcell.NewHexColor(0x002b36), // Background: #002b36 (base03)
|
||||||
|
ContrastBackgroundColor: tcell.ColorDarkCyan, // Selected option: cyan (#2aa198)
|
||||||
|
MoreContrastBackgroundColor: tcell.ColorDarkSlateGray, // Non-selected options: dark blue (#073642)
|
||||||
|
BorderColor: tcell.ColorLightBlue, // Light blue (#839496)
|
||||||
|
TitleColor: tcell.ColorRed, // Red (#dc322f)
|
||||||
|
GraphicsColor: tcell.ColorBlue, // Blue (#268bd2)
|
||||||
|
PrimaryTextColor: tcell.ColorWhite, // White (#fdf6e3)
|
||||||
|
SecondaryTextColor: tcell.ColorYellow, // Yellow (#b58900)
|
||||||
|
TertiaryTextColor: tcell.ColorOrange, // Orange (#cb4b16)
|
||||||
|
InverseTextColor: tcell.ColorWhite, // White (#eee8d5) for selected text
|
||||||
|
ContrastSecondaryTextColor: tcell.ColorLightCyan, // Light cyan (#93a1a1)
|
||||||
|
},
|
||||||
|
"dracula": tview.Theme{
|
||||||
|
PrimitiveBackgroundColor: tcell.NewHexColor(0x282a36), // Background: #282a36
|
||||||
|
ContrastBackgroundColor: tcell.ColorDarkMagenta, // Selected option: magenta (#bd93f9)
|
||||||
|
MoreContrastBackgroundColor: tcell.ColorDarkGray, // Non-selected options: dark gray (#44475a)
|
||||||
|
BorderColor: tcell.ColorLightGray, // Light gray (#f8f8f2)
|
||||||
|
TitleColor: tcell.ColorRed, // Red (#ff5555)
|
||||||
|
GraphicsColor: tcell.ColorDarkCyan, // Cyan (#8be9fd)
|
||||||
|
PrimaryTextColor: tcell.ColorWhite, // White (#f8f8f2)
|
||||||
|
SecondaryTextColor: tcell.ColorYellow, // Yellow (#f1fa8c)
|
||||||
|
TertiaryTextColor: tcell.ColorOrange, // Orange (#ffb86c)
|
||||||
|
InverseTextColor: tcell.ColorWhite, // White (#f8f8f2) for selected text
|
||||||
|
ContrastSecondaryTextColor: tcell.ColorLightGreen, // Light green (#50fa7b)
|
||||||
|
},
|
||||||
|
}
|
||||||
|
)
|
||||||
@@ -10,7 +10,10 @@ DeepSeekModel = "deepseek-reasoner"
|
|||||||
OpenRouterCompletionAPI = "https://openrouter.ai/api/v1/completions"
|
OpenRouterCompletionAPI = "https://openrouter.ai/api/v1/completions"
|
||||||
OpenRouterChatAPI = "https://openrouter.ai/api/v1/chat/completions"
|
OpenRouterChatAPI = "https://openrouter.ai/api/v1/chat/completions"
|
||||||
# OpenRouterToken = ""
|
# OpenRouterToken = ""
|
||||||
|
# embeddings
|
||||||
EmbedURL = "http://localhost:8082/v1/embeddings"
|
EmbedURL = "http://localhost:8082/v1/embeddings"
|
||||||
|
HFToken = ""
|
||||||
|
#
|
||||||
ShowSys = true
|
ShowSys = true
|
||||||
LogFile = "log.txt"
|
LogFile = "log.txt"
|
||||||
UserRole = "user"
|
UserRole = "user"
|
||||||
@@ -21,6 +24,7 @@ ChunkLimit = 100000
|
|||||||
AutoScrollEnabled = true
|
AutoScrollEnabled = true
|
||||||
AutoCleanToolCallsFromCtx = false
|
AutoCleanToolCallsFromCtx = false
|
||||||
# rag settings
|
# rag settings
|
||||||
|
RAGEnabled = false
|
||||||
RAGBatchSize = 1
|
RAGBatchSize = 1
|
||||||
RAGWordLimit = 80
|
RAGWordLimit = 80
|
||||||
RAGWorkers = 2
|
RAGWorkers = 2
|
||||||
@@ -41,12 +45,15 @@ STT_LANG = "en" # Language for speech recognition (for WHISPER_BINARY mode)
|
|||||||
STT_SR = 16000 # Sample rate for audio recording
|
STT_SR = 16000 # Sample rate for audio recording
|
||||||
#
|
#
|
||||||
DBPATH = "gflt.db"
|
DBPATH = "gflt.db"
|
||||||
FilePickerDir = "." # Directory where file picker should start
|
FilePickerDir = "." # Directory for file picker start and coding assistant file operations (relative paths resolved against this)
|
||||||
FilePickerExts = "png,jpg,jpeg,gif,webp" # Comma-separated list of allowed file extensions for file picker
|
FilePickerExts = "png,jpg,jpeg,gif,webp" # Comma-separated list of allowed file extensions for file picker
|
||||||
CodingDir = "." # Default directory for coding assistant file operations (relative paths resolved against this)
|
|
||||||
EnableMouse = false # Enable mouse support in the UI
|
EnableMouse = false # Enable mouse support in the UI
|
||||||
# character specific context
|
# character specific context
|
||||||
CharSpecificContextEnabled = true
|
CharSpecificContextEnabled = true
|
||||||
CharSpecificContextTag = "@"
|
CharSpecificContextTag = "@"
|
||||||
AutoTurn = true
|
AutoTurn = true
|
||||||
StripThinkingFromAPI = true # Strip <think> blocks from messages before sending to LLM (keeps them in chat history)
|
StripThinkingFromAPI = true # Strip <think> blocks from messages before sending to LLM (keeps them in chat history)
|
||||||
|
# OpenRouter reasoning configuration (only applies to OpenRouter chat API)
|
||||||
|
# Valid values: xhigh, high, medium, low, minimal, none (empty or none = disabled)
|
||||||
|
# Models that support reasoning will include thinking content wrapped in <think> tags
|
||||||
|
ReasoningEffort = "medium"
|
||||||
|
|||||||
@@ -18,8 +18,8 @@ type Config struct {
|
|||||||
UserRole string `toml:"UserRole"`
|
UserRole string `toml:"UserRole"`
|
||||||
ToolRole string `toml:"ToolRole"`
|
ToolRole string `toml:"ToolRole"`
|
||||||
ToolUse bool `toml:"ToolUse"`
|
ToolUse bool `toml:"ToolUse"`
|
||||||
ThinkUse bool `toml:"ThinkUse"`
|
|
||||||
StripThinkingFromAPI bool `toml:"StripThinkingFromAPI"`
|
StripThinkingFromAPI bool `toml:"StripThinkingFromAPI"`
|
||||||
|
ReasoningEffort string `toml:"ReasoningEffort"`
|
||||||
AssistantRole string `toml:"AssistantRole"`
|
AssistantRole string `toml:"AssistantRole"`
|
||||||
SysDir string `toml:"SysDir"`
|
SysDir string `toml:"SysDir"`
|
||||||
ChunkLimit uint32 `toml:"ChunkLimit"`
|
ChunkLimit uint32 `toml:"ChunkLimit"`
|
||||||
@@ -31,15 +31,14 @@ type Config struct {
|
|||||||
DBPATH string `toml:"DBPATH"`
|
DBPATH string `toml:"DBPATH"`
|
||||||
FilePickerDir string `toml:"FilePickerDir"`
|
FilePickerDir string `toml:"FilePickerDir"`
|
||||||
FilePickerExts string `toml:"FilePickerExts"`
|
FilePickerExts string `toml:"FilePickerExts"`
|
||||||
CodingDir string `toml:"CodingDir"`
|
|
||||||
ImagePreview bool `toml:"ImagePreview"`
|
ImagePreview bool `toml:"ImagePreview"`
|
||||||
EnableMouse bool `toml:"EnableMouse"`
|
EnableMouse bool `toml:"EnableMouse"`
|
||||||
// embeddings
|
// embeddings
|
||||||
RAGEnabled bool `toml:"RAGEnabled"`
|
|
||||||
EmbedURL string `toml:"EmbedURL"`
|
EmbedURL string `toml:"EmbedURL"`
|
||||||
HFToken string `toml:"HFToken"`
|
HFToken string `toml:"HFToken"`
|
||||||
RAGDir string `toml:"RAGDir"`
|
|
||||||
// rag settings
|
// rag settings
|
||||||
|
RAGEnabled bool `toml:"RAGEnabled"`
|
||||||
|
RAGDir string `toml:"RAGDir"`
|
||||||
RAGWorkers uint32 `toml:"RAGWorkers"`
|
RAGWorkers uint32 `toml:"RAGWorkers"`
|
||||||
RAGBatchSize int `toml:"RAGBatchSize"`
|
RAGBatchSize int `toml:"RAGBatchSize"`
|
||||||
RAGWordLimit uint32 `toml:"RAGWordLimit"`
|
RAGWordLimit uint32 `toml:"RAGWordLimit"`
|
||||||
@@ -125,6 +124,9 @@ func LoadConfig(fn string) (*Config, error) {
|
|||||||
if config.CompletionAPI != "" {
|
if config.CompletionAPI != "" {
|
||||||
config.ApiLinks = append(config.ApiLinks, config.CompletionAPI)
|
config.ApiLinks = append(config.ApiLinks, config.CompletionAPI)
|
||||||
}
|
}
|
||||||
|
if config.RAGDir == "" {
|
||||||
|
config.RAGDir = "ragimport"
|
||||||
|
}
|
||||||
// if any value is empty fill with default
|
// if any value is empty fill with default
|
||||||
return config, nil
|
return config, nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -140,17 +140,24 @@ This document explains how to set up and configure the application using the `co
|
|||||||
- Path to the SQLite database file used for storing conversation history and other data.
|
- Path to the SQLite database file used for storing conversation history and other data.
|
||||||
|
|
||||||
#### FilePickerDir (`"."`)
|
#### FilePickerDir (`"."`)
|
||||||
- Directory where the file (image) picker should start when selecting files.
|
- Directory where the file picker starts and where relative paths in coding assistant file tools (file_read, file_write, etc.) are resolved against. Use absolute paths (starting with `/`) to bypass this.
|
||||||
|
|
||||||
#### FilePickerExts (`"png,jpg,jpeg,gif,webp"`)
|
|
||||||
- Comma-separated list of allowed file extensions for the file picker.
|
|
||||||
|
|
||||||
#### CodingDir (`"."`)
|
|
||||||
- Default directory for coding assistant file operations. Relative paths in file tools (file_read, file_write, etc.) are resolved against this directory. Use absolute paths (starting with `/`) to bypass this.
|
|
||||||
|
|
||||||
#### EnableMouse (`false`)
|
#### EnableMouse (`false`)
|
||||||
- Enable or disable mouse support in the UI. When set to `true`, allows clicking buttons and interacting with UI elements using the mouse, but prevents the terminal from handling mouse events normally (such as selecting and copying text). When set to `false`, enables default terminal behavior allowing you to select and copy text, but disables mouse interaction with UI elements.
|
- Enable or disable mouse support in the UI. When set to `true`, allows clicking buttons and interacting with UI elements using the mouse, but prevents the terminal from handling mouse events normally (such as selecting and copying text). When set to `false`, enables default terminal behavior allowing you to select and copy text, but disables mouse interaction with UI elements.
|
||||||
|
|
||||||
|
### Character-Specific Context Settings (/completion only)
|
||||||
|
|
||||||
|
[character specific context page for more info](./char-specific-context.md)
|
||||||
|
|
||||||
|
#### CharSpecificContextEnabled (`true`)
|
||||||
|
- Enable or disable character-specific context functionality.
|
||||||
|
|
||||||
|
#### CharSpecificContextTag (`"@"`)
|
||||||
|
- The tag prefix used to reference character-specific context in messages.
|
||||||
|
|
||||||
|
#### AutoTurn (`true`)
|
||||||
|
- Enable or disable automatic turn detection/switching.
|
||||||
|
|
||||||
### Additional Features
|
### Additional Features
|
||||||
|
|
||||||
Those could be switched in program, but also bould be setup in config.
|
Those could be switched in program, but also bould be setup in config.
|
||||||
@@ -158,8 +165,11 @@ Those could be switched in program, but also bould be setup in config.
|
|||||||
#### ToolUse
|
#### ToolUse
|
||||||
- Enable or disable explanation of tools to llm, so it could use them.
|
- Enable or disable explanation of tools to llm, so it could use them.
|
||||||
|
|
||||||
#### ThinkUse
|
### StripThinkingFromAPI (`true`)
|
||||||
- Enable or disable insertion of <think> token at the beggining of llm resp.
|
- Strip thinking blocks from messages before sending to LLM. Keeps them in chat history for local viewing but reduces token usage in API calls.
|
||||||
|
|
||||||
|
#### ReasoningEffort (`"medium"`)
|
||||||
|
- OpenRouter reasoning configuration (only applies to OpenRouter chat API). Valid values: `xhigh`, `high`, `medium`, `low`, `minimal`, `none`. Empty or `none` disables reasoning.
|
||||||
|
|
||||||
## Environment Variables
|
## Environment Variables
|
||||||
|
|
||||||
|
|||||||
50
extra/tts.go
50
extra/tts.go
@@ -92,8 +92,6 @@ func (o *KokoroOrator) stoproutine() {
|
|||||||
|
|
||||||
func (o *KokoroOrator) readroutine() {
|
func (o *KokoroOrator) readroutine() {
|
||||||
tokenizer, _ := english.NewSentenceTokenizer(nil)
|
tokenizer, _ := english.NewSentenceTokenizer(nil)
|
||||||
// var sentenceBuf bytes.Buffer
|
|
||||||
// var remainder strings.Builder
|
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case chunk := <-TTSTextChan:
|
case chunk := <-TTSTextChan:
|
||||||
@@ -106,24 +104,28 @@ func (o *KokoroOrator) readroutine() {
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
text := o.textBuffer.String()
|
text := o.textBuffer.String()
|
||||||
o.mu.Unlock()
|
|
||||||
sentences := tokenizer.Tokenize(text)
|
sentences := tokenizer.Tokenize(text)
|
||||||
o.logger.Debug("adding chunk", "chunk", chunk, "text", text, "sen-len", len(sentences))
|
o.logger.Debug("adding chunk", "chunk", chunk, "text", text, "sen-len", len(sentences))
|
||||||
for i, sentence := range sentences {
|
if len(sentences) <= 1 {
|
||||||
if i == len(sentences)-1 { // last sentence
|
|
||||||
o.mu.Lock()
|
|
||||||
o.textBuffer.Reset()
|
|
||||||
_, err := o.textBuffer.WriteString(sentence.Text)
|
|
||||||
o.mu.Unlock()
|
o.mu.Unlock()
|
||||||
if err != nil {
|
|
||||||
o.logger.Warn("failed to write to stringbuilder", "error", err)
|
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
continue // if only one (often incomplete) sentence; wait for next chunk
|
completeSentences := sentences[:len(sentences)-1]
|
||||||
|
remaining := sentences[len(sentences)-1].Text
|
||||||
|
o.textBuffer.Reset()
|
||||||
|
o.textBuffer.WriteString(remaining)
|
||||||
|
o.mu.Unlock()
|
||||||
|
|
||||||
|
for _, sentence := range completeSentences {
|
||||||
|
o.mu.Lock()
|
||||||
|
interrupted := o.interrupt
|
||||||
|
o.mu.Unlock()
|
||||||
|
if interrupted {
|
||||||
|
return
|
||||||
}
|
}
|
||||||
cleanedText := models.CleanText(sentence.Text)
|
cleanedText := models.CleanText(sentence.Text)
|
||||||
if cleanedText == "" {
|
if cleanedText == "" {
|
||||||
continue // Skip empty text after cleaning
|
continue
|
||||||
}
|
}
|
||||||
o.logger.Debug("calling Speak with sentence", "sent", cleanedText)
|
o.logger.Debug("calling Speak with sentence", "sent", cleanedText)
|
||||||
if err := o.Speak(cleanedText); err != nil {
|
if err := o.Speak(cleanedText); err != nil {
|
||||||
@@ -338,24 +340,28 @@ func (o *GoogleTranslateOrator) readroutine() {
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
text := o.textBuffer.String()
|
text := o.textBuffer.String()
|
||||||
o.mu.Unlock()
|
|
||||||
sentences := tokenizer.Tokenize(text)
|
sentences := tokenizer.Tokenize(text)
|
||||||
o.logger.Debug("adding chunk", "chunk", chunk, "text", text, "sen-len", len(sentences))
|
o.logger.Debug("adding chunk", "chunk", chunk, "text", text, "sen-len", len(sentences))
|
||||||
for i, sentence := range sentences {
|
if len(sentences) <= 1 {
|
||||||
if i == len(sentences)-1 { // last sentence
|
|
||||||
o.mu.Lock()
|
|
||||||
o.textBuffer.Reset()
|
|
||||||
_, err := o.textBuffer.WriteString(sentence.Text)
|
|
||||||
o.mu.Unlock()
|
o.mu.Unlock()
|
||||||
if err != nil {
|
|
||||||
o.logger.Warn("failed to write to stringbuilder", "error", err)
|
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
continue // if only one (often incomplete) sentence; wait for next chunk
|
completeSentences := sentences[:len(sentences)-1]
|
||||||
|
remaining := sentences[len(sentences)-1].Text
|
||||||
|
o.textBuffer.Reset()
|
||||||
|
o.textBuffer.WriteString(remaining)
|
||||||
|
o.mu.Unlock()
|
||||||
|
|
||||||
|
for _, sentence := range completeSentences {
|
||||||
|
o.mu.Lock()
|
||||||
|
interrupted := o.interrupt
|
||||||
|
o.mu.Unlock()
|
||||||
|
if interrupted {
|
||||||
|
return
|
||||||
}
|
}
|
||||||
cleanedText := models.CleanText(sentence.Text)
|
cleanedText := models.CleanText(sentence.Text)
|
||||||
if cleanedText == "" {
|
if cleanedText == "" {
|
||||||
continue // Skip empty text after cleaning
|
continue
|
||||||
}
|
}
|
||||||
o.logger.Debug("calling Speak with sentence", "sent", cleanedText)
|
o.logger.Debug("calling Speak with sentence", "sent", cleanedText)
|
||||||
if err := o.Speak(cleanedText); err != nil {
|
if err := o.Speak(cleanedText); err != nil {
|
||||||
|
|||||||
4
go.mod
4
go.mod
@@ -6,17 +6,19 @@ require (
|
|||||||
github.com/BurntSushi/toml v1.5.0
|
github.com/BurntSushi/toml v1.5.0
|
||||||
github.com/GrailFinder/google-translate-tts v0.1.3
|
github.com/GrailFinder/google-translate-tts v0.1.3
|
||||||
github.com/GrailFinder/searchagent v0.2.0
|
github.com/GrailFinder/searchagent v0.2.0
|
||||||
|
github.com/PuerkitoBio/goquery v1.11.0
|
||||||
github.com/gdamore/tcell/v2 v2.13.2
|
github.com/gdamore/tcell/v2 v2.13.2
|
||||||
github.com/glebarez/go-sqlite v1.22.0
|
github.com/glebarez/go-sqlite v1.22.0
|
||||||
github.com/gopxl/beep/v2 v2.1.1
|
github.com/gopxl/beep/v2 v2.1.1
|
||||||
github.com/gordonklaus/portaudio v0.0.0-20250206071425-98a94950218b
|
github.com/gordonklaus/portaudio v0.0.0-20250206071425-98a94950218b
|
||||||
github.com/jmoiron/sqlx v1.4.0
|
github.com/jmoiron/sqlx v1.4.0
|
||||||
|
github.com/ledongthuc/pdf v0.0.0-20250511090121-5959a4027728
|
||||||
github.com/neurosnap/sentences v1.1.2
|
github.com/neurosnap/sentences v1.1.2
|
||||||
github.com/rivo/tview v0.42.0
|
github.com/rivo/tview v0.42.0
|
||||||
|
github.com/yuin/goldmark v1.4.13
|
||||||
)
|
)
|
||||||
|
|
||||||
require (
|
require (
|
||||||
github.com/PuerkitoBio/goquery v1.11.0 // indirect
|
|
||||||
github.com/andybalholm/cascadia v1.3.3 // indirect
|
github.com/andybalholm/cascadia v1.3.3 // indirect
|
||||||
github.com/dustin/go-humanize v1.0.1 // indirect
|
github.com/dustin/go-humanize v1.0.1 // indirect
|
||||||
github.com/ebitengine/oto/v3 v3.4.0 // indirect
|
github.com/ebitengine/oto/v3 v3.4.0 // indirect
|
||||||
|
|||||||
3
go.sum
3
go.sum
@@ -43,6 +43,8 @@ github.com/hashicorp/golang-lru/v2 v2.0.7 h1:a+bsQ5rvGLjzHuww6tVxozPZFVghXaHOwFs
|
|||||||
github.com/hashicorp/golang-lru/v2 v2.0.7/go.mod h1:QeFd9opnmA6QUJc5vARoKUSoFhyfM2/ZepoAG6RGpeM=
|
github.com/hashicorp/golang-lru/v2 v2.0.7/go.mod h1:QeFd9opnmA6QUJc5vARoKUSoFhyfM2/ZepoAG6RGpeM=
|
||||||
github.com/jmoiron/sqlx v1.4.0 h1:1PLqN7S1UYp5t4SrVVnt4nUVNemrDAtxlulVe+Qgm3o=
|
github.com/jmoiron/sqlx v1.4.0 h1:1PLqN7S1UYp5t4SrVVnt4nUVNemrDAtxlulVe+Qgm3o=
|
||||||
github.com/jmoiron/sqlx v1.4.0/go.mod h1:ZrZ7UsYB/weZdl2Bxg6jCRO9c3YHl8r3ahlKmRT4JLY=
|
github.com/jmoiron/sqlx v1.4.0/go.mod h1:ZrZ7UsYB/weZdl2Bxg6jCRO9c3YHl8r3ahlKmRT4JLY=
|
||||||
|
github.com/ledongthuc/pdf v0.0.0-20250511090121-5959a4027728 h1:QwWKgMY28TAXaDl+ExRDqGQltzXqN/xypdKP86niVn8=
|
||||||
|
github.com/ledongthuc/pdf v0.0.0-20250511090121-5959a4027728/go.mod h1:1fEHWurg7pvf5SG6XNE5Q8UZmOwex51Mkx3SLhrW5B4=
|
||||||
github.com/lib/pq v1.10.9 h1:YXG7RB+JIjhP29X+OtkiDnYaXQwpS4JEWq7dtCCRUEw=
|
github.com/lib/pq v1.10.9 h1:YXG7RB+JIjhP29X+OtkiDnYaXQwpS4JEWq7dtCCRUEw=
|
||||||
github.com/lib/pq v1.10.9/go.mod h1:AlVN5x4E4T544tWzH6hKfbfQvm3HdbOxrmggDNAPY9o=
|
github.com/lib/pq v1.10.9/go.mod h1:AlVN5x4E4T544tWzH6hKfbfQvm3HdbOxrmggDNAPY9o=
|
||||||
github.com/lucasb-eyer/go-colorful v1.3.0 h1:2/yBRLdWBZKrf7gB40FoiKfAWYQ0lqNcbuQwVHXptag=
|
github.com/lucasb-eyer/go-colorful v1.3.0 h1:2/yBRLdWBZKrf7gB40FoiKfAWYQ0lqNcbuQwVHXptag=
|
||||||
@@ -67,6 +69,7 @@ github.com/rivo/uniseg v0.4.7 h1:WUdvkW8uEhrYfLC4ZzdpI2ztxP1I582+49Oc5Mq64VQ=
|
|||||||
github.com/rivo/uniseg v0.4.7/go.mod h1:FN3SvrM+Zdj16jyLfmOkMNblXMcoc8DfTHruCPUcx88=
|
github.com/rivo/uniseg v0.4.7/go.mod h1:FN3SvrM+Zdj16jyLfmOkMNblXMcoc8DfTHruCPUcx88=
|
||||||
github.com/stretchr/testify v1.10.0 h1:Xv5erBjTwe/5IxqUQTdXv5kgmIvbHo3QQyRwhJsOfJA=
|
github.com/stretchr/testify v1.10.0 h1:Xv5erBjTwe/5IxqUQTdXv5kgmIvbHo3QQyRwhJsOfJA=
|
||||||
github.com/stretchr/testify v1.10.0/go.mod h1:r2ic/lqez/lEtzL7wO/rwa5dbSLXVDPFyf8C91i36aY=
|
github.com/stretchr/testify v1.10.0/go.mod h1:r2ic/lqez/lEtzL7wO/rwa5dbSLXVDPFyf8C91i36aY=
|
||||||
|
github.com/yuin/goldmark v1.4.13 h1:fVcFKWvrslecOb/tg+Cc05dkeYx540o0FuFt3nUVDoE=
|
||||||
github.com/yuin/goldmark v1.4.13/go.mod h1:6yULJ656Px+3vBD8DxQVa3kxgyrAnzto9xy5taEt/CY=
|
github.com/yuin/goldmark v1.4.13/go.mod h1:6yULJ656Px+3vBD8DxQVa3kxgyrAnzto9xy5taEt/CY=
|
||||||
golang.org/x/crypto v0.0.0-20190308221718-c2843e01d9a2/go.mod h1:djNgcEr1/C05ACkg1iLfiJU5Ep61QUkGW8qpdssI0+w=
|
golang.org/x/crypto v0.0.0-20190308221718-c2843e01d9a2/go.mod h1:djNgcEr1/C05ACkg1iLfiJU5Ep61QUkGW8qpdssI0+w=
|
||||||
golang.org/x/crypto v0.0.0-20210921155107-089bfa567519/go.mod h1:GvvjBRRGRdwPK5ydBHafDWAxML/pGHZbMvKqRZ5+Abc=
|
golang.org/x/crypto v0.0.0-20210921155107-089bfa567519/go.mod h1:GvvjBRRGRdwPK5ydBHafDWAxML/pGHZbMvKqRZ5+Abc=
|
||||||
|
|||||||
131
helpfuncs.go
131
helpfuncs.go
@@ -9,8 +9,10 @@ import (
|
|||||||
"os"
|
"os"
|
||||||
"os/exec"
|
"os/exec"
|
||||||
"path"
|
"path"
|
||||||
|
"path/filepath"
|
||||||
"slices"
|
"slices"
|
||||||
"strings"
|
"strings"
|
||||||
|
"time"
|
||||||
"unicode"
|
"unicode"
|
||||||
|
|
||||||
"math/rand/v2"
|
"math/rand/v2"
|
||||||
@@ -18,6 +20,45 @@ import (
|
|||||||
"github.com/rivo/tview"
|
"github.com/rivo/tview"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// Cached model color - updated by background goroutine
|
||||||
|
var cachedModelColor string = "orange"
|
||||||
|
|
||||||
|
// startModelColorUpdater starts a background goroutine that periodically updates
|
||||||
|
// the cached model color. Only runs HTTP requests for local llama.cpp APIs.
|
||||||
|
func startModelColorUpdater() {
|
||||||
|
go func() {
|
||||||
|
ticker := time.NewTicker(5 * time.Second)
|
||||||
|
defer ticker.Stop()
|
||||||
|
|
||||||
|
// Initial check
|
||||||
|
updateCachedModelColor()
|
||||||
|
for range ticker.C {
|
||||||
|
updateCachedModelColor()
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
|
// updateCachedModelColor updates the global cachedModelColor variable
|
||||||
|
func updateCachedModelColor() {
|
||||||
|
if !isLocalLlamacpp() {
|
||||||
|
cachedModelColor = "orange"
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check if model is loaded
|
||||||
|
loaded, err := isModelLoaded(chatBody.Model)
|
||||||
|
if err != nil {
|
||||||
|
// On error, assume not loaded (red)
|
||||||
|
cachedModelColor = "red"
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if loaded {
|
||||||
|
cachedModelColor = "green"
|
||||||
|
} else {
|
||||||
|
cachedModelColor = "red"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func isASCII(s string) bool {
|
func isASCII(s string) bool {
|
||||||
for i := 0; i < len(s); i++ {
|
for i := 0; i < len(s); i++ {
|
||||||
if s[i] > unicode.MaxASCII {
|
if s[i] > unicode.MaxASCII {
|
||||||
@@ -33,15 +74,16 @@ func stripThinkingFromMsg(msg *models.RoleMsg) *models.RoleMsg {
|
|||||||
if !cfg.StripThinkingFromAPI {
|
if !cfg.StripThinkingFromAPI {
|
||||||
return msg
|
return msg
|
||||||
}
|
}
|
||||||
// Skip user, tool, and system messages - they might contain thinking examples
|
// Skip user, tool, they might contain thinking and system messages - examples
|
||||||
if msg.Role == cfg.UserRole || msg.Role == cfg.ToolRole || msg.Role == "system" {
|
if msg.Role == cfg.UserRole || msg.Role == cfg.ToolRole || msg.Role == "system" {
|
||||||
return msg
|
return msg
|
||||||
}
|
}
|
||||||
// Strip thinking from assistant messages
|
// Strip thinking from assistant messages
|
||||||
if thinkRE.MatchString(msg.Content) {
|
msgText := msg.GetText()
|
||||||
msg.Content = thinkRE.ReplaceAllString(msg.Content, "")
|
if thinkRE.MatchString(msgText) {
|
||||||
// Clean up any double newlines that might result
|
cleanedText := thinkRE.ReplaceAllString(msgText, "")
|
||||||
msg.Content = strings.TrimSpace(msg.Content)
|
cleanedText = strings.TrimSpace(cleanedText)
|
||||||
|
msg.SetText(cleanedText)
|
||||||
}
|
}
|
||||||
return msg
|
return msg
|
||||||
}
|
}
|
||||||
@@ -60,19 +102,23 @@ func refreshChatDisplay() {
|
|||||||
displayText := chatToText(filteredMessages, cfg.ShowSys)
|
displayText := chatToText(filteredMessages, cfg.ShowSys)
|
||||||
textView.SetText(displayText)
|
textView.SetText(displayText)
|
||||||
colorText()
|
colorText()
|
||||||
|
updateStatusLine()
|
||||||
if scrollToEndEnabled {
|
if scrollToEndEnabled {
|
||||||
textView.ScrollToEnd()
|
textView.ScrollToEnd()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// stopTTSIfNotForUser: character specific context, not meant fot the human to hear
|
||||||
func stopTTSIfNotForUser(msg *models.RoleMsg) {
|
func stopTTSIfNotForUser(msg *models.RoleMsg) {
|
||||||
|
if strings.Contains(cfg.CurrentAPI, "/chat") || !cfg.CharSpecificContextEnabled {
|
||||||
|
return
|
||||||
|
}
|
||||||
viewingAs := cfg.UserRole
|
viewingAs := cfg.UserRole
|
||||||
if cfg.WriteNextMsgAs != "" {
|
if cfg.WriteNextMsgAs != "" {
|
||||||
viewingAs = cfg.WriteNextMsgAs
|
viewingAs = cfg.WriteNextMsgAs
|
||||||
}
|
}
|
||||||
// stop tts if msg is not for user
|
// stop tts if msg is not for user
|
||||||
if cfg.CharSpecificContextEnabled &&
|
if !slices.Contains(msg.KnownTo, viewingAs) && cfg.TTS_ENABLED {
|
||||||
!slices.Contains(msg.KnownTo, viewingAs) && cfg.TTS_ENABLED {
|
|
||||||
TTSDoneChan <- true
|
TTSDoneChan <- true
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -130,8 +176,8 @@ func colorText() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func updateStatusLine() {
|
func updateStatusLine() {
|
||||||
statusLineWidget.SetText(makeStatusLine())
|
status := makeStatusLine()
|
||||||
helpView.SetText(fmt.Sprintf(helpText, makeStatusLine()))
|
statusLineWidget.SetText(status)
|
||||||
}
|
}
|
||||||
|
|
||||||
func initSysCards() ([]string, error) {
|
func initSysCards() ([]string, error) {
|
||||||
@@ -273,22 +319,11 @@ func isLocalLlamacpp() bool {
|
|||||||
return host == "localhost" || host == "127.0.0.1" || host == "::1"
|
return host == "localhost" || host == "127.0.0.1" || host == "::1"
|
||||||
}
|
}
|
||||||
|
|
||||||
// getModelColor returns the color tag for the model name based on its load status.
|
// getModelColor returns the cached color tag for the model name.
|
||||||
|
// The cached value is updated by a background goroutine every 5 seconds.
|
||||||
// For non-local models, returns orange. For local llama.cpp models, returns green if loaded, red if not.
|
// For non-local models, returns orange. For local llama.cpp models, returns green if loaded, red if not.
|
||||||
func getModelColor() string {
|
func getModelColor() string {
|
||||||
if !isLocalLlamacpp() {
|
return cachedModelColor
|
||||||
return "orange"
|
|
||||||
}
|
|
||||||
// Check if model is loaded
|
|
||||||
loaded, err := isModelLoaded(chatBody.Model)
|
|
||||||
if err != nil {
|
|
||||||
// On error, assume not loaded (red)
|
|
||||||
return "red"
|
|
||||||
}
|
|
||||||
if loaded {
|
|
||||||
return "green"
|
|
||||||
}
|
|
||||||
return "red"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func makeStatusLine() string {
|
func makeStatusLine() string {
|
||||||
@@ -322,10 +357,19 @@ func makeStatusLine() string {
|
|||||||
}
|
}
|
||||||
// Get model color based on load status for local llama.cpp models
|
// Get model color based on load status for local llama.cpp models
|
||||||
modelColor := getModelColor()
|
modelColor := getModelColor()
|
||||||
statusLine := fmt.Sprintf(indexLineCompletion, boolColors[botRespMode], botRespMode, activeChatName,
|
statusLine := fmt.Sprintf(statusLineTempl, boolColors[botRespMode], activeChatName,
|
||||||
boolColors[cfg.ToolUse], cfg.ToolUse, modelColor, chatBody.Model, boolColors[cfg.SkipLLMResp],
|
boolColors[cfg.ToolUse], modelColor, chatBody.Model, boolColors[cfg.SkipLLMResp],
|
||||||
cfg.SkipLLMResp, cfg.CurrentAPI, boolColors[isRecording], isRecording, persona,
|
cfg.CurrentAPI, persona, botPersona)
|
||||||
botPersona, boolColors[injectRole], injectRole)
|
if cfg.STT_ENABLED {
|
||||||
|
recordingS := fmt.Sprintf(" | [%s:-:b]voice recording[-:-:-] (ctrl+r)",
|
||||||
|
boolColors[isRecording])
|
||||||
|
statusLine += recordingS
|
||||||
|
}
|
||||||
|
// completion endpoint
|
||||||
|
if !strings.Contains(cfg.CurrentAPI, "chat") {
|
||||||
|
roleInject := fmt.Sprintf(" | [%s:-:b]role injection[-:-:-] (alt+7)", boolColors[injectRole])
|
||||||
|
statusLine += roleInject
|
||||||
|
}
|
||||||
return statusLine + imageInfo + shellModeInfo
|
return statusLine + imageInfo + shellModeInfo
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -706,23 +750,44 @@ func searchPrev() {
|
|||||||
// == tab completion ==
|
// == tab completion ==
|
||||||
|
|
||||||
func scanFiles(dir, filter string) []string {
|
func scanFiles(dir, filter string) []string {
|
||||||
|
const maxDepth = 3
|
||||||
|
const maxFiles = 50
|
||||||
var files []string
|
var files []string
|
||||||
entries, err := os.ReadDir(dir)
|
var scanRecursive func(currentDir string, currentDepth int, relPath string)
|
||||||
|
scanRecursive = func(currentDir string, currentDepth int, relPath string) {
|
||||||
|
if len(files) >= maxFiles {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if currentDepth > maxDepth {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
entries, err := os.ReadDir(currentDir)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return files
|
return
|
||||||
}
|
}
|
||||||
for _, entry := range entries {
|
for _, entry := range entries {
|
||||||
|
if len(files) >= maxFiles {
|
||||||
|
return
|
||||||
|
}
|
||||||
name := entry.Name()
|
name := entry.Name()
|
||||||
if strings.HasPrefix(name, ".") {
|
if strings.HasPrefix(name, ".") {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if filter == "" || strings.HasPrefix(strings.ToLower(name), strings.ToLower(filter)) {
|
fullPath := name
|
||||||
|
if relPath != "" {
|
||||||
|
fullPath = relPath + "/" + name
|
||||||
|
}
|
||||||
if entry.IsDir() {
|
if entry.IsDir() {
|
||||||
files = append(files, name+"/")
|
// Recursively scan subdirectories
|
||||||
} else {
|
scanRecursive(filepath.Join(currentDir, name), currentDepth+1, fullPath)
|
||||||
files = append(files, name)
|
continue
|
||||||
|
}
|
||||||
|
// Check if file matches filter
|
||||||
|
if filter == "" || strings.HasPrefix(strings.ToLower(fullPath), strings.ToLower(filter)) {
|
||||||
|
files = append(files, fullPath)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
scanRecursive(dir, 0, "")
|
||||||
return files
|
return files
|
||||||
}
|
}
|
||||||
|
|||||||
128
llm.go
128
llm.go
@@ -11,7 +11,6 @@ import (
|
|||||||
|
|
||||||
var imageAttachmentPath string // Global variable to track image attachment for next message
|
var imageAttachmentPath string // Global variable to track image attachment for next message
|
||||||
var lastImg string // for ctrl+j
|
var lastImg string // for ctrl+j
|
||||||
var RAGMsg = "Retrieved context for user's query:\n"
|
|
||||||
|
|
||||||
// containsToolSysMsg checks if the toolSysMsg already exists in the chat body
|
// containsToolSysMsg checks if the toolSysMsg already exists in the chat body
|
||||||
func containsToolSysMsg() bool {
|
func containsToolSysMsg() bool {
|
||||||
@@ -142,22 +141,6 @@ func (lcp LCPCompletion) FormMsg(msg, role string, resume bool) (io.Reader, erro
|
|||||||
newMsg = *processMessageTag(&newMsg)
|
newMsg = *processMessageTag(&newMsg)
|
||||||
chatBody.Messages = append(chatBody.Messages, newMsg)
|
chatBody.Messages = append(chatBody.Messages, newMsg)
|
||||||
}
|
}
|
||||||
// if rag - add as system message to avoid conflicts with tool usage
|
|
||||||
if !resume && cfg.RAGEnabled {
|
|
||||||
um := chatBody.Messages[len(chatBody.Messages)-1].Content
|
|
||||||
logger.Debug("RAG is enabled, preparing RAG context", "user_message", um)
|
|
||||||
ragResp, err := chatRagUse(um)
|
|
||||||
if err != nil {
|
|
||||||
logger.Error("failed to form a rag msg", "error", err)
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
logger.Debug("RAG response received", "response_len", len(ragResp),
|
|
||||||
"response_preview", ragResp[:min(len(ragResp), 100)])
|
|
||||||
// Use system role for RAG context to avoid conflicts with tool usage
|
|
||||||
ragMsg := models.RoleMsg{Role: "system", Content: RAGMsg + ragResp}
|
|
||||||
chatBody.Messages = append(chatBody.Messages, ragMsg)
|
|
||||||
logger.Debug("RAG message added to chat body", "message_count", len(chatBody.Messages))
|
|
||||||
}
|
|
||||||
// sending description of the tools and how to use them
|
// sending description of the tools and how to use them
|
||||||
if cfg.ToolUse && !resume && role == cfg.UserRole && !containsToolSysMsg() {
|
if cfg.ToolUse && !resume && role == cfg.UserRole && !containsToolSysMsg() {
|
||||||
chatBody.Messages = append(chatBody.Messages, models.RoleMsg{Role: cfg.ToolRole, Content: toolSysMsg})
|
chatBody.Messages = append(chatBody.Messages, models.RoleMsg{Role: cfg.ToolRole, Content: toolSysMsg})
|
||||||
@@ -184,9 +167,6 @@ func (lcp LCPCompletion) FormMsg(msg, role string, resume bool) (io.Reader, erro
|
|||||||
botMsgStart := "\n" + botPersona + ":\n"
|
botMsgStart := "\n" + botPersona + ":\n"
|
||||||
prompt += botMsgStart
|
prompt += botMsgStart
|
||||||
}
|
}
|
||||||
if cfg.ThinkUse && !cfg.ToolUse {
|
|
||||||
prompt += "<think>"
|
|
||||||
}
|
|
||||||
logger.Debug("checking prompt for /completion", "tool_use", cfg.ToolUse,
|
logger.Debug("checking prompt for /completion", "tool_use", cfg.ToolUse,
|
||||||
"msg", msg, "resume", resume, "prompt", prompt, "multimodal_data_count", len(multimodalData))
|
"msg", msg, "resume", resume, "prompt", prompt, "multimodal_data_count", len(multimodalData))
|
||||||
payload := models.NewLCPReq(prompt, chatBody.Model, multimodalData,
|
payload := models.NewLCPReq(prompt, chatBody.Model, multimodalData,
|
||||||
@@ -236,11 +216,11 @@ func (op LCPChat) ParseChunk(data []byte) (*models.TextChunk, error) {
|
|||||||
logger.Warn("LCPChat ParseChunk: no choices in response", "data", string(data))
|
logger.Warn("LCPChat ParseChunk: no choices in response", "data", string(data))
|
||||||
return &models.TextChunk{Finished: true}, nil
|
return &models.TextChunk{Finished: true}, nil
|
||||||
}
|
}
|
||||||
|
lastChoice := llmchunk.Choices[len(llmchunk.Choices)-1]
|
||||||
resp := &models.TextChunk{
|
resp := &models.TextChunk{
|
||||||
Chunk: llmchunk.Choices[len(llmchunk.Choices)-1].Delta.Content,
|
Chunk: lastChoice.Delta.Content,
|
||||||
|
Reasoning: lastChoice.Delta.ReasoningContent,
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check for tool calls in all choices, not just the last one
|
// Check for tool calls in all choices, not just the last one
|
||||||
for _, choice := range llmchunk.Choices {
|
for _, choice := range llmchunk.Choices {
|
||||||
if len(choice.Delta.ToolCalls) > 0 {
|
if len(choice.Delta.ToolCalls) > 0 {
|
||||||
@@ -255,8 +235,7 @@ func (op LCPChat) ParseChunk(data []byte) (*models.TextChunk, error) {
|
|||||||
break // Process only the first tool call
|
break // Process only the first tool call
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
if lastChoice.FinishReason == "stop" {
|
||||||
if llmchunk.Choices[len(llmchunk.Choices)-1].FinishReason == "stop" {
|
|
||||||
if resp.Chunk != "" {
|
if resp.Chunk != "" {
|
||||||
logger.Error("text inside of finish llmchunk", "chunk", llmchunk)
|
logger.Error("text inside of finish llmchunk", "chunk", llmchunk)
|
||||||
}
|
}
|
||||||
@@ -302,23 +281,6 @@ func (op LCPChat) FormMsg(msg, role string, resume bool) (io.Reader, error) {
|
|||||||
logger.Debug("LCPChat FormMsg: added message to chatBody", "role", newMsg.Role,
|
logger.Debug("LCPChat FormMsg: added message to chatBody", "role", newMsg.Role,
|
||||||
"content_len", len(newMsg.Content), "message_count_after_add", len(chatBody.Messages))
|
"content_len", len(newMsg.Content), "message_count_after_add", len(chatBody.Messages))
|
||||||
}
|
}
|
||||||
// if rag - add as system message to avoid conflicts with tool usage
|
|
||||||
if !resume && cfg.RAGEnabled {
|
|
||||||
um := chatBody.Messages[len(chatBody.Messages)-1].Content
|
|
||||||
logger.Debug("LCPChat: RAG is enabled, preparing RAG context", "user_message", um)
|
|
||||||
ragResp, err := chatRagUse(um)
|
|
||||||
if err != nil {
|
|
||||||
logger.Error("LCPChat: failed to form a rag msg", "error", err)
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
logger.Debug("LCPChat: RAG response received",
|
|
||||||
"response_len", len(ragResp), "response_preview", ragResp[:min(len(ragResp), 100)])
|
|
||||||
// Use system role for RAG context to avoid conflicts with tool usage
|
|
||||||
ragMsg := models.RoleMsg{Role: "system", Content: RAGMsg + ragResp}
|
|
||||||
chatBody.Messages = append(chatBody.Messages, ragMsg)
|
|
||||||
logger.Debug("LCPChat: RAG message added to chat body", "role", ragMsg.Role,
|
|
||||||
"rag_content_len", len(ragMsg.Content), "message_count_after_rag", len(chatBody.Messages))
|
|
||||||
}
|
|
||||||
filteredMessages, _ := filterMessagesForCurrentCharacter(chatBody.Messages)
|
filteredMessages, _ := filterMessagesForCurrentCharacter(chatBody.Messages)
|
||||||
// openai /v1/chat does not support custom roles; needs to be user, assistant, system
|
// openai /v1/chat does not support custom roles; needs to be user, assistant, system
|
||||||
// Add persona suffix to the last user message to indicate who the assistant should reply as
|
// Add persona suffix to the last user message to indicate who the assistant should reply as
|
||||||
@@ -390,22 +352,6 @@ func (ds DeepSeekerCompletion) FormMsg(msg, role string, resume bool) (io.Reader
|
|||||||
newMsg = *processMessageTag(&newMsg)
|
newMsg = *processMessageTag(&newMsg)
|
||||||
chatBody.Messages = append(chatBody.Messages, newMsg)
|
chatBody.Messages = append(chatBody.Messages, newMsg)
|
||||||
}
|
}
|
||||||
// if rag - add as system message to avoid conflicts with tool usage
|
|
||||||
if !resume && cfg.RAGEnabled {
|
|
||||||
um := chatBody.Messages[len(chatBody.Messages)-1].Content
|
|
||||||
logger.Debug("DeepSeekerCompletion: RAG is enabled, preparing RAG context", "user_message", um)
|
|
||||||
ragResp, err := chatRagUse(um)
|
|
||||||
if err != nil {
|
|
||||||
logger.Error("DeepSeekerCompletion: failed to form a rag msg", "error", err)
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
logger.Debug("DeepSeekerCompletion: RAG response received",
|
|
||||||
"response_len", len(ragResp), "response_preview", ragResp[:min(len(ragResp), 100)])
|
|
||||||
// Use system role for RAG context to avoid conflicts with tool usage
|
|
||||||
ragMsg := models.RoleMsg{Role: "system", Content: RAGMsg + ragResp}
|
|
||||||
chatBody.Messages = append(chatBody.Messages, ragMsg)
|
|
||||||
logger.Debug("DeepSeekerCompletion: RAG message added to chat body", "message_count", len(chatBody.Messages))
|
|
||||||
}
|
|
||||||
// sending description of the tools and how to use them
|
// sending description of the tools and how to use them
|
||||||
if cfg.ToolUse && !resume && role == cfg.UserRole && !containsToolSysMsg() {
|
if cfg.ToolUse && !resume && role == cfg.UserRole && !containsToolSysMsg() {
|
||||||
chatBody.Messages = append(chatBody.Messages, models.RoleMsg{Role: cfg.ToolRole, Content: toolSysMsg})
|
chatBody.Messages = append(chatBody.Messages, models.RoleMsg{Role: cfg.ToolRole, Content: toolSysMsg})
|
||||||
@@ -421,9 +367,6 @@ func (ds DeepSeekerCompletion) FormMsg(msg, role string, resume bool) (io.Reader
|
|||||||
botMsgStart := "\n" + botPersona + ":\n"
|
botMsgStart := "\n" + botPersona + ":\n"
|
||||||
prompt += botMsgStart
|
prompt += botMsgStart
|
||||||
}
|
}
|
||||||
if cfg.ThinkUse && !cfg.ToolUse {
|
|
||||||
prompt += "<think>"
|
|
||||||
}
|
|
||||||
logger.Debug("checking prompt for /completion", "tool_use", cfg.ToolUse,
|
logger.Debug("checking prompt for /completion", "tool_use", cfg.ToolUse,
|
||||||
"msg", msg, "resume", resume, "prompt", prompt)
|
"msg", msg, "resume", resume, "prompt", prompt)
|
||||||
payload := models.NewDSCompletionReq(prompt, chatBody.Model,
|
payload := models.NewDSCompletionReq(prompt, chatBody.Model,
|
||||||
@@ -478,22 +421,6 @@ func (ds DeepSeekerChat) FormMsg(msg, role string, resume bool) (io.Reader, erro
|
|||||||
newMsg = *processMessageTag(&newMsg)
|
newMsg = *processMessageTag(&newMsg)
|
||||||
chatBody.Messages = append(chatBody.Messages, newMsg)
|
chatBody.Messages = append(chatBody.Messages, newMsg)
|
||||||
}
|
}
|
||||||
// if rag - add as system message to avoid conflicts with tool usage
|
|
||||||
if !resume && cfg.RAGEnabled {
|
|
||||||
um := chatBody.Messages[len(chatBody.Messages)-1].Content
|
|
||||||
logger.Debug("RAG is enabled, preparing RAG context", "user_message", um)
|
|
||||||
ragResp, err := chatRagUse(um)
|
|
||||||
if err != nil {
|
|
||||||
logger.Error("failed to form a rag msg", "error", err)
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
logger.Debug("RAG response received", "response_len", len(ragResp),
|
|
||||||
"response_preview", ragResp[:min(len(ragResp), 100)])
|
|
||||||
// Use system role for RAG context to avoid conflicts with tool usage
|
|
||||||
ragMsg := models.RoleMsg{Role: "system", Content: RAGMsg + ragResp}
|
|
||||||
chatBody.Messages = append(chatBody.Messages, ragMsg)
|
|
||||||
logger.Debug("RAG message added to chat body", "message_count", len(chatBody.Messages))
|
|
||||||
}
|
|
||||||
// Create copy of chat body with standardized user role
|
// Create copy of chat body with standardized user role
|
||||||
filteredMessages, _ := filterMessagesForCurrentCharacter(chatBody.Messages)
|
filteredMessages, _ := filterMessagesForCurrentCharacter(chatBody.Messages)
|
||||||
// Add persona suffix to the last user message to indicate who the assistant should reply as
|
// Add persona suffix to the last user message to indicate who the assistant should reply as
|
||||||
@@ -556,22 +483,6 @@ func (or OpenRouterCompletion) FormMsg(msg, role string, resume bool) (io.Reader
|
|||||||
newMsg = *processMessageTag(&newMsg)
|
newMsg = *processMessageTag(&newMsg)
|
||||||
chatBody.Messages = append(chatBody.Messages, newMsg)
|
chatBody.Messages = append(chatBody.Messages, newMsg)
|
||||||
}
|
}
|
||||||
// if rag - add as system message to avoid conflicts with tool usage
|
|
||||||
if !resume && cfg.RAGEnabled {
|
|
||||||
um := chatBody.Messages[len(chatBody.Messages)-1].Content
|
|
||||||
logger.Debug("RAG is enabled, preparing RAG context", "user_message", um)
|
|
||||||
ragResp, err := chatRagUse(um)
|
|
||||||
if err != nil {
|
|
||||||
logger.Error("failed to form a rag msg", "error", err)
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
logger.Debug("RAG response received", "response_len",
|
|
||||||
len(ragResp), "response_preview", ragResp[:min(len(ragResp), 100)])
|
|
||||||
// Use system role for RAG context to avoid conflicts with tool usage
|
|
||||||
ragMsg := models.RoleMsg{Role: "system", Content: RAGMsg + ragResp}
|
|
||||||
chatBody.Messages = append(chatBody.Messages, ragMsg)
|
|
||||||
logger.Debug("RAG message added to chat body", "message_count", len(chatBody.Messages))
|
|
||||||
}
|
|
||||||
// sending description of the tools and how to use them
|
// sending description of the tools and how to use them
|
||||||
if cfg.ToolUse && !resume && role == cfg.UserRole && !containsToolSysMsg() {
|
if cfg.ToolUse && !resume && role == cfg.UserRole && !containsToolSysMsg() {
|
||||||
chatBody.Messages = append(chatBody.Messages, models.RoleMsg{Role: cfg.ToolRole, Content: toolSysMsg})
|
chatBody.Messages = append(chatBody.Messages, models.RoleMsg{Role: cfg.ToolRole, Content: toolSysMsg})
|
||||||
@@ -587,9 +498,6 @@ func (or OpenRouterCompletion) FormMsg(msg, role string, resume bool) (io.Reader
|
|||||||
botMsgStart := "\n" + botPersona + ":\n"
|
botMsgStart := "\n" + botPersona + ":\n"
|
||||||
prompt += botMsgStart
|
prompt += botMsgStart
|
||||||
}
|
}
|
||||||
if cfg.ThinkUse && !cfg.ToolUse {
|
|
||||||
prompt += "<think>"
|
|
||||||
}
|
|
||||||
stopSlice := chatBody.MakeStopSliceExcluding("", listChatRoles())
|
stopSlice := chatBody.MakeStopSliceExcluding("", listChatRoles())
|
||||||
logger.Debug("checking prompt for /completion", "tool_use", cfg.ToolUse,
|
logger.Debug("checking prompt for /completion", "tool_use", cfg.ToolUse,
|
||||||
"msg", msg, "resume", resume, "prompt", prompt, "stop_strings", stopSlice)
|
"msg", msg, "resume", resume, "prompt", prompt, "stop_strings", stopSlice)
|
||||||
@@ -614,12 +522,14 @@ func (or OpenRouterChat) ParseChunk(data []byte) (*models.TextChunk, error) {
|
|||||||
logger.Error("failed to decode", "error", err, "line", string(data))
|
logger.Error("failed to decode", "error", err, "line", string(data))
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
lastChoice := llmchunk.Choices[len(llmchunk.Choices)-1]
|
||||||
resp := &models.TextChunk{
|
resp := &models.TextChunk{
|
||||||
Chunk: llmchunk.Choices[len(llmchunk.Choices)-1].Delta.Content,
|
Chunk: lastChoice.Delta.Content,
|
||||||
|
Reasoning: lastChoice.Delta.Reasoning,
|
||||||
}
|
}
|
||||||
// Handle tool calls similar to LCPChat
|
// Handle tool calls similar to LCPChat
|
||||||
if len(llmchunk.Choices[len(llmchunk.Choices)-1].Delta.ToolCalls) > 0 {
|
if len(lastChoice.Delta.ToolCalls) > 0 {
|
||||||
toolCall := llmchunk.Choices[len(llmchunk.Choices)-1].Delta.ToolCalls[0]
|
toolCall := lastChoice.Delta.ToolCalls[0]
|
||||||
resp.ToolChunk = toolCall.Function.Arguments
|
resp.ToolChunk = toolCall.Function.Arguments
|
||||||
fname := toolCall.Function.Name
|
fname := toolCall.Function.Name
|
||||||
if fname != "" {
|
if fname != "" {
|
||||||
@@ -631,7 +541,7 @@ func (or OpenRouterChat) ParseChunk(data []byte) (*models.TextChunk, error) {
|
|||||||
if resp.ToolChunk != "" {
|
if resp.ToolChunk != "" {
|
||||||
resp.ToolResp = true
|
resp.ToolResp = true
|
||||||
}
|
}
|
||||||
if llmchunk.Choices[len(llmchunk.Choices)-1].FinishReason == "stop" {
|
if lastChoice.FinishReason == "stop" {
|
||||||
if resp.Chunk != "" {
|
if resp.Chunk != "" {
|
||||||
logger.Error("text inside of finish llmchunk", "chunk", llmchunk)
|
logger.Error("text inside of finish llmchunk", "chunk", llmchunk)
|
||||||
}
|
}
|
||||||
@@ -675,22 +585,6 @@ func (or OpenRouterChat) FormMsg(msg, role string, resume bool) (io.Reader, erro
|
|||||||
newMsg = *processMessageTag(&newMsg)
|
newMsg = *processMessageTag(&newMsg)
|
||||||
chatBody.Messages = append(chatBody.Messages, newMsg)
|
chatBody.Messages = append(chatBody.Messages, newMsg)
|
||||||
}
|
}
|
||||||
// if rag - add as system message to avoid conflicts with tool usage
|
|
||||||
if !resume && cfg.RAGEnabled {
|
|
||||||
um := chatBody.Messages[len(chatBody.Messages)-1].Content
|
|
||||||
logger.Debug("RAG is enabled, preparing RAG context", "user_message", um)
|
|
||||||
ragResp, err := chatRagUse(um)
|
|
||||||
if err != nil {
|
|
||||||
logger.Error("failed to form a rag msg", "error", err)
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
logger.Debug("RAG response received", "response_len", len(ragResp),
|
|
||||||
"response_preview", ragResp[:min(len(ragResp), 100)])
|
|
||||||
// Use system role for RAG context to avoid conflicts with tool usage
|
|
||||||
ragMsg := models.RoleMsg{Role: "system", Content: RAGMsg + ragResp}
|
|
||||||
chatBody.Messages = append(chatBody.Messages, ragMsg)
|
|
||||||
logger.Debug("RAG message added to chat body", "message_count", len(chatBody.Messages))
|
|
||||||
}
|
|
||||||
// Create copy of chat body with standardized user role
|
// Create copy of chat body with standardized user role
|
||||||
filteredMessages, _ := filterMessagesForCurrentCharacter(chatBody.Messages)
|
filteredMessages, _ := filterMessagesForCurrentCharacter(chatBody.Messages)
|
||||||
// Add persona suffix to the last user message to indicate who the assistant should reply as
|
// Add persona suffix to the last user message to indicate who the assistant should reply as
|
||||||
@@ -710,7 +604,7 @@ func (or OpenRouterChat) FormMsg(msg, role string, resume bool) (io.Reader, erro
|
|||||||
}
|
}
|
||||||
// Clean null/empty messages to prevent API issues
|
// Clean null/empty messages to prevent API issues
|
||||||
bodyCopy.Messages = consolidateAssistantMessages(bodyCopy.Messages)
|
bodyCopy.Messages = consolidateAssistantMessages(bodyCopy.Messages)
|
||||||
orBody := models.NewOpenRouterChatReq(*bodyCopy, defaultLCPProps)
|
orBody := models.NewOpenRouterChatReq(*bodyCopy, defaultLCPProps, cfg.ReasoningEffort)
|
||||||
if cfg.ToolUse && !resume && role != cfg.ToolRole {
|
if cfg.ToolUse && !resume && role != cfg.ToolRole {
|
||||||
orBody.Tools = baseTools // set tools to use
|
orBody.Tools = baseTools // set tools to use
|
||||||
}
|
}
|
||||||
|
|||||||
3
main.go
3
main.go
@@ -7,13 +7,14 @@ import (
|
|||||||
var (
|
var (
|
||||||
boolColors = map[bool]string{true: "green", false: "red"}
|
boolColors = map[bool]string{true: "green", false: "red"}
|
||||||
botRespMode = false
|
botRespMode = false
|
||||||
|
toolRunningMode = false
|
||||||
editMode = false
|
editMode = false
|
||||||
roleEditMode = false
|
roleEditMode = false
|
||||||
injectRole = true
|
injectRole = true
|
||||||
selectedIndex = int(-1)
|
selectedIndex = int(-1)
|
||||||
shellMode = false
|
shellMode = false
|
||||||
thinkingCollapsed = false
|
thinkingCollapsed = false
|
||||||
indexLineCompletion = "F12 to show keys help | llm turn: [%s:-:b]%v[-:-:-] (F6) | chat: [orange:-:b]%s[-:-:-] (F1) | toolUseAdviced: [%s:-:b]%v[-:-:-] (ctrl+k) | model: [%s:-:b]%s[-:-:-] (ctrl+l) | skip LLM resp: [%s:-:b]%v[-:-:-] (F10)\nAPI: [orange:-:b]%s[-:-:-] (ctrl+v) | recording: [%s:-:b]%v[-:-:-] (ctrl+r) | writing as: [orange:-:b]%s[-:-:-] (ctrl+q) | bot will write as [orange:-:b]%s[-:-:-] (ctrl+x) | role injection (alt+7) [%s:-:b]%v[-:-:-]"
|
statusLineTempl = "help (F12) | [%s:-:b]llm writes[-:-:-] (F6 to interrupt) | chat: [orange:-:b]%s[-:-:-] (F1) | [%s:-:b]tool use[-:-:-] (ctrl+k) | model: [%s:-:b]%s[-:-:-] (ctrl+l) | [%s:-:b]skip LLM resp[-:-:-] (F10)\nAPI: [orange:-:b]%s[-:-:-] (ctrl+v) | writing as: [orange:-:b]%s[-:-:-] (ctrl+q) | bot will write as [orange:-:b]%s[-:-:-] (ctrl+x)"
|
||||||
focusSwitcher = map[tview.Primitive]tview.Primitive{}
|
focusSwitcher = map[tview.Primitive]tview.Primitive{}
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|||||||
@@ -65,6 +65,7 @@ type LLMRespChunk struct {
|
|||||||
Index int `json:"index"`
|
Index int `json:"index"`
|
||||||
Delta struct {
|
Delta struct {
|
||||||
Content string `json:"content"`
|
Content string `json:"content"`
|
||||||
|
ReasoningContent string `json:"reasoning_content"`
|
||||||
ToolCalls []ToolDeltaResp `json:"tool_calls"`
|
ToolCalls []ToolDeltaResp `json:"tool_calls"`
|
||||||
} `json:"delta"`
|
} `json:"delta"`
|
||||||
} `json:"choices"`
|
} `json:"choices"`
|
||||||
@@ -86,6 +87,7 @@ type TextChunk struct {
|
|||||||
ToolResp bool
|
ToolResp bool
|
||||||
FuncName string
|
FuncName string
|
||||||
ToolID string
|
ToolID string
|
||||||
|
Reasoning string // For models that send reasoning separately (OpenRouter, etc.)
|
||||||
}
|
}
|
||||||
|
|
||||||
type TextContentPart struct {
|
type TextContentPart struct {
|
||||||
@@ -108,6 +110,7 @@ type RoleMsg struct {
|
|||||||
ContentParts []any `json:"-"`
|
ContentParts []any `json:"-"`
|
||||||
ToolCallID string `json:"tool_call_id,omitempty"` // For tool response messages
|
ToolCallID string `json:"tool_call_id,omitempty"` // For tool response messages
|
||||||
KnownTo []string `json:"known_to,omitempty"`
|
KnownTo []string `json:"known_to,omitempty"`
|
||||||
|
Stats *ResponseStats `json:"stats"`
|
||||||
hasContentParts bool // Flag to indicate which content type to marshal
|
hasContentParts bool // Flag to indicate which content type to marshal
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -181,13 +184,11 @@ func (m *RoleMsg) UnmarshalJSON(data []byte) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (m *RoleMsg) ToText(i int) string {
|
func (m *RoleMsg) ToText(i int) string {
|
||||||
// Convert content to string representation
|
|
||||||
var contentStr string
|
var contentStr string
|
||||||
var imageIndicators []string
|
var imageIndicators []string
|
||||||
if !m.hasContentParts {
|
if !m.hasContentParts {
|
||||||
contentStr = m.Content
|
contentStr = m.Content
|
||||||
} else {
|
} else {
|
||||||
// For structured content, collect text parts and image indicators
|
|
||||||
var textParts []string
|
var textParts []string
|
||||||
for _, part := range m.ContentParts {
|
for _, part := range m.ContentParts {
|
||||||
switch p := part.(type) {
|
switch p := part.(type) {
|
||||||
@@ -196,7 +197,6 @@ func (m *RoleMsg) ToText(i int) string {
|
|||||||
textParts = append(textParts, p.Text)
|
textParts = append(textParts, p.Text)
|
||||||
}
|
}
|
||||||
case ImageContentPart:
|
case ImageContentPart:
|
||||||
// Collect image indicator
|
|
||||||
displayPath := p.Path
|
displayPath := p.Path
|
||||||
if displayPath == "" {
|
if displayPath == "" {
|
||||||
displayPath = "image"
|
displayPath = "image"
|
||||||
@@ -214,7 +214,6 @@ func (m *RoleMsg) ToText(i int) string {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
case "image_url":
|
case "image_url":
|
||||||
// Handle unmarshaled image content
|
|
||||||
var displayPath string
|
var displayPath string
|
||||||
if pathVal, pathExists := p["path"]; pathExists {
|
if pathVal, pathExists := p["path"]; pathExists {
|
||||||
if pathStr, isStr := pathVal.(string); isStr && pathStr != "" {
|
if pathStr, isStr := pathVal.(string); isStr && pathStr != "" {
|
||||||
@@ -231,23 +230,19 @@ func (m *RoleMsg) ToText(i int) string {
|
|||||||
}
|
}
|
||||||
contentStr = strings.Join(textParts, " ") + " "
|
contentStr = strings.Join(textParts, " ") + " "
|
||||||
}
|
}
|
||||||
// check if already has role annotation (/completion makes them)
|
|
||||||
// in that case remove it, and then add to icon
|
|
||||||
// since icon and content are separated by \n
|
|
||||||
contentStr, _ = strings.CutPrefix(contentStr, m.Role+":")
|
contentStr, _ = strings.CutPrefix(contentStr, m.Role+":")
|
||||||
// if !strings.HasPrefix(contentStr, m.Role+":") {
|
|
||||||
icon := fmt.Sprintf("(%d) <%s>: ", i, m.Role)
|
icon := fmt.Sprintf("(%d) <%s>: ", i, m.Role)
|
||||||
// }
|
|
||||||
// Build final message with image indicators before text
|
|
||||||
var finalContent strings.Builder
|
var finalContent strings.Builder
|
||||||
if len(imageIndicators) > 0 {
|
if len(imageIndicators) > 0 {
|
||||||
// Add each image indicator on its own line
|
|
||||||
for _, indicator := range imageIndicators {
|
for _, indicator := range imageIndicators {
|
||||||
finalContent.WriteString(indicator)
|
finalContent.WriteString(indicator)
|
||||||
finalContent.WriteString("\n")
|
finalContent.WriteString("\n")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
finalContent.WriteString(contentStr)
|
finalContent.WriteString(contentStr)
|
||||||
|
if m.Stats != nil {
|
||||||
|
fmt.Fprintf(&finalContent, "\n[gray::i][%d tok, %.1fs, %.1f t/s][-:-:-]", m.Stats.Tokens, m.Stats.Duration, m.Stats.TokensPerSec)
|
||||||
|
}
|
||||||
textMsg := fmt.Sprintf("[-:-:b]%s[-:-:-]\n%s\n", icon, finalContent.String())
|
textMsg := fmt.Sprintf("[-:-:b]%s[-:-:-]\n%s\n", icon, finalContent.String())
|
||||||
return strings.ReplaceAll(textMsg, "\n\n", "\n")
|
return strings.ReplaceAll(textMsg, "\n\n", "\n")
|
||||||
}
|
}
|
||||||
@@ -329,10 +324,71 @@ func (m *RoleMsg) Copy() RoleMsg {
|
|||||||
ContentParts: m.ContentParts,
|
ContentParts: m.ContentParts,
|
||||||
ToolCallID: m.ToolCallID,
|
ToolCallID: m.ToolCallID,
|
||||||
KnownTo: m.KnownTo,
|
KnownTo: m.KnownTo,
|
||||||
|
Stats: m.Stats,
|
||||||
hasContentParts: m.hasContentParts,
|
hasContentParts: m.hasContentParts,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// GetText returns the text content of the message, handling both
|
||||||
|
// simple Content and multimodal ContentParts formats.
|
||||||
|
func (m *RoleMsg) GetText() string {
|
||||||
|
if !m.hasContentParts {
|
||||||
|
return m.Content
|
||||||
|
}
|
||||||
|
var textParts []string
|
||||||
|
for _, part := range m.ContentParts {
|
||||||
|
switch p := part.(type) {
|
||||||
|
case TextContentPart:
|
||||||
|
if p.Type == "text" {
|
||||||
|
textParts = append(textParts, p.Text)
|
||||||
|
}
|
||||||
|
case map[string]any:
|
||||||
|
if partType, exists := p["type"]; exists {
|
||||||
|
if partType == "text" {
|
||||||
|
if textVal, textExists := p["text"]; textExists {
|
||||||
|
if textStr, isStr := textVal.(string); isStr {
|
||||||
|
textParts = append(textParts, textStr)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return strings.Join(textParts, " ")
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetText updates the text content of the message. If the message has
|
||||||
|
// ContentParts (multimodal), it updates the text parts while preserving
|
||||||
|
// images. If not, it sets the simple Content field.
|
||||||
|
func (m *RoleMsg) SetText(text string) {
|
||||||
|
if !m.hasContentParts {
|
||||||
|
m.Content = text
|
||||||
|
return
|
||||||
|
}
|
||||||
|
var newParts []any
|
||||||
|
for _, part := range m.ContentParts {
|
||||||
|
switch p := part.(type) {
|
||||||
|
case TextContentPart:
|
||||||
|
if p.Type == "text" {
|
||||||
|
p.Text = text
|
||||||
|
newParts = append(newParts, p)
|
||||||
|
} else {
|
||||||
|
newParts = append(newParts, p)
|
||||||
|
}
|
||||||
|
case map[string]any:
|
||||||
|
if partType, exists := p["type"]; exists && partType == "text" {
|
||||||
|
p["text"] = text
|
||||||
|
newParts = append(newParts, p)
|
||||||
|
} else {
|
||||||
|
newParts = append(newParts, p)
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
newParts = append(newParts, part)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
m.ContentParts = newParts
|
||||||
|
}
|
||||||
|
|
||||||
// AddTextPart adds a text content part to the message
|
// AddTextPart adds a text content part to the message
|
||||||
func (m *RoleMsg) AddTextPart(text string) {
|
func (m *RoleMsg) AddTextPart(text string) {
|
||||||
if !m.hasContentParts {
|
if !m.hasContentParts {
|
||||||
@@ -344,7 +400,6 @@ func (m *RoleMsg) AddTextPart(text string) {
|
|||||||
}
|
}
|
||||||
m.hasContentParts = true
|
m.hasContentParts = true
|
||||||
}
|
}
|
||||||
|
|
||||||
textPart := TextContentPart{Type: "text", Text: text}
|
textPart := TextContentPart{Type: "text", Text: text}
|
||||||
m.ContentParts = append(m.ContentParts, textPart)
|
m.ContentParts = append(m.ContentParts, textPart)
|
||||||
}
|
}
|
||||||
@@ -360,7 +415,6 @@ func (m *RoleMsg) AddImagePart(imageURL, imagePath string) {
|
|||||||
}
|
}
|
||||||
m.hasContentParts = true
|
m.hasContentParts = true
|
||||||
}
|
}
|
||||||
|
|
||||||
imagePart := ImageContentPart{
|
imagePart := ImageContentPart{
|
||||||
Type: "image_url",
|
Type: "image_url",
|
||||||
Path: imagePath, // Store the original file path
|
Path: imagePath, // Store the original file path
|
||||||
@@ -641,6 +695,12 @@ func (lcp *LCPModels) ListModels() []string {
|
|||||||
return resp
|
return resp
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type ResponseStats struct {
|
||||||
|
Tokens int
|
||||||
|
Duration float64
|
||||||
|
TokensPerSec float64
|
||||||
|
}
|
||||||
|
|
||||||
type ChatRoundReq struct {
|
type ChatRoundReq struct {
|
||||||
UserMsg string
|
UserMsg string
|
||||||
Role string
|
Role string
|
||||||
|
|||||||
@@ -1,10 +1,8 @@
|
|||||||
package models
|
package models
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"strings"
|
"strings"
|
||||||
"testing"
|
"testing"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestRoleMsgToTextWithImages(t *testing.T) {
|
func TestRoleMsgToTextWithImages(t *testing.T) {
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
name string
|
name string
|
||||||
@@ -92,7 +90,6 @@ func TestRoleMsgToTextWithImages(t *testing.T) {
|
|||||||
expected: "[orange::i][image: /old/path/photo.jpg][-:-:-]",
|
expected: "[orange::i][image: /old/path/photo.jpg][-:-:-]",
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tt := range tests {
|
for _, tt := range tests {
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
result := tt.msg.ToText(tt.index)
|
result := tt.msg.ToText(tt.index)
|
||||||
@@ -110,12 +107,10 @@ func TestRoleMsgToTextWithImages(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestExtractDisplayPath(t *testing.T) {
|
func TestExtractDisplayPath(t *testing.T) {
|
||||||
// Save original base dir
|
// Save original base dir
|
||||||
originalBaseDir := imageBaseDir
|
originalBaseDir := imageBaseDir
|
||||||
defer func() { imageBaseDir = originalBaseDir }()
|
defer func() { imageBaseDir = originalBaseDir }()
|
||||||
|
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
name string
|
name string
|
||||||
baseDir string
|
baseDir string
|
||||||
@@ -153,7 +148,6 @@ func TestExtractDisplayPath(t *testing.T) {
|
|||||||
expected: "..._that_exceeds_sixty_characters_limit_yes_it_is_very_long.jpg",
|
expected: "..._that_exceeds_sixty_characters_limit_yes_it_is_very_long.jpg",
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, tt := range tests {
|
for _, tt := range tests {
|
||||||
t.Run(tt.name, func(t *testing.T) {
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
imageBaseDir = tt.baseDir
|
imageBaseDir = tt.baseDir
|
||||||
|
|||||||
@@ -32,10 +32,16 @@ type OpenRouterChatReq struct {
|
|||||||
MinP float32 `json:"min_p"`
|
MinP float32 `json:"min_p"`
|
||||||
NPredict int32 `json:"max_tokens"`
|
NPredict int32 `json:"max_tokens"`
|
||||||
Tools []Tool `json:"tools"`
|
Tools []Tool `json:"tools"`
|
||||||
|
Reasoning *ReasoningConfig `json:"reasoning,omitempty"`
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewOpenRouterChatReq(cb ChatBody, props map[string]float32) OpenRouterChatReq {
|
type ReasoningConfig struct {
|
||||||
return OpenRouterChatReq{
|
Effort string `json:"effort,omitempty"` // xhigh, high, medium, low, minimal, none
|
||||||
|
Summary string `json:"summary,omitempty"` // auto, concise, detailed
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewOpenRouterChatReq(cb ChatBody, props map[string]float32, reasoningEffort string) OpenRouterChatReq {
|
||||||
|
req := OpenRouterChatReq{
|
||||||
Messages: cb.Messages,
|
Messages: cb.Messages,
|
||||||
Model: cb.Model,
|
Model: cb.Model,
|
||||||
Stream: cb.Stream,
|
Stream: cb.Stream,
|
||||||
@@ -43,6 +49,13 @@ func NewOpenRouterChatReq(cb ChatBody, props map[string]float32) OpenRouterChatR
|
|||||||
MinP: props["min_p"],
|
MinP: props["min_p"],
|
||||||
NPredict: int32(props["n_predict"]),
|
NPredict: int32(props["n_predict"]),
|
||||||
}
|
}
|
||||||
|
// Only include reasoning config if effort is specified and not "none"
|
||||||
|
if reasoningEffort != "" && reasoningEffort != "none" {
|
||||||
|
req.Reasoning = &ReasoningConfig{
|
||||||
|
Effort: reasoningEffort,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return req
|
||||||
}
|
}
|
||||||
|
|
||||||
type OpenRouterChatRespNonStream struct {
|
type OpenRouterChatRespNonStream struct {
|
||||||
@@ -82,6 +95,7 @@ type OpenRouterChatResp struct {
|
|||||||
Delta struct {
|
Delta struct {
|
||||||
Role string `json:"role"`
|
Role string `json:"role"`
|
||||||
Content string `json:"content"`
|
Content string `json:"content"`
|
||||||
|
Reasoning string `json:"reasoning"`
|
||||||
ToolCalls []ToolDeltaResp `json:"tool_calls"`
|
ToolCalls []ToolDeltaResp `json:"tool_calls"`
|
||||||
} `json:"delta"`
|
} `json:"delta"`
|
||||||
FinishReason string `json:"finish_reason"`
|
FinishReason string `json:"finish_reason"`
|
||||||
|
|||||||
@@ -62,7 +62,6 @@ func TestORModelsListModels(t *testing.T) {
|
|||||||
t.Errorf("expected 4 total models, got %d", len(allModels))
|
t.Errorf("expected 4 total models, got %d", len(allModels))
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
|
||||||
t.Run("integration with or_models.json", func(t *testing.T) {
|
t.Run("integration with or_models.json", func(t *testing.T) {
|
||||||
// Attempt to load the real data file from the project root
|
// Attempt to load the real data file from the project root
|
||||||
path := filepath.Join("..", "or_models.json")
|
path := filepath.Join("..", "or_models.json")
|
||||||
|
|||||||
151
popups.go
151
popups.go
@@ -17,10 +17,14 @@ func showModelSelectionPopup() {
|
|||||||
} else if strings.Contains(api, "openrouter.ai") {
|
} else if strings.Contains(api, "openrouter.ai") {
|
||||||
return ORFreeModels
|
return ORFreeModels
|
||||||
}
|
}
|
||||||
// Assume local llama.cpp
|
// Assume local llama.cpp - fetch with load status
|
||||||
updateModelLists()
|
models, err := fetchLCPModelsWithLoadStatus()
|
||||||
|
if err != nil {
|
||||||
|
logger.Error("failed to fetch models with load status", "error", err)
|
||||||
return LocalModels
|
return LocalModels
|
||||||
}
|
}
|
||||||
|
return models
|
||||||
|
}
|
||||||
// Get the current model list based on the API
|
// Get the current model list based on the API
|
||||||
modelList := getModelListForAPI(cfg.CurrentAPI)
|
modelList := getModelListForAPI(cfg.CurrentAPI)
|
||||||
// Check for empty options list
|
// Check for empty options list
|
||||||
@@ -47,7 +51,7 @@ func showModelSelectionPopup() {
|
|||||||
// Find the current model index to set as selected
|
// Find the current model index to set as selected
|
||||||
currentModelIndex := -1
|
currentModelIndex := -1
|
||||||
for i, model := range modelList {
|
for i, model := range modelList {
|
||||||
if model == chatBody.Model {
|
if strings.TrimPrefix(model, "(loaded) ") == chatBody.Model {
|
||||||
currentModelIndex = i
|
currentModelIndex = i
|
||||||
}
|
}
|
||||||
modelListWidget.AddItem(model, "", 0, nil)
|
modelListWidget.AddItem(model, "", 0, nil)
|
||||||
@@ -57,21 +61,23 @@ func showModelSelectionPopup() {
|
|||||||
modelListWidget.SetCurrentItem(currentModelIndex)
|
modelListWidget.SetCurrentItem(currentModelIndex)
|
||||||
}
|
}
|
||||||
modelListWidget.SetSelectedFunc(func(index int, mainText string, secondaryText string, shortcut rune) {
|
modelListWidget.SetSelectedFunc(func(index int, mainText string, secondaryText string, shortcut rune) {
|
||||||
// Update the model in both chatBody and config
|
modelName := strings.TrimPrefix(mainText, "(loaded) ")
|
||||||
chatBody.Model = mainText
|
chatBody.Model = modelName
|
||||||
cfg.CurrentModel = chatBody.Model
|
cfg.CurrentModel = chatBody.Model
|
||||||
// Remove the popup page
|
|
||||||
pages.RemovePage("modelSelectionPopup")
|
pages.RemovePage("modelSelectionPopup")
|
||||||
// Update the status line to reflect the change
|
app.SetFocus(textArea)
|
||||||
|
updateCachedModelColor()
|
||||||
updateStatusLine()
|
updateStatusLine()
|
||||||
})
|
})
|
||||||
modelListWidget.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
modelListWidget.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
||||||
if event.Key() == tcell.KeyEscape {
|
if event.Key() == tcell.KeyEscape {
|
||||||
pages.RemovePage("modelSelectionPopup")
|
pages.RemovePage("modelSelectionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyRune && event.Rune() == 'x' {
|
if event.Key() == tcell.KeyRune && event.Rune() == 'x' {
|
||||||
pages.RemovePage("modelSelectionPopup")
|
pages.RemovePage("modelSelectionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
return event
|
return event
|
||||||
@@ -156,19 +162,21 @@ func showAPILinkSelectionPopup() {
|
|||||||
chatBody.Model = newModelList[0]
|
chatBody.Model = newModelList[0]
|
||||||
cfg.CurrentModel = chatBody.Model
|
cfg.CurrentModel = chatBody.Model
|
||||||
}
|
}
|
||||||
// Remove the popup page
|
|
||||||
pages.RemovePage("apiLinkSelectionPopup")
|
pages.RemovePage("apiLinkSelectionPopup")
|
||||||
// Update the parser and status line to reflect the change
|
app.SetFocus(textArea)
|
||||||
choseChunkParser()
|
choseChunkParser()
|
||||||
|
updateCachedModelColor()
|
||||||
updateStatusLine()
|
updateStatusLine()
|
||||||
})
|
})
|
||||||
apiListWidget.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
apiListWidget.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
||||||
if event.Key() == tcell.KeyEscape {
|
if event.Key() == tcell.KeyEscape {
|
||||||
pages.RemovePage("apiLinkSelectionPopup")
|
pages.RemovePage("apiLinkSelectionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyRune && event.Rune() == 'x' {
|
if event.Key() == tcell.KeyRune && event.Rune() == 'x' {
|
||||||
pages.RemovePage("apiLinkSelectionPopup")
|
pages.RemovePage("apiLinkSelectionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
return event
|
return event
|
||||||
@@ -228,6 +236,7 @@ func showUserRoleSelectionPopup() {
|
|||||||
textView.SetText(chatToText(filtered, cfg.ShowSys))
|
textView.SetText(chatToText(filtered, cfg.ShowSys))
|
||||||
// Remove the popup page
|
// Remove the popup page
|
||||||
pages.RemovePage("userRoleSelectionPopup")
|
pages.RemovePage("userRoleSelectionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
// Update the status line to reflect the change
|
// Update the status line to reflect the change
|
||||||
updateStatusLine()
|
updateStatusLine()
|
||||||
colorText()
|
colorText()
|
||||||
@@ -235,10 +244,12 @@ func showUserRoleSelectionPopup() {
|
|||||||
roleListWidget.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
roleListWidget.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
||||||
if event.Key() == tcell.KeyEscape {
|
if event.Key() == tcell.KeyEscape {
|
||||||
pages.RemovePage("userRoleSelectionPopup")
|
pages.RemovePage("userRoleSelectionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyRune && event.Rune() == 'x' {
|
if event.Key() == tcell.KeyRune && event.Rune() == 'x' {
|
||||||
pages.RemovePage("userRoleSelectionPopup")
|
pages.RemovePage("userRoleSelectionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
return event
|
return event
|
||||||
@@ -301,16 +312,19 @@ func showBotRoleSelectionPopup() {
|
|||||||
cfg.WriteNextMsgAsCompletionAgent = mainText
|
cfg.WriteNextMsgAsCompletionAgent = mainText
|
||||||
// Remove the popup page
|
// Remove the popup page
|
||||||
pages.RemovePage("botRoleSelectionPopup")
|
pages.RemovePage("botRoleSelectionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
// Update the status line to reflect the change
|
// Update the status line to reflect the change
|
||||||
updateStatusLine()
|
updateStatusLine()
|
||||||
})
|
})
|
||||||
roleListWidget.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
roleListWidget.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
||||||
if event.Key() == tcell.KeyEscape {
|
if event.Key() == tcell.KeyEscape {
|
||||||
pages.RemovePage("botRoleSelectionPopup")
|
pages.RemovePage("botRoleSelectionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyRune && event.Rune() == 'x' {
|
if event.Key() == tcell.KeyRune && event.Rune() == 'x' {
|
||||||
pages.RemovePage("botRoleSelectionPopup")
|
pages.RemovePage("botRoleSelectionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
return event
|
return event
|
||||||
@@ -330,7 +344,7 @@ func showBotRoleSelectionPopup() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func showFileCompletionPopup(filter string) {
|
func showFileCompletionPopup(filter string) {
|
||||||
baseDir := cfg.CodingDir
|
baseDir := cfg.FilePickerDir
|
||||||
if baseDir == "" {
|
if baseDir == "" {
|
||||||
baseDir = "."
|
baseDir = "."
|
||||||
}
|
}
|
||||||
@@ -362,14 +376,17 @@ func showFileCompletionPopup(filter string) {
|
|||||||
textArea.SetText(before+mainText, true)
|
textArea.SetText(before+mainText, true)
|
||||||
}
|
}
|
||||||
pages.RemovePage("fileCompletionPopup")
|
pages.RemovePage("fileCompletionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
})
|
})
|
||||||
widget.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
widget.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
||||||
if event.Key() == tcell.KeyEscape {
|
if event.Key() == tcell.KeyEscape {
|
||||||
pages.RemovePage("fileCompletionPopup")
|
pages.RemovePage("fileCompletionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyRune && event.Rune() == 'x' {
|
if event.Key() == tcell.KeyRune && event.Rune() == 'x' {
|
||||||
pages.RemovePage("fileCompletionPopup")
|
pages.RemovePage("fileCompletionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
return event
|
return event
|
||||||
@@ -387,3 +404,117 @@ func showFileCompletionPopup(filter string) {
|
|||||||
pages.AddPage("fileCompletionPopup", modal(widget, 80, 20), true, true)
|
pages.AddPage("fileCompletionPopup", modal(widget, 80, 20), true, true)
|
||||||
app.SetFocus(widget)
|
app.SetFocus(widget)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func updateWidgetColors(theme *tview.Theme) {
|
||||||
|
bgColor := theme.PrimitiveBackgroundColor
|
||||||
|
fgColor := theme.PrimaryTextColor
|
||||||
|
borderColor := theme.BorderColor
|
||||||
|
titleColor := theme.TitleColor
|
||||||
|
textView.SetBackgroundColor(bgColor)
|
||||||
|
textView.SetTextColor(fgColor)
|
||||||
|
textView.SetBorderColor(borderColor)
|
||||||
|
textView.SetTitleColor(titleColor)
|
||||||
|
textArea.SetBackgroundColor(bgColor)
|
||||||
|
textArea.SetBorderColor(borderColor)
|
||||||
|
textArea.SetTitleColor(titleColor)
|
||||||
|
textArea.SetTextStyle(tcell.StyleDefault.Background(bgColor).Foreground(fgColor))
|
||||||
|
textArea.SetPlaceholderStyle(tcell.StyleDefault.Background(bgColor).Foreground(fgColor))
|
||||||
|
textArea.SetText(textArea.GetText(), true)
|
||||||
|
editArea.SetBackgroundColor(bgColor)
|
||||||
|
editArea.SetBorderColor(borderColor)
|
||||||
|
editArea.SetTitleColor(titleColor)
|
||||||
|
editArea.SetTextStyle(tcell.StyleDefault.Background(bgColor).Foreground(fgColor))
|
||||||
|
editArea.SetPlaceholderStyle(tcell.StyleDefault.Background(bgColor).Foreground(fgColor))
|
||||||
|
editArea.SetText(editArea.GetText(), true)
|
||||||
|
statusLineWidget.SetBackgroundColor(bgColor)
|
||||||
|
statusLineWidget.SetTextColor(fgColor)
|
||||||
|
statusLineWidget.SetBorderColor(borderColor)
|
||||||
|
statusLineWidget.SetTitleColor(titleColor)
|
||||||
|
helpView.SetBackgroundColor(bgColor)
|
||||||
|
helpView.SetTextColor(fgColor)
|
||||||
|
helpView.SetBorderColor(borderColor)
|
||||||
|
helpView.SetTitleColor(titleColor)
|
||||||
|
searchField.SetBackgroundColor(bgColor)
|
||||||
|
searchField.SetBorderColor(borderColor)
|
||||||
|
searchField.SetTitleColor(titleColor)
|
||||||
|
}
|
||||||
|
|
||||||
|
// showColorschemeSelectionPopup creates a modal popup to select a colorscheme
|
||||||
|
func showColorschemeSelectionPopup() {
|
||||||
|
// Get the list of available colorschemes
|
||||||
|
schemeNames := make([]string, 0, len(colorschemes))
|
||||||
|
for name := range colorschemes {
|
||||||
|
schemeNames = append(schemeNames, name)
|
||||||
|
}
|
||||||
|
slices.Sort(schemeNames)
|
||||||
|
// Check for empty options list
|
||||||
|
if len(schemeNames) == 0 {
|
||||||
|
logger.Warn("no colorschemes available for selection")
|
||||||
|
message := "No colorschemes available."
|
||||||
|
if err := notifyUser("Empty list", message); err != nil {
|
||||||
|
logger.Error("failed to send notification", "error", err)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
// Create a list primitive
|
||||||
|
schemeListWidget := tview.NewList().ShowSecondaryText(false).
|
||||||
|
SetSelectedBackgroundColor(tcell.ColorGray)
|
||||||
|
schemeListWidget.SetTitle("Select Colorscheme").SetBorder(true)
|
||||||
|
currentScheme := "default"
|
||||||
|
for name := range colorschemes {
|
||||||
|
if tview.Styles == colorschemes[name] {
|
||||||
|
currentScheme = name
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
currentSchemeIndex := -1
|
||||||
|
for i, scheme := range schemeNames {
|
||||||
|
if scheme == currentScheme {
|
||||||
|
currentSchemeIndex = i
|
||||||
|
}
|
||||||
|
schemeListWidget.AddItem(scheme, "", 0, nil)
|
||||||
|
}
|
||||||
|
// Set the current selection if found
|
||||||
|
if currentSchemeIndex != -1 {
|
||||||
|
schemeListWidget.SetCurrentItem(currentSchemeIndex)
|
||||||
|
}
|
||||||
|
schemeListWidget.SetSelectedFunc(func(index int, mainText string, secondaryText string, shortcut rune) {
|
||||||
|
// Update the colorscheme
|
||||||
|
if theme, ok := colorschemes[mainText]; ok {
|
||||||
|
tview.Styles = theme
|
||||||
|
go func() {
|
||||||
|
app.QueueUpdateDraw(func() {
|
||||||
|
updateWidgetColors(&theme)
|
||||||
|
})
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
// Remove the popup page
|
||||||
|
pages.RemovePage("colorschemeSelectionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
|
})
|
||||||
|
schemeListWidget.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
||||||
|
if event.Key() == tcell.KeyEscape {
|
||||||
|
pages.RemovePage("colorschemeSelectionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
if event.Key() == tcell.KeyRune && event.Rune() == 'x' {
|
||||||
|
pages.RemovePage("colorschemeSelectionPopup")
|
||||||
|
app.SetFocus(textArea)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return event
|
||||||
|
})
|
||||||
|
modal := func(p tview.Primitive, width, height int) tview.Primitive {
|
||||||
|
return tview.NewFlex().
|
||||||
|
AddItem(nil, 0, 1, false).
|
||||||
|
AddItem(tview.NewFlex().SetDirection(tview.FlexRow).
|
||||||
|
AddItem(nil, 0, 1, false).
|
||||||
|
AddItem(p, height, 1, true).
|
||||||
|
AddItem(nil, 0, 1, false), width, 1, true).
|
||||||
|
AddItem(nil, 0, 1, false)
|
||||||
|
}
|
||||||
|
// Add modal page and make it visible
|
||||||
|
pages.AddPage("colorschemeSelectionPopup", modal(schemeListWidget, 40, len(schemeNames)+2), true, true)
|
||||||
|
app.SetFocus(schemeListWidget)
|
||||||
|
}
|
||||||
|
|||||||
@@ -115,9 +115,6 @@ func makePropsTable(props map[string]float32) *tview.Table {
|
|||||||
row++
|
row++
|
||||||
}
|
}
|
||||||
// Add checkboxes
|
// Add checkboxes
|
||||||
addCheckboxRow("Insert <think> tag (/completion only)", cfg.ThinkUse, func(checked bool) {
|
|
||||||
cfg.ThinkUse = checked
|
|
||||||
})
|
|
||||||
addCheckboxRow("RAG use", cfg.RAGEnabled, func(checked bool) {
|
addCheckboxRow("RAG use", cfg.RAGEnabled, func(checked bool) {
|
||||||
cfg.RAGEnabled = checked
|
cfg.RAGEnabled = checked
|
||||||
})
|
})
|
||||||
@@ -149,6 +146,11 @@ func makePropsTable(props map[string]float32) *tview.Table {
|
|||||||
addListPopupRow("Set log level", logLevels, GetLogLevel(), func(option string) {
|
addListPopupRow("Set log level", logLevels, GetLogLevel(), func(option string) {
|
||||||
setLogLevel(option)
|
setLogLevel(option)
|
||||||
})
|
})
|
||||||
|
// Add reasoning effort dropdown (for OpenRouter and supported APIs)
|
||||||
|
reasoningEfforts := []string{"", "none", "minimal", "low", "medium", "high", "xhigh"}
|
||||||
|
addListPopupRow("Reasoning effort (OR)", reasoningEfforts, cfg.ReasoningEffort, func(option string) {
|
||||||
|
cfg.ReasoningEffort = option
|
||||||
|
})
|
||||||
// Helper function to get model list for a given API
|
// Helper function to get model list for a given API
|
||||||
getModelListForAPI := func(api string) []string {
|
getModelListForAPI := func(api string) []string {
|
||||||
if strings.Contains(api, "api.deepseek.com/") {
|
if strings.Contains(api, "api.deepseek.com/") {
|
||||||
|
|||||||
@@ -131,7 +131,6 @@ func (a *APIEmbedder) EmbedSlice(lines []string) ([][]float32, error) {
|
|||||||
}
|
}
|
||||||
embeddings[data.Index] = data.Embedding
|
embeddings[data.Index] = data.Embedding
|
||||||
}
|
}
|
||||||
|
|
||||||
return embeddings, nil
|
return embeddings, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
181
rag/extractors.go
Normal file
181
rag/extractors.go
Normal file
@@ -0,0 +1,181 @@
|
|||||||
|
package rag
|
||||||
|
|
||||||
|
import (
|
||||||
|
"archive/zip"
|
||||||
|
"bytes"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"os"
|
||||||
|
"os/exec"
|
||||||
|
"path"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/PuerkitoBio/goquery"
|
||||||
|
"github.com/ledongthuc/pdf"
|
||||||
|
"github.com/yuin/goldmark"
|
||||||
|
"github.com/yuin/goldmark/extension"
|
||||||
|
"github.com/yuin/goldmark/parser"
|
||||||
|
"github.com/yuin/goldmark/renderer/html"
|
||||||
|
)
|
||||||
|
|
||||||
|
func ExtractText(fpath string) (string, error) {
|
||||||
|
ext := strings.ToLower(path.Ext(fpath))
|
||||||
|
switch ext {
|
||||||
|
case ".txt":
|
||||||
|
return extractTextFromFile(fpath)
|
||||||
|
case ".md", ".markdown":
|
||||||
|
return extractTextFromMarkdown(fpath)
|
||||||
|
case ".html", ".htm":
|
||||||
|
return extractTextFromHtmlFile(fpath)
|
||||||
|
case ".epub":
|
||||||
|
return extractTextFromEpub(fpath)
|
||||||
|
case ".pdf":
|
||||||
|
return extractTextFromPdf(fpath)
|
||||||
|
default:
|
||||||
|
return "", fmt.Errorf("unsupported file format: %s", ext)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func extractTextFromFile(fpath string) (string, error) {
|
||||||
|
data, err := os.ReadFile(fpath)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
return string(data), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func extractTextFromHtmlFile(fpath string) (string, error) {
|
||||||
|
data, err := os.ReadFile(fpath)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
return extractTextFromHtmlContent(data)
|
||||||
|
}
|
||||||
|
|
||||||
|
// non utf-8 encoding?
|
||||||
|
func extractTextFromHtmlContent(data []byte) (string, error) {
|
||||||
|
doc, err := goquery.NewDocumentFromReader(bytes.NewReader(data))
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
// Remove script and style tags
|
||||||
|
doc.Find("script, style, noscript").Each(func(i int, s *goquery.Selection) {
|
||||||
|
s.Remove()
|
||||||
|
})
|
||||||
|
// Get text and clean it
|
||||||
|
text := doc.Text()
|
||||||
|
// Collapse all whitespace (newlines, tabs, multiple spaces) into single spaces
|
||||||
|
cleaned := strings.Join(strings.Fields(text), " ")
|
||||||
|
return cleaned, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func extractTextFromMarkdown(fpath string) (string, error) {
|
||||||
|
data, err := os.ReadFile(fpath)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
// Convert markdown to HTML
|
||||||
|
md := goldmark.New(
|
||||||
|
goldmark.WithExtensions(extension.GFM),
|
||||||
|
goldmark.WithParserOptions(parser.WithAutoHeadingID()),
|
||||||
|
goldmark.WithRendererOptions(html.WithUnsafe()), // allow raw HTML if needed
|
||||||
|
)
|
||||||
|
var buf bytes.Buffer
|
||||||
|
if err := md.Convert(data, &buf); err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
// Now extract text from the resulting HTML (using goquery or similar)
|
||||||
|
return extractTextFromHtmlContent(buf.Bytes())
|
||||||
|
}
|
||||||
|
|
||||||
|
func extractTextFromEpub(fpath string) (string, error) {
|
||||||
|
r, err := zip.OpenReader(fpath)
|
||||||
|
if err != nil {
|
||||||
|
return "", fmt.Errorf("failed to open epub: %w", err)
|
||||||
|
}
|
||||||
|
defer r.Close()
|
||||||
|
var sb strings.Builder
|
||||||
|
for _, f := range r.File {
|
||||||
|
ext := strings.ToLower(path.Ext(f.Name))
|
||||||
|
if ext != ".xhtml" && ext != ".html" && ext != ".htm" && ext != ".xml" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
// Skip manifest, toc, ncx files - they don't contain book content
|
||||||
|
nameLower := strings.ToLower(f.Name)
|
||||||
|
if strings.Contains(nameLower, "toc") || strings.Contains(nameLower, "nav") ||
|
||||||
|
strings.Contains(nameLower, "manifest") || strings.Contains(nameLower, ".opf") ||
|
||||||
|
strings.HasSuffix(nameLower, ".ncx") {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
rc, err := f.Open()
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
if sb.Len() > 0 {
|
||||||
|
sb.WriteString("\n\n")
|
||||||
|
}
|
||||||
|
sb.WriteString(f.Name)
|
||||||
|
sb.WriteString("\n")
|
||||||
|
|
||||||
|
buf, readErr := io.ReadAll(rc)
|
||||||
|
rc.Close()
|
||||||
|
if readErr == nil {
|
||||||
|
sb.WriteString(stripHTML(string(buf)))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if sb.Len() == 0 {
|
||||||
|
return "", errors.New("no content extracted from epub")
|
||||||
|
}
|
||||||
|
return sb.String(), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func stripHTML(html string) string {
|
||||||
|
var sb strings.Builder
|
||||||
|
inTag := false
|
||||||
|
for _, r := range html {
|
||||||
|
switch r {
|
||||||
|
case '<':
|
||||||
|
inTag = true
|
||||||
|
case '>':
|
||||||
|
inTag = false
|
||||||
|
default:
|
||||||
|
if !inTag {
|
||||||
|
sb.WriteRune(r)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return sb.String()
|
||||||
|
}
|
||||||
|
|
||||||
|
func extractTextFromPdf(fpath string) (string, error) {
|
||||||
|
_, err := exec.LookPath("pdftotext")
|
||||||
|
if err == nil {
|
||||||
|
out, err := exec.Command("pdftotext", "-layout", fpath, "-").Output()
|
||||||
|
if err == nil && len(out) > 0 {
|
||||||
|
return string(out), nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return extractTextFromPdfPureGo(fpath)
|
||||||
|
}
|
||||||
|
|
||||||
|
func extractTextFromPdfPureGo(fpath string) (string, error) {
|
||||||
|
df, r, err := pdf.Open(fpath)
|
||||||
|
if err != nil {
|
||||||
|
return "", fmt.Errorf("failed to open pdf: %w", err)
|
||||||
|
}
|
||||||
|
defer df.Close()
|
||||||
|
textReader, err := r.GetPlainText()
|
||||||
|
if err != nil {
|
||||||
|
return "", fmt.Errorf("failed to extract text from pdf: %w", err)
|
||||||
|
}
|
||||||
|
var buf bytes.Buffer
|
||||||
|
_, err = io.Copy(&buf, textReader)
|
||||||
|
if err != nil {
|
||||||
|
return "", fmt.Errorf("failed to read pdf text: %w", err)
|
||||||
|
}
|
||||||
|
return buf.String(), nil
|
||||||
|
}
|
||||||
504
rag/rag.go
504
rag/rag.go
@@ -7,8 +7,9 @@ import (
|
|||||||
"gf-lt/models"
|
"gf-lt/models"
|
||||||
"gf-lt/storage"
|
"gf-lt/storage"
|
||||||
"log/slog"
|
"log/slog"
|
||||||
"os"
|
|
||||||
"path"
|
"path"
|
||||||
|
"regexp"
|
||||||
|
"sort"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
@@ -23,19 +24,18 @@ var (
|
|||||||
ErrRAGStatus = "some error occurred; failed to transfer data to vector db"
|
ErrRAGStatus = "some error occurred; failed to transfer data to vector db"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
type RAG struct {
|
type RAG struct {
|
||||||
logger *slog.Logger
|
logger *slog.Logger
|
||||||
store storage.FullRepo
|
store storage.FullRepo
|
||||||
cfg *config.Config
|
cfg *config.Config
|
||||||
embedder Embedder
|
embedder Embedder
|
||||||
storage *VectorStorage
|
storage *VectorStorage
|
||||||
|
mu sync.Mutex
|
||||||
}
|
}
|
||||||
|
|
||||||
func New(l *slog.Logger, s storage.FullRepo, cfg *config.Config) *RAG {
|
func New(l *slog.Logger, s storage.FullRepo, cfg *config.Config) *RAG {
|
||||||
// Initialize with API embedder by default, could be configurable later
|
// Initialize with API embedder by default, could be configurable later
|
||||||
embedder := NewAPIEmbedder(l, cfg)
|
embedder := NewAPIEmbedder(l, cfg)
|
||||||
|
|
||||||
rag := &RAG{
|
rag := &RAG{
|
||||||
logger: l,
|
logger: l,
|
||||||
store: s,
|
store: s,
|
||||||
@@ -54,7 +54,9 @@ func wordCounter(sentence string) int {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (r *RAG) LoadRAG(fpath string) error {
|
func (r *RAG) LoadRAG(fpath string) error {
|
||||||
data, err := os.ReadFile(fpath)
|
r.mu.Lock()
|
||||||
|
defer r.mu.Unlock()
|
||||||
|
fileText, err := ExtractText(fpath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@@ -63,10 +65,7 @@ func (r *RAG) LoadRAG(fpath string) error {
|
|||||||
case LongJobStatusCh <- LoadedFileRAGStatus:
|
case LongJobStatusCh <- LoadedFileRAGStatus:
|
||||||
default:
|
default:
|
||||||
r.logger.Warn("LongJobStatusCh channel is full or closed, dropping status message", "message", LoadedFileRAGStatus)
|
r.logger.Warn("LongJobStatusCh channel is full or closed, dropping status message", "message", LoadedFileRAGStatus)
|
||||||
// Channel is full or closed, ignore the message to prevent panic
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fileText := string(data)
|
|
||||||
tokenizer, err := english.NewSentenceTokenizer(nil)
|
tokenizer, err := english.NewSentenceTokenizer(nil)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@@ -76,19 +75,16 @@ func (r *RAG) LoadRAG(fpath string) error {
|
|||||||
for i, s := range sentences {
|
for i, s := range sentences {
|
||||||
sents[i] = s.Text
|
sents[i] = s.Text
|
||||||
}
|
}
|
||||||
|
|
||||||
// Group sentences into paragraphs based on word limit
|
// Group sentences into paragraphs based on word limit
|
||||||
paragraphs := []string{}
|
paragraphs := []string{}
|
||||||
par := strings.Builder{}
|
par := strings.Builder{}
|
||||||
for i := 0; i < len(sents); i++ {
|
for i := 0; i < len(sents); i++ {
|
||||||
// Only add sentences that aren't empty
|
|
||||||
if strings.TrimSpace(sents[i]) != "" {
|
if strings.TrimSpace(sents[i]) != "" {
|
||||||
if par.Len() > 0 {
|
if par.Len() > 0 {
|
||||||
par.WriteString(" ") // Add space between sentences
|
par.WriteString(" ")
|
||||||
}
|
}
|
||||||
par.WriteString(sents[i])
|
par.WriteString(sents[i])
|
||||||
}
|
}
|
||||||
|
|
||||||
if wordCounter(par.String()) > int(r.cfg.RAGWordLimit) {
|
if wordCounter(par.String()) > int(r.cfg.RAGWordLimit) {
|
||||||
paragraph := strings.TrimSpace(par.String())
|
paragraph := strings.TrimSpace(par.String())
|
||||||
if paragraph != "" {
|
if paragraph != "" {
|
||||||
@@ -97,7 +93,6 @@ func (r *RAG) LoadRAG(fpath string) error {
|
|||||||
par.Reset()
|
par.Reset()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Handle any remaining content in the paragraph buffer
|
// Handle any remaining content in the paragraph buffer
|
||||||
if par.Len() > 0 {
|
if par.Len() > 0 {
|
||||||
paragraph := strings.TrimSpace(par.String())
|
paragraph := strings.TrimSpace(par.String())
|
||||||
@@ -105,216 +100,83 @@ func (r *RAG) LoadRAG(fpath string) error {
|
|||||||
paragraphs = append(paragraphs, paragraph)
|
paragraphs = append(paragraphs, paragraph)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Adjust batch size if needed
|
// Adjust batch size if needed
|
||||||
if len(paragraphs) < r.cfg.RAGBatchSize && len(paragraphs) > 0 {
|
if len(paragraphs) < r.cfg.RAGBatchSize && len(paragraphs) > 0 {
|
||||||
r.cfg.RAGBatchSize = len(paragraphs)
|
r.cfg.RAGBatchSize = len(paragraphs)
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(paragraphs) == 0 {
|
if len(paragraphs) == 0 {
|
||||||
return errors.New("no valid paragraphs found in file")
|
return errors.New("no valid paragraphs found in file")
|
||||||
}
|
}
|
||||||
|
// Process paragraphs in batches synchronously
|
||||||
var (
|
batchCount := 0
|
||||||
maxChSize = 100
|
for i := 0; i < len(paragraphs); i += r.cfg.RAGBatchSize {
|
||||||
left = 0
|
end := i + r.cfg.RAGBatchSize
|
||||||
right = r.cfg.RAGBatchSize
|
if end > len(paragraphs) {
|
||||||
batchCh = make(chan map[int][]string, maxChSize)
|
end = len(paragraphs)
|
||||||
vectorCh = make(chan []models.VectorRow, maxChSize)
|
|
||||||
errCh = make(chan error, 1)
|
|
||||||
wg = new(sync.WaitGroup)
|
|
||||||
lock = new(sync.Mutex)
|
|
||||||
)
|
|
||||||
|
|
||||||
defer close(errCh)
|
|
||||||
defer close(batchCh)
|
|
||||||
|
|
||||||
// Fill input channel with batches
|
|
||||||
ctn := 0
|
|
||||||
totalParagraphs := len(paragraphs)
|
|
||||||
for {
|
|
||||||
if right > totalParagraphs {
|
|
||||||
batchCh <- map[int][]string{left: paragraphs[left:]}
|
|
||||||
break
|
|
||||||
}
|
}
|
||||||
batchCh <- map[int][]string{left: paragraphs[left:right]}
|
batch := paragraphs[i:end]
|
||||||
left, right = right, right+r.cfg.RAGBatchSize
|
batchCount++
|
||||||
ctn++
|
// Filter empty paragraphs
|
||||||
|
nonEmptyBatch := make([]string, 0, len(batch))
|
||||||
|
for _, p := range batch {
|
||||||
|
if strings.TrimSpace(p) != "" {
|
||||||
|
nonEmptyBatch = append(nonEmptyBatch, strings.TrimSpace(p))
|
||||||
}
|
}
|
||||||
|
|
||||||
finishedBatchesMsg := fmt.Sprintf("finished batching batches#: %d; paragraphs: %d; sentences: %d\n", ctn+1, len(paragraphs), len(sents))
|
|
||||||
r.logger.Debug(finishedBatchesMsg)
|
|
||||||
select {
|
|
||||||
case LongJobStatusCh <- finishedBatchesMsg:
|
|
||||||
default:
|
|
||||||
r.logger.Warn("LongJobStatusCh channel is full or closed, dropping status message", "message", finishedBatchesMsg)
|
|
||||||
// Channel is full or closed, ignore the message to prevent panic
|
|
||||||
}
|
}
|
||||||
|
if len(nonEmptyBatch) == 0 {
|
||||||
// Start worker goroutines with WaitGroup
|
continue
|
||||||
wg.Add(int(r.cfg.RAGWorkers))
|
|
||||||
for w := 0; w < int(r.cfg.RAGWorkers); w++ {
|
|
||||||
go func(workerID int) {
|
|
||||||
defer wg.Done()
|
|
||||||
r.batchToVectorAsync(lock, workerID, batchCh, vectorCh, errCh, path.Base(fpath))
|
|
||||||
}(w)
|
|
||||||
}
|
}
|
||||||
|
// Embed the batch
|
||||||
// Use a goroutine to close the batchCh when all batches are sent
|
embeddings, err := r.embedder.EmbedSlice(nonEmptyBatch)
|
||||||
go func() {
|
|
||||||
wg.Wait()
|
|
||||||
close(vectorCh) // Close vectorCh when all workers are done
|
|
||||||
}()
|
|
||||||
|
|
||||||
// Check for errors from workers
|
|
||||||
// Use a non-blocking check for errors
|
|
||||||
select {
|
|
||||||
case err := <-errCh:
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
r.logger.Error("error during RAG processing", "error", err)
|
r.logger.Error("failed to embed batch", "error", err, "batch", batchCount)
|
||||||
|
select {
|
||||||
|
case LongJobStatusCh <- ErrRAGStatus:
|
||||||
|
default:
|
||||||
|
r.logger.Warn("LongJobStatusCh channel full, dropping message")
|
||||||
|
}
|
||||||
|
return fmt.Errorf("failed to embed batch %d: %w", batchCount, err)
|
||||||
|
}
|
||||||
|
if len(embeddings) != len(nonEmptyBatch) {
|
||||||
|
err := errors.New("embedding count mismatch")
|
||||||
|
r.logger.Error("embedding mismatch", "expected", len(nonEmptyBatch), "got", len(embeddings))
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
default:
|
// Write vectors to storage
|
||||||
// No immediate error, continue
|
filename := path.Base(fpath)
|
||||||
|
for j, text := range nonEmptyBatch {
|
||||||
|
vector := models.VectorRow{
|
||||||
|
Embeddings: embeddings[j],
|
||||||
|
RawText: text,
|
||||||
|
Slug: fmt.Sprintf("%s_%d_%d", filename, batchCount, j),
|
||||||
|
FileName: filename,
|
||||||
}
|
}
|
||||||
|
|
||||||
// Write vectors to storage - this will block until vectorCh is closed
|
|
||||||
return r.writeVectors(vectorCh)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (r *RAG) writeVectors(vectorCh chan []models.VectorRow) error {
|
|
||||||
for {
|
|
||||||
for batch := range vectorCh {
|
|
||||||
for _, vector := range batch {
|
|
||||||
if err := r.storage.WriteVector(&vector); err != nil {
|
if err := r.storage.WriteVector(&vector); err != nil {
|
||||||
r.logger.Error("failed to write vector to DB", "error", err, "slug", vector.Slug)
|
r.logger.Error("failed to write vector to DB", "error", err, "slug", vector.Slug)
|
||||||
select {
|
select {
|
||||||
case LongJobStatusCh <- ErrRAGStatus:
|
case LongJobStatusCh <- ErrRAGStatus:
|
||||||
default:
|
default:
|
||||||
r.logger.Warn("LongJobStatusCh channel is full or closed, dropping status message", "message", ErrRAGStatus)
|
r.logger.Warn("LongJobStatusCh channel full, dropping message")
|
||||||
// Channel is full or closed, ignore the message to prevent panic
|
|
||||||
}
|
}
|
||||||
return err // Stop the entire RAG operation on DB error
|
return fmt.Errorf("failed to write vector: %w", err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
r.logger.Debug("wrote batch to db", "size", len(batch), "vector_chan_len", len(vectorCh))
|
r.logger.Debug("wrote batch to db", "batch", batchCount, "size", len(nonEmptyBatch))
|
||||||
if len(vectorCh) == 0 {
|
// Send progress status
|
||||||
r.logger.Debug("finished writing vectors")
|
statusMsg := fmt.Sprintf("processed batch %d/%d", batchCount, (len(paragraphs)+r.cfg.RAGBatchSize-1)/r.cfg.RAGBatchSize)
|
||||||
|
select {
|
||||||
|
case LongJobStatusCh <- statusMsg:
|
||||||
|
default:
|
||||||
|
r.logger.Warn("LongJobStatusCh channel full, dropping message")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
r.logger.Debug("finished writing vectors", "batches", batchCount)
|
||||||
select {
|
select {
|
||||||
case LongJobStatusCh <- FinishedRAGStatus:
|
case LongJobStatusCh <- FinishedRAGStatus:
|
||||||
default:
|
default:
|
||||||
r.logger.Warn("LongJobStatusCh channel is full or closed, dropping status message", "message", FinishedRAGStatus)
|
r.logger.Warn("LongJobStatusCh channel is full or closed, dropping status message", "message", FinishedRAGStatus)
|
||||||
// Channel is full or closed, ignore the message to prevent panic
|
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (r *RAG) batchToVectorAsync(lock *sync.Mutex, id int, inputCh <-chan map[int][]string,
|
|
||||||
vectorCh chan<- []models.VectorRow, errCh chan error, filename string) {
|
|
||||||
var err error
|
|
||||||
|
|
||||||
defer func() {
|
|
||||||
// For errCh, make sure we only send if there's actually an error and the channel can accept it
|
|
||||||
if err != nil {
|
|
||||||
select {
|
|
||||||
case errCh <- err:
|
|
||||||
default:
|
|
||||||
// errCh might be full or closed, log but don't panic
|
|
||||||
r.logger.Warn("errCh channel full or closed, skipping error propagation", "worker", id, "error", err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
for {
|
|
||||||
lock.Lock()
|
|
||||||
if len(inputCh) == 0 {
|
|
||||||
lock.Unlock()
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
select {
|
|
||||||
case linesMap := <-inputCh:
|
|
||||||
for leftI, lines := range linesMap {
|
|
||||||
if err := r.fetchEmb(lines, errCh, vectorCh, fmt.Sprintf("%s_%d", filename, leftI), filename); err != nil {
|
|
||||||
r.logger.Error("error fetching embeddings", "error", err, "worker", id)
|
|
||||||
lock.Unlock()
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
lock.Unlock()
|
|
||||||
case err = <-errCh:
|
|
||||||
r.logger.Error("got an error from error channel", "error", err)
|
|
||||||
lock.Unlock()
|
|
||||||
return
|
|
||||||
default:
|
|
||||||
lock.Unlock()
|
|
||||||
}
|
|
||||||
|
|
||||||
r.logger.Debug("processed batch", "batches#", len(inputCh), "worker#", id)
|
|
||||||
statusMsg := fmt.Sprintf("converted to vector; batches: %d, worker#: %d", len(inputCh), id)
|
|
||||||
select {
|
|
||||||
case LongJobStatusCh <- statusMsg:
|
|
||||||
default:
|
|
||||||
r.logger.Warn("LongJobStatusCh channel full or closed, dropping status message", "message", statusMsg)
|
|
||||||
// Channel is full or closed, ignore the message to prevent panic
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (r *RAG) fetchEmb(lines []string, errCh chan error, vectorCh chan<- []models.VectorRow, slug, filename string) error {
|
|
||||||
// Filter out empty lines before sending to embedder
|
|
||||||
nonEmptyLines := make([]string, 0, len(lines))
|
|
||||||
for _, line := range lines {
|
|
||||||
trimmed := strings.TrimSpace(line)
|
|
||||||
if trimmed != "" {
|
|
||||||
nonEmptyLines = append(nonEmptyLines, trimmed)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Skip if no non-empty lines
|
|
||||||
if len(nonEmptyLines) == 0 {
|
|
||||||
// Send empty result but don't error
|
|
||||||
vectorCh <- []models.VectorRow{}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
embeddings, err := r.embedder.EmbedSlice(nonEmptyLines)
|
|
||||||
if err != nil {
|
|
||||||
r.logger.Error("failed to embed lines", "err", err.Error())
|
|
||||||
errCh <- err
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
if len(embeddings) == 0 {
|
|
||||||
err := errors.New("no embeddings returned")
|
|
||||||
r.logger.Error("empty embeddings")
|
|
||||||
errCh <- err
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
if len(embeddings) != len(nonEmptyLines) {
|
|
||||||
err := errors.New("mismatch between number of lines and embeddings returned")
|
|
||||||
r.logger.Error("embedding mismatch", "err", err.Error())
|
|
||||||
errCh <- err
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
// Create a VectorRow for each line in the batch
|
|
||||||
vectors := make([]models.VectorRow, len(nonEmptyLines))
|
|
||||||
for i, line := range nonEmptyLines {
|
|
||||||
vectors[i] = models.VectorRow{
|
|
||||||
Embeddings: embeddings[i],
|
|
||||||
RawText: line,
|
|
||||||
Slug: fmt.Sprintf("%s_%d", slug, i),
|
|
||||||
FileName: filename,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
vectorCh <- vectors
|
|
||||||
return nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *RAG) LineToVector(line string) ([]float32, error) {
|
func (r *RAG) LineToVector(line string) ([]float32, error) {
|
||||||
@@ -332,3 +194,259 @@ func (r *RAG) ListLoaded() ([]string, error) {
|
|||||||
func (r *RAG) RemoveFile(filename string) error {
|
func (r *RAG) RemoveFile(filename string) error {
|
||||||
return r.storage.RemoveEmbByFileName(filename)
|
return r.storage.RemoveEmbByFileName(filename)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
queryRefinementPattern = regexp.MustCompile(`(?i)(based on my (vector db|vector db|vector database|rags?|past (conversations?|chat|messages?))|from my (files?|documents?|data|information|memory)|search (in|my) (vector db|database|rags?)|rag search for)`)
|
||||||
|
importantKeywords = []string{"project", "architecture", "code", "file", "chat", "conversation", "topic", "summary", "details", "history", "previous", "my", "user", "me"}
|
||||||
|
stopWords = []string{"the", "a", "an", "and", "or", "but", "in", "on", "at", "to", "for", "of", "with", "by", "from", "up", "down", "left", "right"}
|
||||||
|
)
|
||||||
|
|
||||||
|
func (r *RAG) RefineQuery(query string) string {
|
||||||
|
original := query
|
||||||
|
query = strings.TrimSpace(query)
|
||||||
|
if len(query) == 0 {
|
||||||
|
return original
|
||||||
|
}
|
||||||
|
if len(query) <= 3 {
|
||||||
|
return original
|
||||||
|
}
|
||||||
|
query = strings.ToLower(query)
|
||||||
|
for _, stopWord := range stopWords {
|
||||||
|
wordPattern := `\b` + stopWord + `\b`
|
||||||
|
re := regexp.MustCompile(wordPattern)
|
||||||
|
query = re.ReplaceAllString(query, "")
|
||||||
|
}
|
||||||
|
query = strings.TrimSpace(query)
|
||||||
|
if len(query) < 5 {
|
||||||
|
return original
|
||||||
|
}
|
||||||
|
if queryRefinementPattern.MatchString(original) {
|
||||||
|
cleaned := queryRefinementPattern.ReplaceAllString(original, "")
|
||||||
|
cleaned = strings.TrimSpace(cleaned)
|
||||||
|
if len(cleaned) >= 5 {
|
||||||
|
return cleaned
|
||||||
|
}
|
||||||
|
}
|
||||||
|
query = r.extractImportantPhrases(query)
|
||||||
|
if len(query) < 5 {
|
||||||
|
return original
|
||||||
|
}
|
||||||
|
return query
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *RAG) extractImportantPhrases(query string) string {
|
||||||
|
words := strings.Fields(query)
|
||||||
|
var important []string
|
||||||
|
for _, word := range words {
|
||||||
|
word = strings.Trim(word, ".,!?;:'\"()[]{}")
|
||||||
|
isImportant := false
|
||||||
|
for _, kw := range importantKeywords {
|
||||||
|
if strings.Contains(strings.ToLower(word), kw) {
|
||||||
|
isImportant = true
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if isImportant || len(word) > 3 {
|
||||||
|
important = append(important, word)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if len(important) == 0 {
|
||||||
|
return query
|
||||||
|
}
|
||||||
|
return strings.Join(important, " ")
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *RAG) GenerateQueryVariations(query string) []string {
|
||||||
|
variations := []string{query}
|
||||||
|
if len(query) < 5 {
|
||||||
|
return variations
|
||||||
|
}
|
||||||
|
parts := strings.Fields(query)
|
||||||
|
if len(parts) == 0 {
|
||||||
|
return variations
|
||||||
|
}
|
||||||
|
if len(parts) >= 2 {
|
||||||
|
trimmed := strings.Join(parts[:len(parts)-1], " ")
|
||||||
|
if len(trimmed) >= 5 {
|
||||||
|
variations = append(variations, trimmed)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if len(parts) >= 2 {
|
||||||
|
trimmed := strings.Join(parts[1:], " ")
|
||||||
|
if len(trimmed) >= 5 {
|
||||||
|
variations = append(variations, trimmed)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if !strings.HasSuffix(query, " explanation") {
|
||||||
|
variations = append(variations, query+" explanation")
|
||||||
|
}
|
||||||
|
if !strings.HasPrefix(query, "what is ") {
|
||||||
|
variations = append(variations, "what is "+query)
|
||||||
|
}
|
||||||
|
if !strings.HasSuffix(query, " details") {
|
||||||
|
variations = append(variations, query+" details")
|
||||||
|
}
|
||||||
|
if !strings.HasSuffix(query, " summary") {
|
||||||
|
variations = append(variations, query+" summary")
|
||||||
|
}
|
||||||
|
return variations
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *RAG) RerankResults(results []models.VectorRow, query string) []models.VectorRow {
|
||||||
|
type scoredResult struct {
|
||||||
|
row models.VectorRow
|
||||||
|
distance float32
|
||||||
|
}
|
||||||
|
scored := make([]scoredResult, 0, len(results))
|
||||||
|
for i := range results {
|
||||||
|
row := results[i]
|
||||||
|
|
||||||
|
score := float32(0)
|
||||||
|
rawTextLower := strings.ToLower(row.RawText)
|
||||||
|
queryLower := strings.ToLower(query)
|
||||||
|
if strings.Contains(rawTextLower, queryLower) {
|
||||||
|
score += 10
|
||||||
|
}
|
||||||
|
queryWords := strings.Fields(queryLower)
|
||||||
|
matchCount := 0
|
||||||
|
for _, word := range queryWords {
|
||||||
|
if len(word) > 2 && strings.Contains(rawTextLower, word) {
|
||||||
|
matchCount++
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if len(queryWords) > 0 {
|
||||||
|
score += float32(matchCount) / float32(len(queryWords)) * 5
|
||||||
|
}
|
||||||
|
if row.FileName == "chat" || strings.Contains(strings.ToLower(row.FileName), "conversation") {
|
||||||
|
score += 3
|
||||||
|
}
|
||||||
|
distance := row.Distance - score/100
|
||||||
|
scored = append(scored, scoredResult{row: row, distance: distance})
|
||||||
|
}
|
||||||
|
sort.Slice(scored, func(i, j int) bool {
|
||||||
|
return scored[i].distance < scored[j].distance
|
||||||
|
})
|
||||||
|
unique := make([]models.VectorRow, 0)
|
||||||
|
seen := make(map[string]bool)
|
||||||
|
for i := range scored {
|
||||||
|
if !seen[scored[i].row.Slug] {
|
||||||
|
seen[scored[i].row.Slug] = true
|
||||||
|
unique = append(unique, scored[i].row)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if len(unique) > 10 {
|
||||||
|
unique = unique[:10]
|
||||||
|
}
|
||||||
|
return unique
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *RAG) SynthesizeAnswer(results []models.VectorRow, query string) (string, error) {
|
||||||
|
if len(results) == 0 {
|
||||||
|
return "No relevant information found in the vector database.", nil
|
||||||
|
}
|
||||||
|
var contextBuilder strings.Builder
|
||||||
|
contextBuilder.WriteString("User Query: ")
|
||||||
|
contextBuilder.WriteString(query)
|
||||||
|
contextBuilder.WriteString("\n\nRetrieved Context:\n")
|
||||||
|
for i, row := range results {
|
||||||
|
fmt.Fprintf(&contextBuilder, "[Source %d: %s]\n", i+1, row.FileName)
|
||||||
|
contextBuilder.WriteString(row.RawText)
|
||||||
|
contextBuilder.WriteString("\n\n")
|
||||||
|
}
|
||||||
|
contextBuilder.WriteString("Instructions: ")
|
||||||
|
contextBuilder.WriteString("Based on the retrieved context above, provide a concise, coherent answer to the user's query. ")
|
||||||
|
contextBuilder.WriteString("Extract only the most relevant information. ")
|
||||||
|
contextBuilder.WriteString("If no relevant information is found, state that clearly. ")
|
||||||
|
contextBuilder.WriteString("Cite sources by filename when relevant. ")
|
||||||
|
contextBuilder.WriteString("Do not include unnecessary preamble or explanations.")
|
||||||
|
synthesisPrompt := contextBuilder.String()
|
||||||
|
emb, err := r.LineToVector(synthesisPrompt)
|
||||||
|
if err != nil {
|
||||||
|
r.logger.Error("failed to embed synthesis prompt", "error", err)
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
embResp := &models.EmbeddingResp{
|
||||||
|
Embedding: emb,
|
||||||
|
Index: 0,
|
||||||
|
}
|
||||||
|
topResults, err := r.SearchEmb(embResp)
|
||||||
|
if err != nil {
|
||||||
|
r.logger.Error("failed to search for synthesis context", "error", err)
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
if len(topResults) > 0 && topResults[0].RawText != synthesisPrompt {
|
||||||
|
return topResults[0].RawText, nil
|
||||||
|
}
|
||||||
|
var finalAnswer strings.Builder
|
||||||
|
finalAnswer.WriteString("Based on the retrieved context:\n\n")
|
||||||
|
for i, row := range results {
|
||||||
|
if i >= 5 {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
fmt.Fprintf(&finalAnswer, "- From %s: %s\n", row.FileName, truncateString(row.RawText, 200))
|
||||||
|
}
|
||||||
|
return finalAnswer.String(), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func truncateString(s string, maxLen int) string {
|
||||||
|
if len(s) <= maxLen {
|
||||||
|
return s
|
||||||
|
}
|
||||||
|
return s[:maxLen] + "..."
|
||||||
|
}
|
||||||
|
|
||||||
|
func (r *RAG) Search(query string, limit int) ([]models.VectorRow, error) {
|
||||||
|
refined := r.RefineQuery(query)
|
||||||
|
variations := r.GenerateQueryVariations(refined)
|
||||||
|
allResults := make([]models.VectorRow, 0)
|
||||||
|
seen := make(map[string]bool)
|
||||||
|
for _, q := range variations {
|
||||||
|
emb, err := r.LineToVector(q)
|
||||||
|
if err != nil {
|
||||||
|
r.logger.Error("failed to embed query variation", "error", err, "query", q)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
embResp := &models.EmbeddingResp{
|
||||||
|
Embedding: emb,
|
||||||
|
Index: 0,
|
||||||
|
}
|
||||||
|
|
||||||
|
results, err := r.SearchEmb(embResp)
|
||||||
|
if err != nil {
|
||||||
|
r.logger.Error("failed to search embeddings", "error", err, "query", q)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, row := range results {
|
||||||
|
if !seen[row.Slug] {
|
||||||
|
seen[row.Slug] = true
|
||||||
|
allResults = append(allResults, row)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
reranked := r.RerankResults(allResults, query)
|
||||||
|
if len(reranked) > limit {
|
||||||
|
reranked = reranked[:limit]
|
||||||
|
}
|
||||||
|
return reranked, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
ragInstance *RAG
|
||||||
|
ragOnce sync.Once
|
||||||
|
)
|
||||||
|
|
||||||
|
func Init(c *config.Config, l *slog.Logger, s storage.FullRepo) error {
|
||||||
|
ragOnce.Do(func() {
|
||||||
|
if c == nil || l == nil || s == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
ragInstance = New(l, s, c)
|
||||||
|
})
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func GetInstance() *RAG {
|
||||||
|
return ragInstance
|
||||||
|
}
|
||||||
|
|||||||
@@ -28,7 +28,6 @@ func NewVectorStorage(logger *slog.Logger, store storage.FullRepo) *VectorStorag
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
// SerializeVector converts []float32 to binary blob
|
// SerializeVector converts []float32 to binary blob
|
||||||
func SerializeVector(vec []float32) []byte {
|
func SerializeVector(vec []float32) []byte {
|
||||||
buf := make([]byte, len(vec)*4) // 4 bytes per float32
|
buf := make([]byte, len(vec)*4) // 4 bytes per float32
|
||||||
@@ -66,17 +65,14 @@ func (vs *VectorStorage) WriteVector(row *models.VectorRow) error {
|
|||||||
|
|
||||||
// Serialize the embeddings to binary
|
// Serialize the embeddings to binary
|
||||||
serializedEmbeddings := SerializeVector(row.Embeddings)
|
serializedEmbeddings := SerializeVector(row.Embeddings)
|
||||||
|
|
||||||
query := fmt.Sprintf(
|
query := fmt.Sprintf(
|
||||||
"INSERT INTO %s (embeddings, slug, raw_text, filename) VALUES (?, ?, ?, ?)",
|
"INSERT INTO %s (embeddings, slug, raw_text, filename) VALUES (?, ?, ?, ?)",
|
||||||
tableName,
|
tableName,
|
||||||
)
|
)
|
||||||
|
|
||||||
if _, err := vs.sqlxDB.Exec(query, serializedEmbeddings, row.Slug, row.RawText, row.FileName); err != nil {
|
if _, err := vs.sqlxDB.Exec(query, serializedEmbeddings, row.Slug, row.RawText, row.FileName); err != nil {
|
||||||
vs.logger.Error("failed to write vector", "error", err, "slug", row.Slug)
|
vs.logger.Error("failed to write vector", "error", err, "slug", row.Slug)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -95,11 +91,9 @@ func (vs *VectorStorage) getTableName(emb []float32) (string, error) {
|
|||||||
4096: true,
|
4096: true,
|
||||||
5120: true,
|
5120: true,
|
||||||
}
|
}
|
||||||
|
|
||||||
if supportedSizes[size] {
|
if supportedSizes[size] {
|
||||||
return fmt.Sprintf("embeddings_%d", size), nil
|
return fmt.Sprintf("embeddings_%d", size), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
return "", fmt.Errorf("no table for embedding size of %d", size)
|
return "", fmt.Errorf("no table for embedding size of %d", size)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -126,9 +120,7 @@ func (vs *VectorStorage) SearchClosest(query []float32) ([]models.VectorRow, err
|
|||||||
vector models.VectorRow
|
vector models.VectorRow
|
||||||
distance float32
|
distance float32
|
||||||
}
|
}
|
||||||
|
|
||||||
var topResults []SearchResult
|
var topResults []SearchResult
|
||||||
|
|
||||||
// Process vectors one by one to avoid loading everything into memory
|
// Process vectors one by one to avoid loading everything into memory
|
||||||
for rows.Next() {
|
for rows.Next() {
|
||||||
var (
|
var (
|
||||||
@@ -176,14 +168,12 @@ func (vs *VectorStorage) SearchClosest(query []float32) ([]models.VectorRow, err
|
|||||||
result.vector.Distance = result.distance
|
result.vector.Distance = result.distance
|
||||||
results = append(results, result.vector)
|
results = append(results, result.vector)
|
||||||
}
|
}
|
||||||
|
|
||||||
return results, nil
|
return results, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// ListFiles returns a list of all loaded files
|
// ListFiles returns a list of all loaded files
|
||||||
func (vs *VectorStorage) ListFiles() ([]string, error) {
|
func (vs *VectorStorage) ListFiles() ([]string, error) {
|
||||||
fileLists := make([][]string, 0)
|
fileLists := make([][]string, 0)
|
||||||
|
|
||||||
// Query all supported tables and combine results
|
// Query all supported tables and combine results
|
||||||
embeddingSizes := []int{384, 768, 1024, 1536, 2048, 3072, 4096, 5120}
|
embeddingSizes := []int{384, 768, 1024, 1536, 2048, 3072, 4096, 5120}
|
||||||
for _, size := range embeddingSizes {
|
for _, size := range embeddingSizes {
|
||||||
@@ -219,14 +209,12 @@ func (vs *VectorStorage) ListFiles() ([]string, error) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return allFiles, nil
|
return allFiles, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// RemoveEmbByFileName removes all embeddings associated with a specific filename
|
// RemoveEmbByFileName removes all embeddings associated with a specific filename
|
||||||
func (vs *VectorStorage) RemoveEmbByFileName(filename string) error {
|
func (vs *VectorStorage) RemoveEmbByFileName(filename string) error {
|
||||||
var errors []string
|
var errors []string
|
||||||
|
|
||||||
embeddingSizes := []int{384, 768, 1024, 1536, 2048, 3072, 4096, 5120}
|
embeddingSizes := []int{384, 768, 1024, 1536, 2048, 3072, 4096, 5120}
|
||||||
for _, size := range embeddingSizes {
|
for _, size := range embeddingSizes {
|
||||||
table := fmt.Sprintf("embeddings_%d", size)
|
table := fmt.Sprintf("embeddings_%d", size)
|
||||||
@@ -235,11 +223,9 @@ func (vs *VectorStorage) RemoveEmbByFileName(filename string) error {
|
|||||||
errors = append(errors, err.Error())
|
errors = append(errors, err.Error())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(errors) > 0 {
|
if len(errors) > 0 {
|
||||||
return fmt.Errorf("errors occurred: %s", strings.Join(errors, "; "))
|
return fmt.Errorf("errors occurred: %s", strings.Join(errors, "; "))
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -248,18 +234,15 @@ func cosineSimilarity(a, b []float32) float32 {
|
|||||||
if len(a) != len(b) {
|
if len(a) != len(b) {
|
||||||
return 0.0
|
return 0.0
|
||||||
}
|
}
|
||||||
|
|
||||||
var dotProduct, normA, normB float32
|
var dotProduct, normA, normB float32
|
||||||
for i := 0; i < len(a); i++ {
|
for i := 0; i < len(a); i++ {
|
||||||
dotProduct += a[i] * b[i]
|
dotProduct += a[i] * b[i]
|
||||||
normA += a[i] * a[i]
|
normA += a[i] * a[i]
|
||||||
normB += b[i] * b[i]
|
normB += b[i] * b[i]
|
||||||
}
|
}
|
||||||
|
|
||||||
if normA == 0 || normB == 0 {
|
if normA == 0 || normB == 0 {
|
||||||
return 0.0
|
return 0.0
|
||||||
}
|
}
|
||||||
|
|
||||||
return dotProduct / (sqrt(normA) * sqrt(normB))
|
return dotProduct / (sqrt(normA) * sqrt(normB))
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -275,4 +258,3 @@ func sqrt(f float32) float32 {
|
|||||||
}
|
}
|
||||||
return guess
|
return guess
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -103,7 +103,6 @@ func NewProviderSQL(dbPath string, logger *slog.Logger) FullRepo {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
p := ProviderSQL{db: db, logger: logger}
|
p := ProviderSQL{db: db, logger: logger}
|
||||||
|
|
||||||
p.Migrate()
|
p.Migrate()
|
||||||
return p
|
return p
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -73,12 +73,9 @@ func (p ProviderSQL) WriteVector(row *models.VectorRow) error {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
serializedEmbeddings := SerializeVector(row.Embeddings)
|
serializedEmbeddings := SerializeVector(row.Embeddings)
|
||||||
|
|
||||||
query := fmt.Sprintf("INSERT INTO %s(embeddings, slug, raw_text, filename) VALUES (?, ?, ?, ?)", tableName)
|
query := fmt.Sprintf("INSERT INTO %s(embeddings, slug, raw_text, filename) VALUES (?, ?, ?, ?)", tableName)
|
||||||
_, err = p.db.Exec(query, serializedEmbeddings, row.Slug, row.RawText, row.FileName)
|
_, err = p.db.Exec(query, serializedEmbeddings, row.Slug, row.RawText, row.FileName)
|
||||||
|
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -87,27 +84,22 @@ func (p ProviderSQL) SearchClosest(q []float32) ([]models.VectorRow, error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
querySQL := "SELECT embeddings, slug, raw_text, filename FROM " + tableName
|
querySQL := "SELECT embeddings, slug, raw_text, filename FROM " + tableName
|
||||||
rows, err := p.db.Query(querySQL)
|
rows, err := p.db.Query(querySQL)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
defer rows.Close()
|
defer rows.Close()
|
||||||
|
|
||||||
type SearchResult struct {
|
type SearchResult struct {
|
||||||
vector models.VectorRow
|
vector models.VectorRow
|
||||||
distance float32
|
distance float32
|
||||||
}
|
}
|
||||||
|
|
||||||
var topResults []SearchResult
|
var topResults []SearchResult
|
||||||
|
|
||||||
for rows.Next() {
|
for rows.Next() {
|
||||||
var (
|
var (
|
||||||
embeddingsBlob []byte
|
embeddingsBlob []byte
|
||||||
slug, rawText, fileName string
|
slug, rawText, fileName string
|
||||||
)
|
)
|
||||||
|
|
||||||
if err := rows.Scan(&embeddingsBlob, &slug, &rawText, &fileName); err != nil {
|
if err := rows.Scan(&embeddingsBlob, &slug, &rawText, &fileName); err != nil {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
@@ -152,7 +144,6 @@ func (p ProviderSQL) SearchClosest(q []float32) ([]models.VectorRow, error) {
|
|||||||
result.vector.Distance = result.distance
|
result.vector.Distance = result.distance
|
||||||
results[i] = result.vector
|
results[i] = result.vector
|
||||||
}
|
}
|
||||||
|
|
||||||
return results, nil
|
return results, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -161,18 +152,15 @@ func cosineSimilarity(a, b []float32) float32 {
|
|||||||
if len(a) != len(b) {
|
if len(a) != len(b) {
|
||||||
return 0.0
|
return 0.0
|
||||||
}
|
}
|
||||||
|
|
||||||
var dotProduct, normA, normB float32
|
var dotProduct, normA, normB float32
|
||||||
for i := 0; i < len(a); i++ {
|
for i := 0; i < len(a); i++ {
|
||||||
dotProduct += a[i] * b[i]
|
dotProduct += a[i] * b[i]
|
||||||
normA += a[i] * a[i]
|
normA += a[i] * a[i]
|
||||||
normB += b[i] * b[i]
|
normB += b[i] * b[i]
|
||||||
}
|
}
|
||||||
|
|
||||||
if normA == 0 || normB == 0 {
|
if normA == 0 || normB == 0 {
|
||||||
return 0.0
|
return 0.0
|
||||||
}
|
}
|
||||||
|
|
||||||
return dotProduct / (sqrt(normA) * sqrt(normB))
|
return dotProduct / (sqrt(normA) * sqrt(normB))
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -229,13 +217,11 @@ func (p ProviderSQL) ListFiles() ([]string, error) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return allFiles, nil
|
return allFiles, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p ProviderSQL) RemoveEmbByFileName(filename string) error {
|
func (p ProviderSQL) RemoveEmbByFileName(filename string) error {
|
||||||
var errors []string
|
var errors []string
|
||||||
|
|
||||||
tableNames := []string{
|
tableNames := []string{
|
||||||
"embeddings_384", "embeddings_768", "embeddings_1024", "embeddings_1536",
|
"embeddings_384", "embeddings_768", "embeddings_1024", "embeddings_1536",
|
||||||
"embeddings_2048", "embeddings_3072", "embeddings_4096", "embeddings_5120",
|
"embeddings_2048", "embeddings_3072", "embeddings_4096", "embeddings_5120",
|
||||||
@@ -246,10 +232,8 @@ func (p ProviderSQL) RemoveEmbByFileName(filename string) error {
|
|||||||
errors = append(errors, err.Error())
|
errors = append(errors, err.Error())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if len(errors) > 0 {
|
if len(errors) > 0 {
|
||||||
return fmt.Errorf("errors occurred: %v", errors)
|
return fmt.Errorf("errors occurred: %v", errors)
|
||||||
}
|
}
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,5 +1,5 @@
|
|||||||
{
|
{
|
||||||
"sys_prompt": "You are an expert software engineering assistant. Your goal is to help users with coding tasks, debugging, refactoring, and software development.\n\n## Core Principles\n1. **Security First**: Never expose secrets, keys, or credentials. Never commit sensitive data.\n2. **No Git Actions**: You can READ git info (status, log, diff) for context, but NEVER perform git actions (commit, add, push, checkout, reset, rm, etc.). Let the user handle all git operations.\n3. **Explore Before Execute**: Always understand the codebase structure before making changes.\n4. **Follow Conventions**: Match existing code style, patterns, and frameworks used in the project.\n5. **Be Concise**: Minimize output tokens while maintaining quality. Avoid unnecessary explanations.\n\n## Workflow for Complex Tasks\nFor multi-step tasks, ALWAYS use the todo system to track progress:\n\n1. **Create Todo List**: At the start of complex tasks, use `todo_create` to break down work into actionable items.\n2. **Update Progress**: Mark items as `in_progress` when working on them, and `completed` when done.\n3. **Check Status**: Use `todo_read` to review your progress.\n\nExample workflow:\n- User: \"Add user authentication to this app\"\n- You: Create todos: [\"Analyze existing auth structure\", \"Check frameworks in use\", \"Implement auth middleware\", \"Add login endpoints\", \"Test implementation\"]\n\n## Task Execution Flow\n\n### Phase 1: Exploration (Always First)\n- Use `file_list` to understand directory structure (path defaults to CodingDir if not specified)\n- Use `file_read` to examine relevant files (paths are relative to CodingDir unless starting with `/`)\n- Use `execute_command` with `grep`/`find` to search for patterns\n- Check `README` or documentation files\n- Identify: frameworks, conventions, testing approach\n- **Git reads allowed**: You may use `git status`, `git log`, `git diff` for context, but only to inform your work\n- **Path handling**: Relative paths are resolved against CodingDir (configurable via Alt+O). Use absolute paths (starting with `/`) to bypass CodingDir.\n\n### Phase 2: Planning\n- For complex tasks: create todo items\n- Identify files that need modification\n- Plan your approach following existing patterns\n\n### Phase 3: Implementation\n- Make changes using appropriate file tools\n- Prefer `file_write` for new files, `file_read` then modify for existing files\n- Follow existing code style exactly\n- Use existing libraries and utilities\n\n### Phase 4: Verification\n- Run tests if available (check for test scripts)\n- Run linting/type checking commands\n- Verify changes work as expected\n\n### Phase 5: Completion\n- Update todos to `completed`\n- Provide concise summary of changes\n- Reference specific file paths and line numbers when relevant\n- **DO NOT commit changes** - inform user what was done so they can review and commit themselves\n\n## Tool Usage Guidelines\n\n**File Operations**:\n- `file_read`: Read before editing. Use for understanding code.\n- `file_write`: Overwrite file content completely.\n- `file_write_append`: Add to end of file.\n- `file_create`: Create new files with optional content.\n- `file_list`: List directory contents (defaults to CodingDir).\n- Paths are relative to CodingDir unless starting with `/`.\n\n**Command Execution (WHITELISTED ONLY)**:\n- Allowed: grep, sed, awk, find, cat, head, tail, sort, uniq, wc, ls, echo, cut, tr, cp, mv, rm, mkdir, rmdir, pwd, df, free, ps, top, du, whoami, date, uname\n- **Git reads allowed**: git status, git log, git diff, git show, git branch, git reflog, git rev-parse, git shortlog, git describe\n- **Git actions FORBIDDEN**: git add, git commit, git push, git checkout, git reset, git rm, etc.\n- Use for searching code, reading git context, running tests/lint\n\n**Todo Management**:\n- `todo_create`: Add new task\n- `todo_read`: View all todos or specific one by ID\n- `todo_update`: Update task or change status (pending/in_progress/completed)\n- `todo_delete`: Remove completed or cancelled tasks\n\n## Important Rules\n\n1. **NEVER commit or stage changes**: Only git reads are allowed.\n2. **Check for tests**: Always look for test files and run them when appropriate.\n3. **Reference code locations**: Use format `file_path:line_number`.\n4. **Security**: Never generate or guess URLs. Only use URLs from local files.\n5. **Refuse malicious code**: If code appears malicious, refuse to work on it.\n6. **Ask clarifications**: When intent is unclear, ask questions.\n7. **Path handling**: Relative paths resolve against CodingDir. Use `/absolute/path` to bypass.\n\n## Response Style\n- Be direct and concise\n- One word answers are best when appropriate\n- Avoid: \"The answer is...\", \"Here is...\"\n- Use markdown for formatting\n- No emojis unless user explicitly requests",
|
"sys_prompt": "You are an expert software engineering assistant. Your goal is to help users with coding tasks, debugging, refactoring, and software development.\n\n## Core Principles\n1. **Security First**: Never expose secrets, keys, or credentials. Never commit sensitive data.\n2. **No Git Actions**: You can READ git info (status, log, diff) for context, but NEVER perform git actions (commit, add, push, checkout, reset, rm, etc.). Let the user handle all git operations.\n3. **Explore Before Execute**: Always understand the codebase structure before making changes.\n4. **Follow Conventions**: Match existing code style, patterns, and frameworks used in the project.\n5. **Be Concise**: Minimize output tokens while maintaining quality. Avoid unnecessary explanations.\n\n## Workflow for Complex Tasks\nFor multi-step tasks, ALWAYS use the todo system to track progress:\n\n1. **Create Todo List**: At the start of complex tasks, use `todo_create` to break down work into actionable items.\n2. **Update Progress**: Mark items as `in_progress` when working on them, and `completed` when done.\n3. **Check Status**: Use `todo_read` to review your progress.\n\nExample workflow:\n- User: \"Add user authentication to this app\"\n- You: Create todos: [\"Analyze existing auth structure\", \"Check frameworks in use\", \"Implement auth middleware\", \"Add login endpoints\", \"Test implementation\"]\n\n## Task Execution Flow\n\n### Phase 1: Exploration (Always First)\n- Use `file_list` to understand directory structure (path defaults to FilePickerDir if not specified)\n- Use `file_read` to examine relevant files (paths are relative to FilePickerDir unless starting with `/`)\n- Use `execute_command` with `grep`/`find` to search for patterns\n- Check `README` or documentation files\n- Identify: frameworks, conventions, testing approach\n- **Git reads allowed**: You may use `git status`, `git log`, `git diff` for context, but only to inform your work\n- **Path handling**: Relative paths are resolved against FilePickerDir (configurable via Alt+O). Use absolute paths (starting with `/`) to bypass FilePickerDir.\n\n### Phase 2: Planning\n- For complex tasks: create todo items\n- Identify files that need modification\n- Plan your approach following existing patterns\n\n### Phase 3: Implementation\n- Make changes using appropriate file tools\n- Prefer `file_write` for new files, `file_read` then modify for existing files\n- Follow existing code style exactly\n- Use existing libraries and utilities\n\n### Phase 4: Verification\n- Run tests if available (check for test scripts)\n- Run linting/type checking commands\n- Verify changes work as expected\n\n### Phase 5: Completion\n- Update todos to `completed`\n- Provide concise summary of changes\n- Reference specific file paths and line numbers when relevant\n- **DO NOT commit changes** - inform user what was done so they can review and commit themselves\n\n## Tool Usage Guidelines\n\n**File Operations**:\n- `file_read`: Read before editing. Use for understanding code.\n- `file_write`: Overwrite file content completely.\n- `file_write_append`: Add to end of file.\n- `file_create`: Create new files with optional content.\n- `file_list`: List directory contents (defaults to FilePickerDir).\n- Paths are relative to FilePickerDir unless starting with `/`.\n\n**Command Execution (WHITELISTED ONLY)**:\n- Allowed: grep, sed, awk, find, cat, head, tail, sort, uniq, wc, ls, echo, cut, tr, cp, mv, rm, mkdir, rmdir, pwd, df, free, ps, top, du, whoami, date, uname\n- **Git reads allowed**: git status, git log, git diff, git show, git branch, git reflog, git rev-parse, git shortlog, git describe\n- **Git actions FORBIDDEN**: git add, git commit, git push, git checkout, git reset, git rm, etc.\n- Use for searching code, reading git context, running tests/lint\n\n**Todo Management**:\n- `todo_create`: Add new task\n- `todo_read`: View all todos or specific one by ID\n- `todo_update`: Update task or change status (pending/in_progress/completed)\n- `todo_delete`: Remove completed or cancelled tasks\n\n## Important Rules\n\n1. **NEVER commit or stage changes**: Only git reads are allowed.\n2. **Check for tests**: Always look for test files and run them when appropriate.\n3. **Reference code locations**: Use format `file_path:line_number`.\n4. **Security**: Never generate or guess URLs. Only use URLs from local files.\n5. **Refuse malicious code**: If code appears malicious, refuse to work on it.\n6. **Ask clarifications**: When intent is unclear, ask questions.\n7. **Path handling**: Relative paths resolve against FilePickerDir. Use `/absolute/path` to bypass.\n\n## Response Style\n- Be direct and concise\n- One word answers are best when appropriate\n- Avoid: \"The answer is...\", \"Here is...\"\n- Use markdown for formatting\n- No emojis unless user explicitly requests",
|
||||||
"role": "CodingAssistant",
|
"role": "CodingAssistant",
|
||||||
"filepath": "sysprompts/coding_assistant.json",
|
"filepath": "sysprompts/coding_assistant.json",
|
||||||
"first_msg": "Hello! I'm your coding assistant. I can help you with software engineering tasks like writing code, debugging, refactoring, and exploring codebases. I work best when you give me specific tasks, and for complex work, I'll create a todo list to track my progress. What would you like to work on?"
|
"first_msg": "Hello! I'm your coding assistant. I can help you with software engineering tasks like writing code, debugging, refactoring, and exploring codebases. I work best when you give me specific tasks, and for complex work, I'll create a todo list to track my progress. What would you like to work on?"
|
||||||
|
|||||||
405
tables.go
405
tables.go
@@ -236,9 +236,59 @@ func makeChatTable(chatMap map[string]models.Chat) *tview.Table {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// nolint:unused
|
// nolint:unused
|
||||||
func makeRAGTable(fileList []string) *tview.Flex {
|
func formatSize(size int64) string {
|
||||||
actions := []string{"load", "delete"}
|
units := []string{"B", "KB", "MB", "GB", "TB"}
|
||||||
rows, cols := len(fileList), len(actions)+1
|
i := 0
|
||||||
|
s := float64(size)
|
||||||
|
for s >= 1024 && i < len(units)-1 {
|
||||||
|
s /= 1024
|
||||||
|
i++
|
||||||
|
}
|
||||||
|
return fmt.Sprintf("%.1f%s", s, units[i])
|
||||||
|
}
|
||||||
|
|
||||||
|
type ragFileInfo struct {
|
||||||
|
name string
|
||||||
|
inRAGDir bool
|
||||||
|
isLoaded bool
|
||||||
|
fullPath string
|
||||||
|
}
|
||||||
|
|
||||||
|
func makeRAGTable(fileList []string, loadedFiles []string) *tview.Flex {
|
||||||
|
// Build set of loaded files for quick lookup
|
||||||
|
loadedSet := make(map[string]bool)
|
||||||
|
for _, f := range loadedFiles {
|
||||||
|
loadedSet[f] = true
|
||||||
|
}
|
||||||
|
|
||||||
|
// Build merged list: files from ragdir + orphaned files from DB
|
||||||
|
ragFiles := make([]ragFileInfo, 0, len(fileList)+len(loadedFiles))
|
||||||
|
seen := make(map[string]bool)
|
||||||
|
|
||||||
|
// Add files from ragdir
|
||||||
|
for _, f := range fileList {
|
||||||
|
ragFiles = append(ragFiles, ragFileInfo{
|
||||||
|
name: f,
|
||||||
|
inRAGDir: true,
|
||||||
|
isLoaded: loadedSet[f],
|
||||||
|
fullPath: path.Join(cfg.RAGDir, f),
|
||||||
|
})
|
||||||
|
seen[f] = true
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add orphaned files (in DB but not in ragdir)
|
||||||
|
for _, f := range loadedFiles {
|
||||||
|
if !seen[f] {
|
||||||
|
ragFiles = append(ragFiles, ragFileInfo{
|
||||||
|
name: f,
|
||||||
|
inRAGDir: false,
|
||||||
|
isLoaded: true,
|
||||||
|
fullPath: "",
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
rows := len(ragFiles)
|
||||||
|
cols := 4 // File Name | Preview | Action | Delete
|
||||||
fileTable := tview.NewTable().
|
fileTable := tview.NewTable().
|
||||||
SetBorders(true)
|
SetBorders(true)
|
||||||
longStatusView := tview.NewTextView()
|
longStatusView := tview.NewTextView()
|
||||||
@@ -252,41 +302,92 @@ func makeRAGTable(fileList []string) *tview.Flex {
|
|||||||
AddItem(fileTable, 0, 60, true)
|
AddItem(fileTable, 0, 60, true)
|
||||||
// Add the exit option as the first row (row 0)
|
// Add the exit option as the first row (row 0)
|
||||||
fileTable.SetCell(0, 0,
|
fileTable.SetCell(0, 0,
|
||||||
tview.NewTableCell("Exit RAG manager").
|
tview.NewTableCell("File Name").
|
||||||
SetTextColor(tcell.ColorWhite).
|
SetTextColor(tcell.ColorWhite).
|
||||||
SetAlign(tview.AlignCenter).
|
SetAlign(tview.AlignCenter).
|
||||||
SetSelectable(false))
|
SetSelectable(false))
|
||||||
fileTable.SetCell(0, 1,
|
fileTable.SetCell(0, 1,
|
||||||
tview.NewTableCell("(Close without action)").
|
tview.NewTableCell("Preview").
|
||||||
SetTextColor(tcell.ColorGray).
|
SetTextColor(tcell.ColorWhite).
|
||||||
SetAlign(tview.AlignCenter).
|
SetAlign(tview.AlignCenter).
|
||||||
SetSelectable(false))
|
SetSelectable(false))
|
||||||
fileTable.SetCell(0, 2,
|
fileTable.SetCell(0, 2,
|
||||||
tview.NewTableCell("exit").
|
tview.NewTableCell("Load/Unload").
|
||||||
SetTextColor(tcell.ColorGray).
|
SetTextColor(tcell.ColorWhite).
|
||||||
SetAlign(tview.AlignCenter))
|
SetAlign(tview.AlignCenter).
|
||||||
|
SetSelectable(false))
|
||||||
|
fileTable.SetCell(0, 3,
|
||||||
|
tview.NewTableCell("Delete").
|
||||||
|
SetTextColor(tcell.ColorWhite).
|
||||||
|
SetAlign(tview.AlignCenter).
|
||||||
|
SetSelectable(false))
|
||||||
// Add the file rows starting from row 1
|
// Add the file rows starting from row 1
|
||||||
for r := 0; r < rows; r++ {
|
for r := 0; r < rows; r++ {
|
||||||
|
f := ragFiles[r]
|
||||||
for c := 0; c < cols; c++ {
|
for c := 0; c < cols; c++ {
|
||||||
color := tcell.ColorWhite
|
color := tcell.ColorWhite
|
||||||
switch {
|
switch c {
|
||||||
case c < 1:
|
case 0:
|
||||||
fileTable.SetCell(r+1, c, // +1 to account for the exit row at index 0
|
displayName := f.name
|
||||||
tview.NewTableCell(fileList[r]).
|
if !f.inRAGDir {
|
||||||
|
displayName = f.name + " (orphaned)"
|
||||||
|
}
|
||||||
|
fileTable.SetCell(r+1, c,
|
||||||
|
tview.NewTableCell(displayName).
|
||||||
SetTextColor(color).
|
SetTextColor(color).
|
||||||
SetAlign(tview.AlignCenter).
|
SetAlign(tview.AlignCenter).
|
||||||
SetSelectable(false))
|
SetSelectable(false))
|
||||||
case c == 1: // Action description column - not selectable
|
case 1:
|
||||||
fileTable.SetCell(r+1, c, // +1 to account for the exit row at index 0
|
if !f.inRAGDir {
|
||||||
tview.NewTableCell("(Action)").
|
// Orphaned file - no preview available
|
||||||
|
fileTable.SetCell(r+1, c,
|
||||||
|
tview.NewTableCell("not in ragdir").
|
||||||
|
SetTextColor(tcell.ColorYellow).
|
||||||
|
SetAlign(tview.AlignCenter).
|
||||||
|
SetSelectable(false))
|
||||||
|
} else if fi, err := os.Stat(f.fullPath); err == nil {
|
||||||
|
size := fi.Size()
|
||||||
|
modTime := fi.ModTime()
|
||||||
|
preview := fmt.Sprintf("%s | %s", formatSize(size), modTime.Format("2006-01-02 15:04"))
|
||||||
|
fileTable.SetCell(r+1, c,
|
||||||
|
tview.NewTableCell(preview).
|
||||||
SetTextColor(color).
|
SetTextColor(color).
|
||||||
SetAlign(tview.AlignCenter).
|
SetAlign(tview.AlignCenter).
|
||||||
SetSelectable(false))
|
SetSelectable(false))
|
||||||
default: // Action button column - selectable
|
} else {
|
||||||
fileTable.SetCell(r+1, c, // +1 to account for the exit row at index 0
|
fileTable.SetCell(r+1, c,
|
||||||
tview.NewTableCell(actions[c-1]).
|
tview.NewTableCell("error").
|
||||||
|
SetTextColor(color).
|
||||||
|
SetAlign(tview.AlignCenter).
|
||||||
|
SetSelectable(false))
|
||||||
|
}
|
||||||
|
case 2:
|
||||||
|
actionText := "load"
|
||||||
|
if f.isLoaded {
|
||||||
|
actionText = "unload"
|
||||||
|
}
|
||||||
|
if !f.inRAGDir {
|
||||||
|
// Orphaned file - can only unload
|
||||||
|
actionText = "unload"
|
||||||
|
}
|
||||||
|
fileTable.SetCell(r+1, c,
|
||||||
|
tview.NewTableCell(actionText).
|
||||||
SetTextColor(color).
|
SetTextColor(color).
|
||||||
SetAlign(tview.AlignCenter))
|
SetAlign(tview.AlignCenter))
|
||||||
|
case 3:
|
||||||
|
if !f.inRAGDir {
|
||||||
|
// Orphaned file - cannot delete from ragdir (not there)
|
||||||
|
fileTable.SetCell(r+1, c,
|
||||||
|
tview.NewTableCell("-").
|
||||||
|
SetTextColor(tcell.ColorDarkGray).
|
||||||
|
SetAlign(tview.AlignCenter).
|
||||||
|
SetSelectable(false))
|
||||||
|
} else {
|
||||||
|
fileTable.SetCell(r+1, c,
|
||||||
|
tview.NewTableCell("delete").
|
||||||
|
SetTextColor(color).
|
||||||
|
SetAlign(tview.AlignCenter))
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -318,7 +419,7 @@ func makeRAGTable(fileList []string) *tview.Flex {
|
|||||||
}()
|
}()
|
||||||
fileTable.Select(0, 0).
|
fileTable.Select(0, 0).
|
||||||
SetFixed(1, 1).
|
SetFixed(1, 1).
|
||||||
SetSelectable(true, false).
|
SetSelectable(true, true).
|
||||||
SetSelectedStyle(tcell.StyleDefault.Background(tcell.ColorGray).Foreground(tcell.ColorWhite)).
|
SetSelectedStyle(tcell.StyleDefault.Background(tcell.ColorGray).Foreground(tcell.ColorWhite)).
|
||||||
SetDoneFunc(func(key tcell.Key) {
|
SetDoneFunc(func(key tcell.Key) {
|
||||||
if key == tcell.KeyEsc || key == tcell.KeyF1 || key == tcell.Key('x') || key == tcell.KeyCtrlX {
|
if key == tcell.KeyEsc || key == tcell.KeyF1 || key == tcell.Key('x') || key == tcell.KeyCtrlX {
|
||||||
@@ -335,30 +436,58 @@ func makeRAGTable(fileList []string) *tview.Flex {
|
|||||||
}
|
}
|
||||||
// defer pages.RemovePage(RAGPage)
|
// defer pages.RemovePage(RAGPage)
|
||||||
tc := fileTable.GetCell(row, column)
|
tc := fileTable.GetCell(row, column)
|
||||||
|
tc.SetTextColor(tcell.ColorRed)
|
||||||
|
fileTable.SetSelectable(false, false)
|
||||||
// Check if the selected row is the exit row (row 0) - do this first to avoid index issues
|
// Check if the selected row is the exit row (row 0) - do this first to avoid index issues
|
||||||
if row == 0 {
|
if row == 0 {
|
||||||
pages.RemovePage(RAGPage)
|
pages.RemovePage(RAGPage)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
// For file rows, get the filename (row index - 1 because of the exit row at index 0)
|
// For file rows, get the file info (row index - 1 because of the exit row at index 0)
|
||||||
fpath := fileList[row-1] // -1 to account for the exit row at index 0
|
f := ragFiles[row-1]
|
||||||
// notification := fmt.Sprintf("chat: %s; action: %s", fpath, tc.Text)
|
// Handle "-" case (orphaned file with no delete option)
|
||||||
|
if tc.Text == "-" {
|
||||||
|
pages.RemovePage(RAGPage)
|
||||||
|
return
|
||||||
|
}
|
||||||
switch tc.Text {
|
switch tc.Text {
|
||||||
case "load":
|
case "load":
|
||||||
fpath = path.Join(cfg.RAGDir, fpath)
|
fpath := path.Join(cfg.RAGDir, f.name)
|
||||||
longStatusView.SetText("clicked load")
|
longStatusView.SetText("clicked load")
|
||||||
go func() {
|
go func() {
|
||||||
if err := ragger.LoadRAG(fpath); err != nil {
|
if err := ragger.LoadRAG(fpath); err != nil {
|
||||||
logger.Error("failed to embed file", "chat", fpath, "error", err)
|
logger.Error("failed to embed file", "chat", fpath, "error", err)
|
||||||
_ = notifyUser("RAG", "failed to embed file; error: "+err.Error())
|
_ = notifyUser("RAG", "failed to embed file; error: "+err.Error())
|
||||||
errCh <- err
|
app.QueueUpdate(func() {
|
||||||
// pages.RemovePage(RAGPage)
|
pages.RemovePage(RAGPage)
|
||||||
|
})
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
_ = notifyUser("RAG", "file loaded successfully")
|
||||||
|
app.QueueUpdate(func() {
|
||||||
|
pages.RemovePage(RAGPage)
|
||||||
|
})
|
||||||
|
}()
|
||||||
|
return
|
||||||
|
case "unload":
|
||||||
|
longStatusView.SetText("clicked unload")
|
||||||
|
go func() {
|
||||||
|
if err := ragger.RemoveFile(f.name); err != nil {
|
||||||
|
logger.Error("failed to unload file from RAG", "filename", f.name, "error", err)
|
||||||
|
_ = notifyUser("RAG", "failed to unload file; error: "+err.Error())
|
||||||
|
app.QueueUpdate(func() {
|
||||||
|
pages.RemovePage(RAGPage)
|
||||||
|
})
|
||||||
|
return
|
||||||
|
}
|
||||||
|
_ = notifyUser("RAG", "file unloaded successfully")
|
||||||
|
app.QueueUpdate(func() {
|
||||||
|
pages.RemovePage(RAGPage)
|
||||||
|
})
|
||||||
}()
|
}()
|
||||||
return
|
return
|
||||||
case "delete":
|
case "delete":
|
||||||
fpath = path.Join(cfg.RAGDir, fpath)
|
fpath := path.Join(cfg.RAGDir, f.name)
|
||||||
if err := os.Remove(fpath); err != nil {
|
if err := os.Remove(fpath); err != nil {
|
||||||
logger.Error("failed to delete file", "filename", fpath, "error", err)
|
logger.Error("failed to delete file", "filename", fpath, "error", err)
|
||||||
return
|
return
|
||||||
@@ -383,114 +512,6 @@ func makeRAGTable(fileList []string) *tview.Flex {
|
|||||||
return ragflex
|
return ragflex
|
||||||
}
|
}
|
||||||
|
|
||||||
func makeLoadedRAGTable(fileList []string) *tview.Flex {
|
|
||||||
actions := []string{"delete"}
|
|
||||||
rows, cols := len(fileList), len(actions)+1
|
|
||||||
// Add 1 extra row for the "exit" option at the top
|
|
||||||
fileTable := tview.NewTable().
|
|
||||||
SetBorders(true)
|
|
||||||
longStatusView := tview.NewTextView()
|
|
||||||
longStatusView.SetText("Loaded RAG files list")
|
|
||||||
longStatusView.SetBorder(true).SetTitle("status")
|
|
||||||
longStatusView.SetChangedFunc(func() {
|
|
||||||
app.Draw()
|
|
||||||
})
|
|
||||||
ragflex := tview.NewFlex().SetDirection(tview.FlexRow).
|
|
||||||
AddItem(longStatusView, 0, 10, false).
|
|
||||||
AddItem(fileTable, 0, 60, true)
|
|
||||||
// Add the exit option as the first row (row 0)
|
|
||||||
fileTable.SetCell(0, 0,
|
|
||||||
tview.NewTableCell("Exit Loaded Files manager").
|
|
||||||
SetTextColor(tcell.ColorWhite).
|
|
||||||
SetAlign(tview.AlignCenter).
|
|
||||||
SetSelectable(false))
|
|
||||||
fileTable.SetCell(0, 1,
|
|
||||||
tview.NewTableCell("(Close without action)").
|
|
||||||
SetTextColor(tcell.ColorGray).
|
|
||||||
SetAlign(tview.AlignCenter).
|
|
||||||
SetSelectable(false))
|
|
||||||
fileTable.SetCell(0, 2,
|
|
||||||
tview.NewTableCell("exit").
|
|
||||||
SetTextColor(tcell.ColorGray).
|
|
||||||
SetAlign(tview.AlignCenter))
|
|
||||||
// Add the file rows starting from row 1
|
|
||||||
for r := 0; r < rows; r++ {
|
|
||||||
for c := 0; c < cols; c++ {
|
|
||||||
color := tcell.ColorWhite
|
|
||||||
switch {
|
|
||||||
case c < 1:
|
|
||||||
fileTable.SetCell(r+1, c, // +1 to account for the exit row at index 0
|
|
||||||
tview.NewTableCell(fileList[r]).
|
|
||||||
SetTextColor(color).
|
|
||||||
SetAlign(tview.AlignCenter).
|
|
||||||
SetSelectable(false))
|
|
||||||
case c == 1: // Action description column - not selectable
|
|
||||||
fileTable.SetCell(r+1, c, // +1 to account for the exit row at index 0
|
|
||||||
tview.NewTableCell("(Action)").
|
|
||||||
SetTextColor(color).
|
|
||||||
SetAlign(tview.AlignCenter).
|
|
||||||
SetSelectable(false))
|
|
||||||
default: // Action button column - selectable
|
|
||||||
fileTable.SetCell(r+1, c, // +1 to account for the exit row at index 0
|
|
||||||
tview.NewTableCell(actions[c-1]).
|
|
||||||
SetTextColor(color).
|
|
||||||
SetAlign(tview.AlignCenter))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
fileTable.Select(0, 0).
|
|
||||||
SetFixed(1, 1).
|
|
||||||
SetSelectable(true, false).
|
|
||||||
SetSelectedStyle(tcell.StyleDefault.Background(tcell.ColorGray).Foreground(tcell.ColorWhite)).
|
|
||||||
SetDoneFunc(func(key tcell.Key) {
|
|
||||||
if key == tcell.KeyEsc || key == tcell.KeyF1 || key == tcell.Key('x') || key == tcell.KeyCtrlX {
|
|
||||||
pages.RemovePage(RAGLoadedPage)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}).SetSelectedFunc(func(row int, column int) {
|
|
||||||
// If user selects a non-actionable column (0 or 1), move to first action column (2)
|
|
||||||
if column <= 1 {
|
|
||||||
if fileTable.GetColumnCount() > 2 {
|
|
||||||
fileTable.Select(row, 2) // Select first action column
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
tc := fileTable.GetCell(row, column)
|
|
||||||
// Check if the selected row is the exit row (row 0) - do this first to avoid index issues
|
|
||||||
if row == 0 {
|
|
||||||
pages.RemovePage(RAGLoadedPage)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
// For file rows, get the filename (row index - 1 because of the exit row at index 0)
|
|
||||||
fpath := fileList[row-1] // -1 to account for the exit row at index 0
|
|
||||||
switch tc.Text {
|
|
||||||
case "delete":
|
|
||||||
if err := ragger.RemoveFile(fpath); err != nil {
|
|
||||||
logger.Error("failed to delete file from RAG", "filename", fpath, "error", err)
|
|
||||||
longStatusView.SetText(fmt.Sprintf("Error deleting file: %v", err))
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if err := notifyUser("RAG file deleted", fpath+" was deleted from RAG system"); err != nil {
|
|
||||||
logger.Error("failed to send notification", "error", err)
|
|
||||||
}
|
|
||||||
longStatusView.SetText(fpath + " was deleted from RAG system")
|
|
||||||
return
|
|
||||||
default:
|
|
||||||
pages.RemovePage(RAGLoadedPage)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
})
|
|
||||||
// Add input capture to the flex container to handle 'x' key for closing
|
|
||||||
ragflex.SetInputCapture(func(event *tcell.EventKey) *tcell.EventKey {
|
|
||||||
if event.Key() == tcell.KeyRune && event.Rune() == 'x' {
|
|
||||||
pages.RemovePage(RAGLoadedPage)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
return event
|
|
||||||
})
|
|
||||||
return ragflex
|
|
||||||
}
|
|
||||||
|
|
||||||
func makeAgentTable(agentList []string) *tview.Table {
|
func makeAgentTable(agentList []string) *tview.Table {
|
||||||
actions := []string{"filepath", "load"}
|
actions := []string{"filepath", "load"}
|
||||||
rows, cols := len(agentList), len(actions)+1
|
rows, cols := len(agentList), len(actions)+1
|
||||||
@@ -499,14 +520,14 @@ func makeAgentTable(agentList []string) *tview.Table {
|
|||||||
for r := 0; r < rows; r++ {
|
for r := 0; r < rows; r++ {
|
||||||
for c := 0; c < cols; c++ {
|
for c := 0; c < cols; c++ {
|
||||||
color := tcell.ColorWhite
|
color := tcell.ColorWhite
|
||||||
switch {
|
switch c {
|
||||||
case c < 1:
|
case 0:
|
||||||
chatActTable.SetCell(r, c,
|
chatActTable.SetCell(r, c,
|
||||||
tview.NewTableCell(agentList[r]).
|
tview.NewTableCell(agentList[r]).
|
||||||
SetTextColor(color).
|
SetTextColor(color).
|
||||||
SetAlign(tview.AlignCenter).
|
SetAlign(tview.AlignCenter).
|
||||||
SetSelectable(false))
|
SetSelectable(false))
|
||||||
case c == 1:
|
case 1:
|
||||||
if actions[c-1] == "filepath" {
|
if actions[c-1] == "filepath" {
|
||||||
cc, ok := sysMap[agentList[r]]
|
cc, ok := sysMap[agentList[r]]
|
||||||
if !ok {
|
if !ok {
|
||||||
@@ -533,7 +554,7 @@ func makeAgentTable(agentList []string) *tview.Table {
|
|||||||
}
|
}
|
||||||
chatActTable.Select(0, 0).
|
chatActTable.Select(0, 0).
|
||||||
SetFixed(1, 1).
|
SetFixed(1, 1).
|
||||||
SetSelectable(true, false).
|
SetSelectable(true, true).
|
||||||
SetSelectedStyle(tcell.StyleDefault.Background(tcell.ColorGray).Foreground(tcell.ColorWhite)).
|
SetSelectedStyle(tcell.StyleDefault.Background(tcell.ColorGray).Foreground(tcell.ColorWhite)).
|
||||||
SetDoneFunc(func(key tcell.Key) {
|
SetDoneFunc(func(key tcell.Key) {
|
||||||
if key == tcell.KeyEsc || key == tcell.KeyF1 || key == tcell.Key('x') {
|
if key == tcell.KeyEsc || key == tcell.KeyF1 || key == tcell.Key('x') {
|
||||||
@@ -549,6 +570,8 @@ func makeAgentTable(agentList []string) *tview.Table {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
tc := chatActTable.GetCell(row, column)
|
tc := chatActTable.GetCell(row, column)
|
||||||
|
tc.SetTextColor(tcell.ColorRed)
|
||||||
|
chatActTable.SetSelectable(false, false)
|
||||||
selected := agentList[row]
|
selected := agentList[row]
|
||||||
// notification := fmt.Sprintf("chat: %s; action: %s", selectedChat, tc.Text)
|
// notification := fmt.Sprintf("chat: %s; action: %s", selectedChat, tc.Text)
|
||||||
switch tc.Text {
|
switch tc.Text {
|
||||||
@@ -630,7 +653,7 @@ func makeCodeBlockTable(codeBlocks []string) *tview.Table {
|
|||||||
}
|
}
|
||||||
table.Select(0, 0).
|
table.Select(0, 0).
|
||||||
SetFixed(1, 1).
|
SetFixed(1, 1).
|
||||||
SetSelectable(true, false).
|
SetSelectable(true, true).
|
||||||
SetSelectedStyle(tcell.StyleDefault.Background(tcell.ColorGray).Foreground(tcell.ColorWhite)).
|
SetSelectedStyle(tcell.StyleDefault.Background(tcell.ColorGray).Foreground(tcell.ColorWhite)).
|
||||||
SetDoneFunc(func(key tcell.Key) {
|
SetDoneFunc(func(key tcell.Key) {
|
||||||
if key == tcell.KeyEsc || key == tcell.KeyF1 || key == tcell.Key('x') {
|
if key == tcell.KeyEsc || key == tcell.KeyF1 || key == tcell.Key('x') {
|
||||||
@@ -646,6 +669,8 @@ func makeCodeBlockTable(codeBlocks []string) *tview.Table {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
tc := table.GetCell(row, column)
|
tc := table.GetCell(row, column)
|
||||||
|
tc.SetTextColor(tcell.ColorRed)
|
||||||
|
table.SetSelectable(false, false)
|
||||||
selected := codeBlocks[row]
|
selected := codeBlocks[row]
|
||||||
// notification := fmt.Sprintf("chat: %s; action: %s", selectedChat, tc.Text)
|
// notification := fmt.Sprintf("chat: %s; action: %s", selectedChat, tc.Text)
|
||||||
switch tc.Text {
|
switch tc.Text {
|
||||||
@@ -702,7 +727,7 @@ func makeImportChatTable(filenames []string) *tview.Table {
|
|||||||
}
|
}
|
||||||
chatActTable.Select(0, 0).
|
chatActTable.Select(0, 0).
|
||||||
SetFixed(1, 1).
|
SetFixed(1, 1).
|
||||||
SetSelectable(true, false).
|
SetSelectable(true, true).
|
||||||
SetSelectedStyle(tcell.StyleDefault.Background(tcell.ColorGray).Foreground(tcell.ColorWhite)).
|
SetSelectedStyle(tcell.StyleDefault.Background(tcell.ColorGray).Foreground(tcell.ColorWhite)).
|
||||||
SetDoneFunc(func(key tcell.Key) {
|
SetDoneFunc(func(key tcell.Key) {
|
||||||
if key == tcell.KeyEsc || key == tcell.KeyF1 || key == tcell.Key('x') {
|
if key == tcell.KeyEsc || key == tcell.KeyF1 || key == tcell.Key('x') {
|
||||||
@@ -718,6 +743,8 @@ func makeImportChatTable(filenames []string) *tview.Table {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
tc := chatActTable.GetCell(row, column)
|
tc := chatActTable.GetCell(row, column)
|
||||||
|
tc.SetTextColor(tcell.ColorRed)
|
||||||
|
chatActTable.SetSelectable(false, false)
|
||||||
selected := filenames[row]
|
selected := filenames[row]
|
||||||
// notification := fmt.Sprintf("chat: %s; action: %s", selectedChat, tc.Text)
|
// notification := fmt.Sprintf("chat: %s; action: %s", selectedChat, tc.Text)
|
||||||
switch tc.Text {
|
switch tc.Text {
|
||||||
@@ -792,6 +819,7 @@ func makeFilePicker() *tview.Flex {
|
|||||||
// --- NEW: search state ---
|
// --- NEW: search state ---
|
||||||
searching := false
|
searching := false
|
||||||
searchQuery := ""
|
searchQuery := ""
|
||||||
|
searchInputMode := false
|
||||||
// Helper function to check if a file has an allowed extension from config
|
// Helper function to check if a file has an allowed extension from config
|
||||||
hasAllowedExtension := func(filename string) bool {
|
hasAllowedExtension := func(filename string) bool {
|
||||||
if cfg.FilePickerExts == "" {
|
if cfg.FilePickerExts == "" {
|
||||||
@@ -820,7 +848,9 @@ func makeFilePicker() *tview.Flex {
|
|||||||
}
|
}
|
||||||
// Create UI elements
|
// Create UI elements
|
||||||
listView := tview.NewList()
|
listView := tview.NewList()
|
||||||
listView.SetBorder(true).SetTitle("Files & Directories [c: set CodingDir]").SetTitleAlign(tview.AlignLeft)
|
listView.SetBorder(true).
|
||||||
|
SetTitle("Files & Directories [s: set FilePickerDir]. Current base dir: " + cfg.FilePickerDir).
|
||||||
|
SetTitleAlign(tview.AlignLeft)
|
||||||
// Status view for selected file information
|
// Status view for selected file information
|
||||||
statusView := tview.NewTextView()
|
statusView := tview.NewTextView()
|
||||||
statusView.SetBorder(true).SetTitle("Selected File").SetTitleAlign(tview.AlignLeft)
|
statusView.SetBorder(true).SetTitle("Selected File").SetTitleAlign(tview.AlignLeft)
|
||||||
@@ -982,6 +1012,7 @@ func makeFilePicker() *tview.Flex {
|
|||||||
case tcell.KeyEsc:
|
case tcell.KeyEsc:
|
||||||
// Exit search, clear filter
|
// Exit search, clear filter
|
||||||
searching = false
|
searching = false
|
||||||
|
searchInputMode = false
|
||||||
searchQuery = ""
|
searchQuery = ""
|
||||||
refreshList(currentDisplayDir, "")
|
refreshList(currentDisplayDir, "")
|
||||||
return nil
|
return nil
|
||||||
@@ -991,16 +1022,80 @@ func makeFilePicker() *tview.Flex {
|
|||||||
refreshList(currentDisplayDir, searchQuery)
|
refreshList(currentDisplayDir, searchQuery)
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
case tcell.KeyRune:
|
case tcell.KeyEnter:
|
||||||
r := event.Rune()
|
// Exit search input mode and let normal processing handle selection
|
||||||
if r != 0 {
|
searchInputMode = false
|
||||||
searchQuery += string(r)
|
// Get the currently highlighted item in the list
|
||||||
refreshList(currentDisplayDir, searchQuery)
|
itemIndex := listView.GetCurrentItem()
|
||||||
|
if itemIndex >= 0 && itemIndex < listView.GetItemCount() {
|
||||||
|
itemText, _ := listView.GetItemText(itemIndex)
|
||||||
|
// Check for the exit option first
|
||||||
|
if strings.HasPrefix(itemText, "Exit file picker") {
|
||||||
|
pages.RemovePage(filePickerPage)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
// Extract the actual filename/directory name by removing the type info
|
||||||
|
actualItemName := itemText
|
||||||
|
if bracketPos := strings.Index(itemText, " ["); bracketPos != -1 {
|
||||||
|
actualItemName = itemText[:bracketPos]
|
||||||
|
}
|
||||||
|
// Check if it's a directory (ends with /)
|
||||||
|
if strings.HasSuffix(actualItemName, "/") {
|
||||||
|
var targetDir string
|
||||||
|
if strings.HasPrefix(actualItemName, "../") {
|
||||||
|
// Parent directory
|
||||||
|
targetDir = path.Dir(currentDisplayDir)
|
||||||
|
if targetDir == currentDisplayDir && currentDisplayDir == "/" {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// Regular subdirectory
|
||||||
|
dirName := strings.TrimSuffix(actualItemName, "/")
|
||||||
|
targetDir = path.Join(currentDisplayDir, dirName)
|
||||||
|
}
|
||||||
|
// Navigate – clear search
|
||||||
|
if cfg.ImagePreview && imgPreview != nil {
|
||||||
|
imgPreview.SetImage(nil)
|
||||||
|
}
|
||||||
|
searching = false
|
||||||
|
searchInputMode = false
|
||||||
|
searchQuery = ""
|
||||||
|
refreshList(targetDir, "")
|
||||||
|
dirStack = append(dirStack, targetDir)
|
||||||
|
currentStackPos = len(dirStack) - 1
|
||||||
|
statusView.SetText("Current: " + targetDir)
|
||||||
|
return nil
|
||||||
|
} else {
|
||||||
|
// It's a file
|
||||||
|
filePath := path.Join(currentDisplayDir, actualItemName)
|
||||||
|
if info, err := os.Stat(filePath); err == nil && !info.IsDir() {
|
||||||
|
if isImageFile(actualItemName) {
|
||||||
|
SetImageAttachment(filePath)
|
||||||
|
statusView.SetText("Image attached: " + filePath + " (will be sent with next message)")
|
||||||
|
pages.RemovePage(filePickerPage)
|
||||||
|
} else {
|
||||||
|
textArea.SetText(filePath, true)
|
||||||
|
app.SetFocus(textArea)
|
||||||
|
pages.RemovePage(filePickerPage)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
case tcell.KeyRune:
|
||||||
|
r := event.Rune()
|
||||||
|
if searchInputMode && r != 0 {
|
||||||
|
searchQuery += string(r)
|
||||||
|
refreshList(currentDisplayDir, searchQuery)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
// If not in search input mode, pass through for navigation
|
||||||
|
return event
|
||||||
default:
|
default:
|
||||||
// Pass all other keys (arrows, Enter, etc.) to normal processing
|
// Exit search input mode but keep filter active for navigation
|
||||||
// This allows selecting items while still in search mode
|
searchInputMode = false
|
||||||
|
// Pass all other keys (arrows, etc.) to normal processing
|
||||||
return event
|
return event
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -1028,12 +1123,13 @@ func makeFilePicker() *tview.Flex {
|
|||||||
if event.Rune() == '/' {
|
if event.Rune() == '/' {
|
||||||
// Enter search mode
|
// Enter search mode
|
||||||
searching = true
|
searching = true
|
||||||
|
searchInputMode = true
|
||||||
searchQuery = ""
|
searchQuery = ""
|
||||||
refreshList(currentDisplayDir, "")
|
refreshList(currentDisplayDir, "")
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Rune() == 'c' {
|
if event.Rune() == 's' {
|
||||||
// Set CodingDir to current directory
|
// Set FilePickerDir to current directory
|
||||||
itemIndex := listView.GetCurrentItem()
|
itemIndex := listView.GetCurrentItem()
|
||||||
if itemIndex >= 0 && itemIndex < listView.GetItemCount() {
|
if itemIndex >= 0 && itemIndex < listView.GetItemCount() {
|
||||||
itemText, _ := listView.GetItemText(itemIndex)
|
itemText, _ := listView.GetItemText(itemIndex)
|
||||||
@@ -1046,6 +1142,7 @@ func makeFilePicker() *tview.Flex {
|
|||||||
if bracketPos := strings.Index(itemText, " ["); bracketPos != -1 {
|
if bracketPos := strings.Index(itemText, " ["); bracketPos != -1 {
|
||||||
actualItemName = itemText[:bracketPos]
|
actualItemName = itemText[:bracketPos]
|
||||||
}
|
}
|
||||||
|
// nolint: gocritic
|
||||||
if strings.HasPrefix(actualItemName, "../") {
|
if strings.HasPrefix(actualItemName, "../") {
|
||||||
targetDir = path.Dir(currentDisplayDir)
|
targetDir = path.Dir(currentDisplayDir)
|
||||||
} else if strings.HasSuffix(actualItemName, "/") {
|
} else if strings.HasSuffix(actualItemName, "/") {
|
||||||
@@ -1055,11 +1152,11 @@ func makeFilePicker() *tview.Flex {
|
|||||||
targetDir = currentDisplayDir
|
targetDir = currentDisplayDir
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
cfg.CodingDir = targetDir
|
cfg.FilePickerDir = targetDir
|
||||||
if err := notifyUser("CodingDir", "Set to: "+targetDir); err != nil {
|
if err := notifyUser("FilePickerDir", "Set to: "+targetDir); err != nil {
|
||||||
logger.Error("failed to notify user", "error", err)
|
logger.Error("failed to notify user", "error", err)
|
||||||
}
|
}
|
||||||
pages.RemovePage(filePickerPage)
|
// pages.RemovePage(filePickerPage)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
142
tools.go
142
tools.go
@@ -16,6 +16,8 @@ import (
|
|||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"gf-lt/rag"
|
||||||
|
|
||||||
"github.com/GrailFinder/searchagent/searcher"
|
"github.com/GrailFinder/searchagent/searcher"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -58,9 +60,9 @@ Your current tools:
|
|||||||
"when_to_use": "when asked to search the web for information; returns clean summary without html,css and other web elements; limit is optional (default 3)"
|
"when_to_use": "when asked to search the web for information; returns clean summary without html,css and other web elements; limit is optional (default 3)"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name":"websearch_raw",
|
"name":"rag_search",
|
||||||
"args": ["query", "limit"],
|
"args": ["query", "limit"],
|
||||||
"when_to_use": "when asked to search the web for information; returns raw data as is without processing; limit is optional (default 3)"
|
"when_to_use": "when asked to search the local document database for information; performs query refinement, semantic search, reranking, and synthesis; returns clean summary with sources; limit is optional (default 3)"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"name":"read_url",
|
"name":"read_url",
|
||||||
@@ -146,6 +148,7 @@ under the topic: Adam's number is stored:
|
|||||||
After that you are free to respond to the user.
|
After that you are free to respond to the user.
|
||||||
`
|
`
|
||||||
webSearchSysPrompt = `Summarize the web search results, extracting key information and presenting a concise answer. Provide sources and URLs where relevant.`
|
webSearchSysPrompt = `Summarize the web search results, extracting key information and presenting a concise answer. Provide sources and URLs where relevant.`
|
||||||
|
ragSearchSysPrompt = `Synthesize the document search results, extracting key information and presenting a concise answer. Provide sources and document IDs where relevant.`
|
||||||
readURLSysPrompt = `Extract and summarize the content from the webpage. Provide key information, main points, and any relevant details.`
|
readURLSysPrompt = `Extract and summarize the content from the webpage. Provide key information, main points, and any relevant details.`
|
||||||
summarySysPrompt = `Please provide a concise summary of the following conversation. Focus on key points, decisions, and actions. Provide only the summary, no additional commentary.`
|
summarySysPrompt = `Please provide a concise summary of the following conversation. Focus on key points, decisions, and actions. Provide only the summary, no additional commentary.`
|
||||||
basicCard = &models.CharCard{
|
basicCard = &models.CharCard{
|
||||||
@@ -170,6 +173,9 @@ func init() {
|
|||||||
panic("failed to init seachagent; error: " + err.Error())
|
panic("failed to init seachagent; error: " + err.Error())
|
||||||
}
|
}
|
||||||
WebSearcher = sa
|
WebSearcher = sa
|
||||||
|
if err := rag.Init(cfg, logger, store); err != nil {
|
||||||
|
logger.Warn("failed to init rag; rag_search tool will not be available", "error", err)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// getWebAgentClient returns a singleton AgentClient for web agents.
|
// getWebAgentClient returns a singleton AgentClient for web agents.
|
||||||
@@ -196,6 +202,8 @@ func getWebAgentClient() *agent.AgentClient {
|
|||||||
func registerWebAgents() {
|
func registerWebAgents() {
|
||||||
webAgentsOnce.Do(func() {
|
webAgentsOnce.Do(func() {
|
||||||
client := getWebAgentClient()
|
client := getWebAgentClient()
|
||||||
|
// Register rag_search agent
|
||||||
|
agent.Register("rag_search", agent.NewWebAgentB(client, ragSearchSysPrompt))
|
||||||
// Register websearch agent
|
// Register websearch agent
|
||||||
agent.Register("websearch", agent.NewWebAgentB(client, webSearchSysPrompt))
|
agent.Register("websearch", agent.NewWebAgentB(client, webSearchSysPrompt))
|
||||||
// Register read_url agent
|
// Register read_url agent
|
||||||
@@ -239,6 +247,45 @@ func websearch(args map[string]string) []byte {
|
|||||||
return data
|
return data
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// rag search (searches local document database)
|
||||||
|
func ragsearch(args map[string]string) []byte {
|
||||||
|
query, ok := args["query"]
|
||||||
|
if !ok || query == "" {
|
||||||
|
msg := "query not provided to rag_search tool"
|
||||||
|
logger.Error(msg)
|
||||||
|
return []byte(msg)
|
||||||
|
}
|
||||||
|
limitS, ok := args["limit"]
|
||||||
|
if !ok || limitS == "" {
|
||||||
|
limitS = "3"
|
||||||
|
}
|
||||||
|
limit, err := strconv.Atoi(limitS)
|
||||||
|
if err != nil || limit == 0 {
|
||||||
|
logger.Warn("ragsearch limit; passed bad value; setting to default (3)",
|
||||||
|
"limit_arg", limitS, "error", err)
|
||||||
|
limit = 3
|
||||||
|
}
|
||||||
|
ragInstance := rag.GetInstance()
|
||||||
|
if ragInstance == nil {
|
||||||
|
msg := "rag not initialized; rag_search tool is not available"
|
||||||
|
logger.Error(msg)
|
||||||
|
return []byte(msg)
|
||||||
|
}
|
||||||
|
results, err := ragInstance.Search(query, limit)
|
||||||
|
if err != nil {
|
||||||
|
msg := "rag search failed; error: " + err.Error()
|
||||||
|
logger.Error(msg)
|
||||||
|
return []byte(msg)
|
||||||
|
}
|
||||||
|
data, err := json.Marshal(results)
|
||||||
|
if err != nil {
|
||||||
|
msg := "failed to marshal rag search result; error: " + err.Error()
|
||||||
|
logger.Error(msg)
|
||||||
|
return []byte(msg)
|
||||||
|
}
|
||||||
|
return data
|
||||||
|
}
|
||||||
|
|
||||||
// web search raw (returns raw data without processing)
|
// web search raw (returns raw data without processing)
|
||||||
func websearchRaw(args map[string]string) []byte {
|
func websearchRaw(args map[string]string) []byte {
|
||||||
// make http request return bytes
|
// make http request return bytes
|
||||||
@@ -369,7 +416,6 @@ func recallTopics(args map[string]string) []byte {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// File Manipulation Tools
|
// File Manipulation Tools
|
||||||
|
|
||||||
func fileCreate(args map[string]string) []byte {
|
func fileCreate(args map[string]string) []byte {
|
||||||
path, ok := args["path"]
|
path, ok := args["path"]
|
||||||
if !ok || path == "" {
|
if !ok || path == "" {
|
||||||
@@ -377,20 +423,16 @@ func fileCreate(args map[string]string) []byte {
|
|||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
path = resolvePath(path)
|
path = resolvePath(path)
|
||||||
|
|
||||||
content, ok := args["content"]
|
content, ok := args["content"]
|
||||||
if !ok {
|
if !ok {
|
||||||
content = ""
|
content = ""
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := writeStringToFile(path, content); err != nil {
|
if err := writeStringToFile(path, content); err != nil {
|
||||||
msg := "failed to create file; error: " + err.Error()
|
msg := "failed to create file; error: " + err.Error()
|
||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
msg := "file created successfully at " + path
|
msg := "file created successfully at " + path
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
@@ -402,16 +444,13 @@ func fileRead(args map[string]string) []byte {
|
|||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
path = resolvePath(path)
|
path = resolvePath(path)
|
||||||
|
|
||||||
content, err := readStringFromFile(path)
|
content, err := readStringFromFile(path)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
msg := "failed to read file; error: " + err.Error()
|
msg := "failed to read file; error: " + err.Error()
|
||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
result := map[string]string{
|
result := map[string]string{
|
||||||
"content": content,
|
"content": content,
|
||||||
"path": path,
|
"path": path,
|
||||||
@@ -422,7 +461,6 @@ func fileRead(args map[string]string) []byte {
|
|||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
return jsonResult
|
return jsonResult
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -475,15 +513,12 @@ func fileDelete(args map[string]string) []byte {
|
|||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
path = resolvePath(path)
|
path = resolvePath(path)
|
||||||
|
|
||||||
if err := removeFile(path); err != nil {
|
if err := removeFile(path); err != nil {
|
||||||
msg := "failed to delete file; error: " + err.Error()
|
msg := "failed to delete file; error: " + err.Error()
|
||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
msg := "file deleted successfully at " + path
|
msg := "file deleted successfully at " + path
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
@@ -496,7 +531,6 @@ func fileMove(args map[string]string) []byte {
|
|||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
src = resolvePath(src)
|
src = resolvePath(src)
|
||||||
|
|
||||||
dst, ok := args["dst"]
|
dst, ok := args["dst"]
|
||||||
if !ok || dst == "" {
|
if !ok || dst == "" {
|
||||||
msg := "destination path not provided to file_move tool"
|
msg := "destination path not provided to file_move tool"
|
||||||
@@ -504,13 +538,11 @@ func fileMove(args map[string]string) []byte {
|
|||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
dst = resolvePath(dst)
|
dst = resolvePath(dst)
|
||||||
|
|
||||||
if err := moveFile(src, dst); err != nil {
|
if err := moveFile(src, dst); err != nil {
|
||||||
msg := "failed to move file; error: " + err.Error()
|
msg := "failed to move file; error: " + err.Error()
|
||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
msg := fmt.Sprintf("file moved successfully from %s to %s", src, dst)
|
msg := fmt.Sprintf("file moved successfully from %s to %s", src, dst)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
@@ -523,7 +555,6 @@ func fileCopy(args map[string]string) []byte {
|
|||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
src = resolvePath(src)
|
src = resolvePath(src)
|
||||||
|
|
||||||
dst, ok := args["dst"]
|
dst, ok := args["dst"]
|
||||||
if !ok || dst == "" {
|
if !ok || dst == "" {
|
||||||
msg := "destination path not provided to file_copy tool"
|
msg := "destination path not provided to file_copy tool"
|
||||||
@@ -531,13 +562,11 @@ func fileCopy(args map[string]string) []byte {
|
|||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
dst = resolvePath(dst)
|
dst = resolvePath(dst)
|
||||||
|
|
||||||
if err := copyFile(src, dst); err != nil {
|
if err := copyFile(src, dst); err != nil {
|
||||||
msg := "failed to copy file; error: " + err.Error()
|
msg := "failed to copy file; error: " + err.Error()
|
||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
msg := fmt.Sprintf("file copied successfully from %s to %s", src, dst)
|
msg := fmt.Sprintf("file copied successfully from %s to %s", src, dst)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
@@ -547,16 +576,13 @@ func fileList(args map[string]string) []byte {
|
|||||||
if !ok || path == "" {
|
if !ok || path == "" {
|
||||||
path = "." // default to current directory
|
path = "." // default to current directory
|
||||||
}
|
}
|
||||||
|
|
||||||
path = resolvePath(path)
|
path = resolvePath(path)
|
||||||
|
|
||||||
files, err := listDirectory(path)
|
files, err := listDirectory(path)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
msg := "failed to list directory; error: " + err.Error()
|
msg := "failed to list directory; error: " + err.Error()
|
||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
result := map[string]interface{}{
|
result := map[string]interface{}{
|
||||||
"directory": path,
|
"directory": path,
|
||||||
"files": files,
|
"files": files,
|
||||||
@@ -567,17 +593,15 @@ func fileList(args map[string]string) []byte {
|
|||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
return jsonResult
|
return jsonResult
|
||||||
}
|
}
|
||||||
|
|
||||||
// Helper functions for file operations
|
// Helper functions for file operations
|
||||||
|
|
||||||
func resolvePath(p string) string {
|
func resolvePath(p string) string {
|
||||||
if filepath.IsAbs(p) {
|
if filepath.IsAbs(p) {
|
||||||
return p
|
return p
|
||||||
}
|
}
|
||||||
return filepath.Join(cfg.CodingDir, p)
|
return filepath.Join(cfg.FilePickerDir, p)
|
||||||
}
|
}
|
||||||
|
|
||||||
func readStringFromFile(filename string) (string, error) {
|
func readStringFromFile(filename string) (string, error) {
|
||||||
@@ -598,7 +622,6 @@ func appendStringToFile(filename string, data string) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
defer file.Close()
|
defer file.Close()
|
||||||
|
|
||||||
_, err = file.WriteString(data)
|
_, err = file.WriteString(data)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@@ -622,13 +645,11 @@ func copyFile(src, dst string) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
defer srcFile.Close()
|
defer srcFile.Close()
|
||||||
|
|
||||||
dstFile, err := os.Create(dst)
|
dstFile, err := os.Create(dst)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
defer dstFile.Close()
|
defer dstFile.Close()
|
||||||
|
|
||||||
_, err = io.Copy(dstFile, srcFile)
|
_, err = io.Copy(dstFile, srcFile)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@@ -647,7 +668,6 @@ func listDirectory(path string) ([]string, error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
var files []string
|
var files []string
|
||||||
for _, entry := range entries {
|
for _, entry := range entries {
|
||||||
if entry.IsDir() {
|
if entry.IsDir() {
|
||||||
@@ -656,12 +676,10 @@ func listDirectory(path string) ([]string, error) {
|
|||||||
files = append(files, entry.Name())
|
files = append(files, entry.Name())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return files, nil
|
return files, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Command Execution Tool
|
// Command Execution Tool
|
||||||
|
|
||||||
func executeCommand(args map[string]string) []byte {
|
func executeCommand(args map[string]string) []byte {
|
||||||
command, ok := args["command"]
|
command, ok := args["command"]
|
||||||
if !ok || command == "" {
|
if !ok || command == "" {
|
||||||
@@ -669,7 +687,6 @@ func executeCommand(args map[string]string) []byte {
|
|||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Get arguments - handle both single arg and multiple args
|
// Get arguments - handle both single arg and multiple args
|
||||||
var cmdArgs []string
|
var cmdArgs []string
|
||||||
if args["args"] != "" {
|
if args["args"] != "" {
|
||||||
@@ -688,43 +705,36 @@ func executeCommand(args map[string]string) []byte {
|
|||||||
argNum++
|
argNum++
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if !isCommandAllowed(command, cmdArgs...) {
|
if !isCommandAllowed(command, cmdArgs...) {
|
||||||
msg := fmt.Sprintf("command '%s' is not allowed", command)
|
msg := fmt.Sprintf("command '%s' is not allowed", command)
|
||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Execute with timeout for safety
|
// Execute with timeout for safety
|
||||||
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
||||||
defer cancel()
|
defer cancel()
|
||||||
cmd := exec.CommandContext(ctx, command, cmdArgs...)
|
cmd := exec.CommandContext(ctx, command, cmdArgs...)
|
||||||
|
|
||||||
output, err := cmd.CombinedOutput()
|
output, err := cmd.CombinedOutput()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
msg := fmt.Sprintf("command '%s' failed; error: %v; output: %s", command, err, string(output))
|
msg := fmt.Sprintf("command '%s' failed; error: %v; output: %s", command, err, string(output))
|
||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check if output is empty and return success message
|
// Check if output is empty and return success message
|
||||||
if len(output) == 0 {
|
if len(output) == 0 {
|
||||||
successMsg := fmt.Sprintf("command '%s %s' executed successfully and exited with code 0", command, strings.Join(cmdArgs, " "))
|
successMsg := fmt.Sprintf("command '%s %s' executed successfully and exited with code 0", command, strings.Join(cmdArgs, " "))
|
||||||
return []byte(successMsg)
|
return []byte(successMsg)
|
||||||
}
|
}
|
||||||
|
|
||||||
return output
|
return output
|
||||||
}
|
}
|
||||||
|
|
||||||
// Helper functions for command execution
|
// Helper functions for command execution
|
||||||
|
|
||||||
// Todo structure
|
// Todo structure
|
||||||
type TodoItem struct {
|
type TodoItem struct {
|
||||||
ID string `json:"id"`
|
ID string `json:"id"`
|
||||||
Task string `json:"task"`
|
Task string `json:"task"`
|
||||||
Status string `json:"status"` // "pending", "in_progress", "completed"
|
Status string `json:"status"` // "pending", "in_progress", "completed"
|
||||||
}
|
}
|
||||||
|
|
||||||
type TodoList struct {
|
type TodoList struct {
|
||||||
Items []TodoItem `json:"items"`
|
Items []TodoItem `json:"items"`
|
||||||
}
|
}
|
||||||
@@ -742,32 +752,26 @@ func todoCreate(args map[string]string) []byte {
|
|||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Generate simple ID
|
// Generate simple ID
|
||||||
id := fmt.Sprintf("todo_%d", len(globalTodoList.Items)+1)
|
id := fmt.Sprintf("todo_%d", len(globalTodoList.Items)+1)
|
||||||
|
|
||||||
newItem := TodoItem{
|
newItem := TodoItem{
|
||||||
ID: id,
|
ID: id,
|
||||||
Task: task,
|
Task: task,
|
||||||
Status: "pending",
|
Status: "pending",
|
||||||
}
|
}
|
||||||
|
|
||||||
globalTodoList.Items = append(globalTodoList.Items, newItem)
|
globalTodoList.Items = append(globalTodoList.Items, newItem)
|
||||||
|
|
||||||
result := map[string]string{
|
result := map[string]string{
|
||||||
"message": "todo created successfully",
|
"message": "todo created successfully",
|
||||||
"id": id,
|
"id": id,
|
||||||
"task": task,
|
"task": task,
|
||||||
"status": "pending",
|
"status": "pending",
|
||||||
}
|
}
|
||||||
|
|
||||||
jsonResult, err := json.Marshal(result)
|
jsonResult, err := json.Marshal(result)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
msg := "failed to marshal result; error: " + err.Error()
|
msg := "failed to marshal result; error: " + err.Error()
|
||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
return jsonResult
|
return jsonResult
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -801,7 +805,6 @@ func todoRead(args map[string]string) []byte {
|
|||||||
}
|
}
|
||||||
return jsonResult
|
return jsonResult
|
||||||
}
|
}
|
||||||
|
|
||||||
// Return all todos if no ID specified
|
// Return all todos if no ID specified
|
||||||
result := map[string]interface{}{
|
result := map[string]interface{}{
|
||||||
"todos": globalTodoList.Items,
|
"todos": globalTodoList.Items,
|
||||||
@@ -812,7 +815,6 @@ func todoRead(args map[string]string) []byte {
|
|||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
return jsonResult
|
return jsonResult
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -823,16 +825,13 @@ func todoUpdate(args map[string]string) []byte {
|
|||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
task, taskOk := args["task"]
|
task, taskOk := args["task"]
|
||||||
status, statusOk := args["status"]
|
status, statusOk := args["status"]
|
||||||
|
|
||||||
if !taskOk && !statusOk {
|
if !taskOk && !statusOk {
|
||||||
msg := "neither task nor status provided to todo_update tool"
|
msg := "neither task nor status provided to todo_update tool"
|
||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Find and update the todo
|
// Find and update the todo
|
||||||
for i, item := range globalTodoList.Items {
|
for i, item := range globalTodoList.Items {
|
||||||
if item.ID == id {
|
if item.ID == id {
|
||||||
@@ -856,23 +855,19 @@ func todoUpdate(args map[string]string) []byte {
|
|||||||
return jsonResult
|
return jsonResult
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
result := map[string]string{
|
result := map[string]string{
|
||||||
"message": "todo updated successfully",
|
"message": "todo updated successfully",
|
||||||
"id": id,
|
"id": id,
|
||||||
}
|
}
|
||||||
|
|
||||||
jsonResult, err := json.Marshal(result)
|
jsonResult, err := json.Marshal(result)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
msg := "failed to marshal result; error: " + err.Error()
|
msg := "failed to marshal result; error: " + err.Error()
|
||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
return jsonResult
|
return jsonResult
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// ID not found
|
// ID not found
|
||||||
result := map[string]string{
|
result := map[string]string{
|
||||||
"error": fmt.Sprintf("todo with id %s not found", id),
|
"error": fmt.Sprintf("todo with id %s not found", id),
|
||||||
@@ -893,29 +888,24 @@ func todoDelete(args map[string]string) []byte {
|
|||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Find and remove the todo
|
// Find and remove the todo
|
||||||
for i, item := range globalTodoList.Items {
|
for i, item := range globalTodoList.Items {
|
||||||
if item.ID == id {
|
if item.ID == id {
|
||||||
// Remove item from slice
|
// Remove item from slice
|
||||||
globalTodoList.Items = append(globalTodoList.Items[:i], globalTodoList.Items[i+1:]...)
|
globalTodoList.Items = append(globalTodoList.Items[:i], globalTodoList.Items[i+1:]...)
|
||||||
|
|
||||||
result := map[string]string{
|
result := map[string]string{
|
||||||
"message": "todo deleted successfully",
|
"message": "todo deleted successfully",
|
||||||
"id": id,
|
"id": id,
|
||||||
}
|
}
|
||||||
|
|
||||||
jsonResult, err := json.Marshal(result)
|
jsonResult, err := json.Marshal(result)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
msg := "failed to marshal result; error: " + err.Error()
|
msg := "failed to marshal result; error: " + err.Error()
|
||||||
logger.Error(msg)
|
logger.Error(msg)
|
||||||
return []byte(msg)
|
return []byte(msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
return jsonResult
|
return jsonResult
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// ID not found
|
// ID not found
|
||||||
result := map[string]string{
|
result := map[string]string{
|
||||||
"error": fmt.Sprintf("todo with id %s not found", id),
|
"error": fmt.Sprintf("todo with id %s not found", id),
|
||||||
@@ -997,6 +987,7 @@ var fnMap = map[string]fnSig{
|
|||||||
"recall": recall,
|
"recall": recall,
|
||||||
"recall_topics": recallTopics,
|
"recall_topics": recallTopics,
|
||||||
"memorise": memorise,
|
"memorise": memorise,
|
||||||
|
"rag_search": ragsearch,
|
||||||
"websearch": websearch,
|
"websearch": websearch,
|
||||||
"websearch_raw": websearchRaw,
|
"websearch_raw": websearchRaw,
|
||||||
"read_url": readURL,
|
"read_url": readURL,
|
||||||
@@ -1033,6 +1024,28 @@ func callToolWithAgent(name string, args map[string]string) []byte {
|
|||||||
|
|
||||||
// openai style def
|
// openai style def
|
||||||
var baseTools = []models.Tool{
|
var baseTools = []models.Tool{
|
||||||
|
// rag_search
|
||||||
|
models.Tool{
|
||||||
|
Type: "function",
|
||||||
|
Function: models.ToolFunc{
|
||||||
|
Name: "rag_search",
|
||||||
|
Description: "Search local document database given query, limit of sources (default 3). Performs query refinement, semantic search, reranking, and synthesis.",
|
||||||
|
Parameters: models.ToolFuncParams{
|
||||||
|
Type: "object",
|
||||||
|
Required: []string{"query", "limit"},
|
||||||
|
Properties: map[string]models.ToolArgProps{
|
||||||
|
"query": models.ToolArgProps{
|
||||||
|
Type: "string",
|
||||||
|
Description: "search query",
|
||||||
|
},
|
||||||
|
"limit": models.ToolArgProps{
|
||||||
|
Type: "string",
|
||||||
|
Description: "limit of the document results",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
// websearch
|
// websearch
|
||||||
models.Tool{
|
models.Tool{
|
||||||
Type: "function",
|
Type: "function",
|
||||||
@@ -1166,7 +1179,6 @@ var baseTools = []models.Tool{
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
// file_create
|
// file_create
|
||||||
models.Tool{
|
models.Tool{
|
||||||
Type: "function",
|
Type: "function",
|
||||||
@@ -1189,7 +1201,6 @@ var baseTools = []models.Tool{
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
// file_read
|
// file_read
|
||||||
models.Tool{
|
models.Tool{
|
||||||
Type: "function",
|
Type: "function",
|
||||||
@@ -1208,7 +1219,6 @@ var baseTools = []models.Tool{
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
// file_write
|
// file_write
|
||||||
models.Tool{
|
models.Tool{
|
||||||
Type: "function",
|
Type: "function",
|
||||||
@@ -1231,7 +1241,6 @@ var baseTools = []models.Tool{
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
// file_write_append
|
// file_write_append
|
||||||
models.Tool{
|
models.Tool{
|
||||||
Type: "function",
|
Type: "function",
|
||||||
@@ -1254,7 +1263,6 @@ var baseTools = []models.Tool{
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
// file_delete
|
// file_delete
|
||||||
models.Tool{
|
models.Tool{
|
||||||
Type: "function",
|
Type: "function",
|
||||||
@@ -1273,7 +1281,6 @@ var baseTools = []models.Tool{
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
// file_move
|
// file_move
|
||||||
models.Tool{
|
models.Tool{
|
||||||
Type: "function",
|
Type: "function",
|
||||||
@@ -1296,7 +1303,6 @@ var baseTools = []models.Tool{
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
// file_copy
|
// file_copy
|
||||||
models.Tool{
|
models.Tool{
|
||||||
Type: "function",
|
Type: "function",
|
||||||
@@ -1319,7 +1325,6 @@ var baseTools = []models.Tool{
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
// file_list
|
// file_list
|
||||||
models.Tool{
|
models.Tool{
|
||||||
Type: "function",
|
Type: "function",
|
||||||
@@ -1338,7 +1343,6 @@ var baseTools = []models.Tool{
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
|
|
||||||
// execute_command
|
// execute_command
|
||||||
models.Tool{
|
models.Tool{
|
||||||
Type: "function",
|
Type: "function",
|
||||||
|
|||||||
153
tui.go
153
tui.go
@@ -10,13 +10,15 @@ import (
|
|||||||
"path"
|
"path"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
|
||||||
|
|
||||||
"github.com/gdamore/tcell/v2"
|
"github.com/gdamore/tcell/v2"
|
||||||
"github.com/rivo/tview"
|
"github.com/rivo/tview"
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = sync.RWMutex{}
|
func isFullScreenPageActive() bool {
|
||||||
|
name, _ := pages.GetFrontPage()
|
||||||
|
return name != "main"
|
||||||
|
}
|
||||||
|
|
||||||
var (
|
var (
|
||||||
app *tview.Application
|
app *tview.Application
|
||||||
@@ -96,6 +98,7 @@ var (
|
|||||||
[yellow]Alt+8[white]: show char img or last picked img
|
[yellow]Alt+8[white]: show char img or last picked img
|
||||||
[yellow]Alt+9[white]: warm up (load) selected llama.cpp model
|
[yellow]Alt+9[white]: warm up (load) selected llama.cpp model
|
||||||
[yellow]Alt+t[white]: toggle thinking blocks visibility (collapse/expand <think> blocks)
|
[yellow]Alt+t[white]: toggle thinking blocks visibility (collapse/expand <think> blocks)
|
||||||
|
[yellow]Alt+i[white]: show colorscheme selection popup
|
||||||
|
|
||||||
=== scrolling chat window (some keys similar to vim) ===
|
=== scrolling chat window (some keys similar to vim) ===
|
||||||
[yellow]arrows up/down and j/k[white]: scroll up and down
|
[yellow]arrows up/down and j/k[white]: scroll up and down
|
||||||
@@ -108,74 +111,22 @@ var (
|
|||||||
[yellow]x[white]: to exit the table page
|
[yellow]x[white]: to exit the table page
|
||||||
|
|
||||||
=== filepicker ===
|
=== filepicker ===
|
||||||
[yellow]c[white]: (in file picker) set current dir as CodingDir
|
[yellow]s[white]: (in file picker) set current dir as FilePickerDir
|
||||||
[yellow]x[white]: to exit
|
[yellow]x[white]: to exit
|
||||||
|
|
||||||
=== shell mode ===
|
=== shell mode ===
|
||||||
[yellow]@match->Tab[white]: file completion (type @ in input to get file suggestions)
|
[yellow]@match->Tab[white]: file completion with relative paths (recursive, depth 3, max 50 files)
|
||||||
|
|
||||||
=== status line ===
|
=== status line ===
|
||||||
%s
|
%s
|
||||||
|
|
||||||
Press <Enter> or 'x' to return
|
Press <Enter> or 'x' to return
|
||||||
`
|
`
|
||||||
colorschemes = map[string]tview.Theme{
|
|
||||||
"default": tview.Theme{
|
|
||||||
PrimitiveBackgroundColor: tcell.ColorDefault,
|
|
||||||
ContrastBackgroundColor: tcell.ColorGray,
|
|
||||||
MoreContrastBackgroundColor: tcell.ColorSteelBlue,
|
|
||||||
BorderColor: tcell.ColorGray,
|
|
||||||
TitleColor: tcell.ColorRed,
|
|
||||||
GraphicsColor: tcell.ColorBlue,
|
|
||||||
PrimaryTextColor: tcell.ColorLightGray,
|
|
||||||
SecondaryTextColor: tcell.ColorYellow,
|
|
||||||
TertiaryTextColor: tcell.ColorOrange,
|
|
||||||
InverseTextColor: tcell.ColorPurple,
|
|
||||||
ContrastSecondaryTextColor: tcell.ColorLime,
|
|
||||||
},
|
|
||||||
"gruvbox": tview.Theme{
|
|
||||||
PrimitiveBackgroundColor: tcell.ColorBlack, // Matches #1e1e2e
|
|
||||||
ContrastBackgroundColor: tcell.ColorDarkGoldenrod, // Selected option: warm yellow (#b57614)
|
|
||||||
MoreContrastBackgroundColor: tcell.ColorDarkSlateGray, // Non-selected options: dark grayish-blue (#32302f)
|
|
||||||
BorderColor: tcell.ColorLightGray, // Light gray (#a89984)
|
|
||||||
TitleColor: tcell.ColorRed, // Red (#fb4934)
|
|
||||||
GraphicsColor: tcell.ColorDarkCyan, // Cyan (#689d6a)
|
|
||||||
PrimaryTextColor: tcell.ColorLightGray, // Light gray (#d5c4a1)
|
|
||||||
SecondaryTextColor: tcell.ColorYellow, // Yellow (#fabd2f)
|
|
||||||
TertiaryTextColor: tcell.ColorOrange, // Orange (#fe8019)
|
|
||||||
InverseTextColor: tcell.ColorWhite, // White (#f9f5d7) for selected text
|
|
||||||
ContrastSecondaryTextColor: tcell.ColorLightGreen, // Light green (#b8bb26)
|
|
||||||
},
|
|
||||||
"solarized": tview.Theme{
|
|
||||||
PrimitiveBackgroundColor: tcell.NewHexColor(0x1e1e2e), // #1e1e2e for main dropdown box
|
|
||||||
ContrastBackgroundColor: tcell.ColorDarkCyan, // Selected option: cyan (#2aa198)
|
|
||||||
MoreContrastBackgroundColor: tcell.ColorDarkSlateGray, // Non-selected options: dark blue (#073642)
|
|
||||||
BorderColor: tcell.ColorLightBlue, // Light blue (#839496)
|
|
||||||
TitleColor: tcell.ColorRed, // Red (#dc322f)
|
|
||||||
GraphicsColor: tcell.ColorBlue, // Blue (#268bd2)
|
|
||||||
PrimaryTextColor: tcell.ColorWhite, // White (#fdf6e3)
|
|
||||||
SecondaryTextColor: tcell.ColorYellow, // Yellow (#b58900)
|
|
||||||
TertiaryTextColor: tcell.ColorOrange, // Orange (#cb4b16)
|
|
||||||
InverseTextColor: tcell.ColorWhite, // White (#eee8d5) for selected text
|
|
||||||
ContrastSecondaryTextColor: tcell.ColorLightCyan, // Light cyan (#93a1a1)
|
|
||||||
},
|
|
||||||
"dracula": tview.Theme{
|
|
||||||
PrimitiveBackgroundColor: tcell.NewHexColor(0x1e1e2e), // #1e1e2e for main dropdown box
|
|
||||||
ContrastBackgroundColor: tcell.ColorDarkMagenta, // Selected option: magenta (#bd93f9)
|
|
||||||
MoreContrastBackgroundColor: tcell.ColorDarkGray, // Non-selected options: dark gray (#44475a)
|
|
||||||
BorderColor: tcell.ColorLightGray, // Light gray (#f8f8f2)
|
|
||||||
TitleColor: tcell.ColorRed, // Red (#ff5555)
|
|
||||||
GraphicsColor: tcell.ColorDarkCyan, // Cyan (#8be9fd)
|
|
||||||
PrimaryTextColor: tcell.ColorWhite, // White (#f8f8f2)
|
|
||||||
SecondaryTextColor: tcell.ColorYellow, // Yellow (#f1fa8c)
|
|
||||||
TertiaryTextColor: tcell.ColorOrange, // Orange (#ffb86c)
|
|
||||||
InverseTextColor: tcell.ColorWhite, // White (#f8f8f2) for selected text
|
|
||||||
ContrastSecondaryTextColor: tcell.ColorLightGreen, // Light green (#50fa7b)
|
|
||||||
},
|
|
||||||
}
|
|
||||||
)
|
)
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
|
// Start background goroutine to update model color cache
|
||||||
|
startModelColorUpdater()
|
||||||
tview.Styles = colorschemes["default"]
|
tview.Styles = colorschemes["default"]
|
||||||
app = tview.NewApplication()
|
app = tview.NewApplication()
|
||||||
pages = tview.NewPages()
|
pages = tview.NewPages()
|
||||||
@@ -286,6 +237,11 @@ func init() {
|
|||||||
statusLineWidget = tview.NewTextView().
|
statusLineWidget = tview.NewTextView().
|
||||||
SetDynamicColors(true).
|
SetDynamicColors(true).
|
||||||
SetTextAlign(tview.AlignCenter)
|
SetTextAlign(tview.AlignCenter)
|
||||||
|
// // vertical text center alignment
|
||||||
|
// statusLineWidget.SetDrawFunc(func(screen tcell.Screen, x, y, w, h int) (int, int, int, int) {
|
||||||
|
// y += h / 2
|
||||||
|
// return x, y, w, h
|
||||||
|
// })
|
||||||
// Initially set up flex without search bar
|
// Initially set up flex without search bar
|
||||||
flex = tview.NewFlex().SetDirection(tview.FlexRow).
|
flex = tview.NewFlex().SetDirection(tview.FlexRow).
|
||||||
AddItem(textView, 0, 40, false).
|
AddItem(textView, 0, 40, false).
|
||||||
@@ -308,7 +264,7 @@ func init() {
|
|||||||
pages.RemovePage(editMsgPage)
|
pages.RemovePage(editMsgPage)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
chatBody.Messages[selectedIndex].Content = editedMsg
|
chatBody.Messages[selectedIndex].SetText(editedMsg)
|
||||||
// change textarea
|
// change textarea
|
||||||
textView.SetText(chatToText(chatBody.Messages, cfg.ShowSys))
|
textView.SetText(chatToText(chatBody.Messages, cfg.ShowSys))
|
||||||
pages.RemovePage(editMsgPage)
|
pages.RemovePage(editMsgPage)
|
||||||
@@ -396,13 +352,14 @@ func init() {
|
|||||||
case editMode:
|
case editMode:
|
||||||
hideIndexBar() // Hide overlay first
|
hideIndexBar() // Hide overlay first
|
||||||
pages.AddPage(editMsgPage, editArea, true, true)
|
pages.AddPage(editMsgPage, editArea, true, true)
|
||||||
editArea.SetText(m.Content, true)
|
editArea.SetText(m.GetText(), true)
|
||||||
default:
|
default:
|
||||||
if err := copyToClipboard(m.Content); err != nil {
|
msgText := m.GetText()
|
||||||
|
if err := copyToClipboard(msgText); err != nil {
|
||||||
logger.Error("failed to copy to clipboard", "error", err)
|
logger.Error("failed to copy to clipboard", "error", err)
|
||||||
}
|
}
|
||||||
previewLen := min(30, len(m.Content))
|
previewLen := min(30, len(msgText))
|
||||||
notification := fmt.Sprintf("msg '%s' was copied to the clipboard", m.Content[:previewLen])
|
notification := fmt.Sprintf("msg '%s' was copied to the clipboard", msgText[:previewLen])
|
||||||
if err := notifyUser("copied", notification); err != nil {
|
if err := notifyUser("copied", notification); err != nil {
|
||||||
logger.Error("failed to send notification", "error", err)
|
logger.Error("failed to send notification", "error", err)
|
||||||
}
|
}
|
||||||
@@ -482,6 +439,19 @@ func init() {
|
|||||||
pages.RemovePage(helpPage)
|
pages.RemovePage(helpPage)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
// Allow scrolling keys to pass through to the TextView
|
||||||
|
switch event.Key() {
|
||||||
|
case tcell.KeyUp, tcell.KeyDown,
|
||||||
|
tcell.KeyPgUp, tcell.KeyPgDn,
|
||||||
|
tcell.KeyHome, tcell.KeyEnd:
|
||||||
|
return event
|
||||||
|
}
|
||||||
|
if event.Key() == tcell.KeyRune {
|
||||||
|
switch event.Rune() {
|
||||||
|
case 'j', 'k', 'g', 'G':
|
||||||
|
return event
|
||||||
|
}
|
||||||
|
}
|
||||||
return nil
|
return nil
|
||||||
})
|
})
|
||||||
//
|
//
|
||||||
@@ -560,6 +530,13 @@ func init() {
|
|||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
if event.Key() == tcell.KeyRune && event.Rune() == 'i' && event.Modifiers()&tcell.ModAlt != 0 {
|
||||||
|
if isFullScreenPageActive() {
|
||||||
|
return event
|
||||||
|
}
|
||||||
|
showColorschemeSelectionPopup()
|
||||||
|
return nil
|
||||||
|
}
|
||||||
if event.Key() == tcell.KeyF1 {
|
if event.Key() == tcell.KeyF1 {
|
||||||
// chatList, err := loadHistoryChats()
|
// chatList, err := loadHistoryChats()
|
||||||
chatList, err := store.GetChatByChar(cfg.AssistantRole)
|
chatList, err := store.GetChatByChar(cfg.AssistantRole)
|
||||||
@@ -672,11 +649,12 @@ func init() {
|
|||||||
// copy msg to clipboard
|
// copy msg to clipboard
|
||||||
editMode = false
|
editMode = false
|
||||||
m := chatBody.Messages[len(chatBody.Messages)-1]
|
m := chatBody.Messages[len(chatBody.Messages)-1]
|
||||||
if err := copyToClipboard(m.Content); err != nil {
|
msgText := m.GetText()
|
||||||
|
if err := copyToClipboard(msgText); err != nil {
|
||||||
logger.Error("failed to copy to clipboard", "error", err)
|
logger.Error("failed to copy to clipboard", "error", err)
|
||||||
}
|
}
|
||||||
previewLen := min(30, len(m.Content))
|
previewLen := min(30, len(msgText))
|
||||||
notification := fmt.Sprintf("msg '%s' was copied to the clipboard", m.Content[:previewLen])
|
notification := fmt.Sprintf("msg '%s' was copied to the clipboard", msgText[:previewLen])
|
||||||
if err := notifyUser("copied", notification); err != nil {
|
if err := notifyUser("copied", notification); err != nil {
|
||||||
logger.Error("failed to send notification", "error", err)
|
logger.Error("failed to send notification", "error", err)
|
||||||
}
|
}
|
||||||
@@ -731,6 +709,8 @@ func init() {
|
|||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyF12 {
|
if event.Key() == tcell.KeyF12 {
|
||||||
// help window cheatsheet
|
// help window cheatsheet
|
||||||
|
// Update help text with current status before showing
|
||||||
|
helpView.SetText(fmt.Sprintf(helpText, makeStatusLine()))
|
||||||
pages.AddPage(helpPage, helpView, true, true)
|
pages.AddPage(helpPage, helpView, true, true)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@@ -761,6 +741,9 @@ func init() {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyCtrlL {
|
if event.Key() == tcell.KeyCtrlL {
|
||||||
|
if isFullScreenPageActive() {
|
||||||
|
return event
|
||||||
|
}
|
||||||
// Show model selection popup instead of rotating models
|
// Show model selection popup instead of rotating models
|
||||||
showModelSelectionPopup()
|
showModelSelectionPopup()
|
||||||
return nil
|
return nil
|
||||||
@@ -774,6 +757,9 @@ func init() {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyCtrlV {
|
if event.Key() == tcell.KeyCtrlV {
|
||||||
|
if isFullScreenPageActive() {
|
||||||
|
return event
|
||||||
|
}
|
||||||
// Show API link selection popup instead of rotating APIs
|
// Show API link selection popup instead of rotating APIs
|
||||||
showAPILinkSelectionPopup()
|
showAPILinkSelectionPopup()
|
||||||
return nil
|
return nil
|
||||||
@@ -863,8 +849,9 @@ func init() {
|
|||||||
// Stop any currently playing TTS first
|
// Stop any currently playing TTS first
|
||||||
TTSDoneChan <- true
|
TTSDoneChan <- true
|
||||||
lastMsg := chatBody.Messages[len(chatBody.Messages)-1]
|
lastMsg := chatBody.Messages[len(chatBody.Messages)-1]
|
||||||
cleanedText := models.CleanText(lastMsg.Content)
|
cleanedText := models.CleanText(lastMsg.GetText())
|
||||||
if cleanedText != "" {
|
if cleanedText != "" {
|
||||||
|
// nolint: errcheck
|
||||||
go orator.Speak(cleanedText)
|
go orator.Speak(cleanedText)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -879,11 +866,17 @@ func init() {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyCtrlQ {
|
if event.Key() == tcell.KeyCtrlQ {
|
||||||
|
if isFullScreenPageActive() {
|
||||||
|
return event
|
||||||
|
}
|
||||||
// Show user role selection popup instead of cycling through roles
|
// Show user role selection popup instead of cycling through roles
|
||||||
showUserRoleSelectionPopup()
|
showUserRoleSelectionPopup()
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyCtrlX {
|
if event.Key() == tcell.KeyCtrlX {
|
||||||
|
if isFullScreenPageActive() {
|
||||||
|
return event
|
||||||
|
}
|
||||||
// Show bot role selection popup instead of cycling through roles
|
// Show bot role selection popup instead of cycling through roles
|
||||||
showBotRoleSelectionPopup()
|
showBotRoleSelectionPopup()
|
||||||
return nil
|
return nil
|
||||||
@@ -922,6 +915,7 @@ func init() {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
// Get files from ragdir
|
||||||
fileList := []string{}
|
fileList := []string{}
|
||||||
for _, f := range files {
|
for _, f := range files {
|
||||||
if f.IsDir() {
|
if f.IsDir() {
|
||||||
@@ -929,22 +923,14 @@ func init() {
|
|||||||
}
|
}
|
||||||
fileList = append(fileList, f.Name())
|
fileList = append(fileList, f.Name())
|
||||||
}
|
}
|
||||||
chatRAGTable := makeRAGTable(fileList)
|
// Get loaded files from vector DB
|
||||||
pages.AddPage(RAGPage, chatRAGTable, true, true)
|
loadedFiles, err := ragger.ListLoaded()
|
||||||
return nil
|
|
||||||
}
|
|
||||||
if event.Key() == tcell.KeyCtrlY { // Use Ctrl+Y to list loaded RAG files
|
|
||||||
// List files already loaded into the RAG system
|
|
||||||
fileList, err := ragger.ListLoaded()
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
logger.Error("failed to list loaded RAG files", "error", err)
|
logger.Error("failed to list loaded RAG files", "error", err)
|
||||||
if notifyerr := notifyUser("failed to list RAG files", err.Error()); notifyerr != nil {
|
loadedFiles = []string{} // Continue with empty list on error
|
||||||
logger.Error("failed to send notification", "error", notifyerr)
|
|
||||||
}
|
}
|
||||||
return nil
|
chatRAGTable := makeRAGTable(fileList, loadedFiles)
|
||||||
}
|
pages.AddPage(RAGPage, chatRAGTable, true, true)
|
||||||
chatLoadedRAGTable := makeLoadedRAGTable(fileList)
|
|
||||||
pages.AddPage(RAGLoadedPage, chatLoadedRAGTable, true, true)
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if event.Key() == tcell.KeyRune && event.Modifiers() == tcell.ModAlt && event.Rune() == '1' {
|
if event.Key() == tcell.KeyRune && event.Modifiers() == tcell.ModAlt && event.Rune() == '1' {
|
||||||
@@ -1004,13 +990,6 @@ func init() {
|
|||||||
}
|
}
|
||||||
// go chatRound(msgText, persona, textView, false, false)
|
// go chatRound(msgText, persona, textView, false, false)
|
||||||
chatRoundChan <- &models.ChatRoundReq{Role: persona, UserMsg: msgText}
|
chatRoundChan <- &models.ChatRoundReq{Role: persona, UserMsg: msgText}
|
||||||
// Also clear any image attachment after sending the message
|
|
||||||
go func() {
|
|
||||||
// Wait a short moment for the message to be processed, then clear the image attachment
|
|
||||||
// This allows the image to be sent with the current message if it was attached
|
|
||||||
// But clears it for the next message
|
|
||||||
ClearImageAttachment()
|
|
||||||
}()
|
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user