Enha (llama.cpp): show loaded model on startup

This commit is contained in:
Grail Finder
2026-03-01 08:22:02 +03:00
parent 1f112259d2
commit cdfccf9a24
6 changed files with 37 additions and 43 deletions

12
models/consts.go Normal file
View File

@@ -0,0 +1,12 @@
package models
const (
LoadedMark = "(loaded) "
)
type APIType int
const (
APITypeChat APIType = iota
APITypeCompletion
)

View File

@@ -519,24 +519,6 @@ type OpenAIReq struct {
// ===
// type LLMModels struct {
// Object string `json:"object"`
// Data []struct {
// ID string `json:"id"`
// Object string `json:"object"`
// Created int `json:"created"`
// OwnedBy string `json:"owned_by"`
// Meta struct {
// VocabType int `json:"vocab_type"`
// NVocab int `json:"n_vocab"`
// NCtxTrain int `json:"n_ctx_train"`
// NEmbd int `json:"n_embd"`
// NParams int64 `json:"n_params"`
// Size int64 `json:"size"`
// } `json:"meta"`
// } `json:"data"`
// }
type LlamaCPPReq struct {
Model string `json:"model"`
Stream bool `json:"stream"`
@@ -641,10 +623,3 @@ type ChatRoundReq struct {
Regen bool
Resume bool
}
type APIType int
const (
APITypeChat APIType = iota
APITypeCompletion
)