33 lines
1,012 B
Markdown
33 lines
1,012 B
Markdown
# Redvau.lt AI Monorepo
|
|
|
|
## Short Term Todo
|
|
- [x] Prepare proxy man for embedding
|
|
- [-] Improve markdown rendering in forge chat
|
|
- [ ] Embed proxy man & add simple ui in forge
|
|
- [ ] dumb embed process on startup
|
|
- [ ] View current instances/models/virtual endpoints
|
|
- [ ] Edit
|
|
- [ ] Add new (from configurable model folder)
|
|
|
|
|
|
## Current Repos
|
|
|
|
#### llama-forge-rs:
|
|
|
|
- old alpha-state webview based GUI app for chatting with ai
|
|
- basic streaming
|
|
- manages a llama-server in the background
|
|
- currently unmaintained
|
|
|
|
#### llama-proxy-man:
|
|
|
|
- proxy which auto starts/stops llama.cpp instances for you
|
|
- retries requests in the background and keeps the connection open
|
|
- enables you to run more models than your VRAM can fit, with HTTP-API requests working like they do (they'll just be slow if you have to start instances to make them happen)
|
|
|
|
## Ideas
|
|
|
|
- emacs-qwen-plugin
|
|
- use emacs-module-rs + llama.cpp/proxy to get awesome qwen integration into emacs
|
|
- agent
|
|
- add experimental rag+agent framework
|