No description
Find a file
2025-02-11 23:17:20 +01:00
.dist Cleanup llama_forge_rs stuff for monorepo 2025-02-11 00:58:14 +01:00
.forgejo/workflows Cleanup llama_forge_rs stuff for monorepo 2025-02-11 00:58:14 +01:00
.nix Fix llama_forge_rs build & maintain a bit 2025-02-10 15:51:46 +01:00
llama_forge_rs WIP 3dbutton 2025-02-11 23:17:20 +01:00
llama_proxy_man feat: Embedded proxy_man for forge 2025-02-11 04:24:43 +01:00
redvault_el_rs feat: Embedded proxy_man for forge 2025-02-11 04:24:43 +01:00
.envrc Init 2024-07-21 02:42:48 +02:00
.gitignore Fix llama_forge_rs build & maintain a bit 2025-02-10 15:51:46 +01:00
Cargo.lock feat: Embedded proxy_man for forge 2025-02-11 04:24:43 +01:00
Cargo.toml Fix llama_forge_rs build & maintain a bit 2025-02-10 15:51:46 +01:00
clippy.yml Remove unused pkg 2024-09-09 15:49:41 +02:00
config.yaml feat: Embedded proxy_man for forge 2025-02-11 04:24:43 +01:00
flake.lock chore: Update llama.cpp dependency to b4681 2025-02-10 16:05:26 +01:00
flake.nix chore: Update llama.cpp dependency to b4681 2025-02-10 16:05:26 +01:00
Makefile.toml Try another sitedir fix 2025-02-11 16:43:39 +01:00
package-lock.json Fix llama_forge_rs build & maintain a bit 2025-02-10 15:51:46 +01:00
package.json Fix llama_forge_rs build & maintain a bit 2025-02-10 15:51:46 +01:00
README.md docs: Update README with short term todo list 2025-02-10 23:44:25 +01:00
rust-toolchain.toml Update deps 2025-01-31 13:19:28 +01:00

Redvau.lt AI Monorepo

Short Term Todo

  • Prepare proxy man for embedding
  • [-] Improve markdown rendering in forge chat
  • Embed proxy man & add simple ui in forge
    • dumb embed process on startup
    • View current instances/models/virtual endpoints
    • Edit
    • Add new (from configurable model folder)

Current Repos

llama-forge-rs:

  • old alpha-state webview based GUI app for chatting with ai
  • basic streaming
  • manages a llama-server in the background
  • currently unmaintained

llama-proxy-man:

  • proxy which auto starts/stops llama.cpp instances for you
  • retries requests in the background and keeps the connection open
  • enables you to run more models than your VRAM can fit, with HTTP-API requests working like they do (they'll just be slow if you have to start instances to make them happen)

Ideas

  • emacs-qwen-plugin
    • use emacs-module-rs + llama.cpp/proxy to get awesome qwen integration into emacs
  • agent
    • add experimental rag+agent framework