Skip to content

Musings on AI

Musings on AI

  • Home
    • Home
    • Llama.cpp
Llama.cpp Mac Oobabooga

Llama 3 on Web UI

Dracones April 22, 2024

When doing inference with Llama 3 Instruct on Text Generation Web UI, up front you can get pretty decent inference speeds on a the M1 Mac Ultra, even with a…

Llama.cpp Mac Nvidia

The Case for Mac: Power Usage

Dracones April 22, 2024

I have two dual Nvidia 3090 Linux servers for inference and they’ve worked very well for running large language models. 48GB of VRAM will load models up to 70B at…

Recent Posts

  • Adventures in Training: Axolotl
  • Llama 3 on Web UI
  • The Case for Mac: Power Usage
  • Text Generation Web UI(llama.cpp)
  • Mac for a Linux Nerd

Recent Comments

No comments to show.

Musings on AI