Current LLM tech stack
because people ask, not because I recommend
November 2024:
For coding, I use Cody in my local IDE VSCode and whatever is on offer in cloud IDEs (recently CoPilot and ReplitAI).
For getting my code to talk with LLMs, I use Simon Willison's llm
.
For running models in the cloud, I use Replicate. I've used Ollama locally – but honestly I haven't touched a local model recent months. I use DrawThings to explore image-making locally (but while generative, that's not using an LLM).
For general stuff (including explaining code and learning new stuff), I currently use msty, which allows chat-like interfaces with several AIs via their API. This lets me set up the context for a chat with system prompts, and handily means that I can pay-as-I-go with AIs, which is rather cheaper for me than a monthly subscription.
I'm currently using Claude-3.5-Sonnet, GPT 4o and GPT 4o-mini via their API
Comments
Sign in or become a Workroom Productions member to read and leave comments.