Found 401 bookmarks
Newest
Impeccable: Design skills for AI harnesses
Impeccable: Design skills for AI harnesses
1 skill, 23 commands, and curated anti-patterns for impeccable frontend design. Works with Cursor, Claude Code, Gemini CLI, and Codex CLI.
·impeccable.style·
Impeccable: Design skills for AI harnesses
Harness teams of coding agents with Squad
Harness teams of coding agents with Squad
How do we fix code fast when the bug reports arrive faster? Multi-agent orchestration tools like Squad may be the answer.
·infoworld.com·
Harness teams of coding agents with Squad
Making AI work through eval hygiene
Making AI work through eval hygiene
Most companies don't have an AI quality problem. They have a measurement problem.
·infoworld.com·
Making AI work through eval hygiene
Small language models: Rethinking enterprise AI architecture
Small language models: Rethinking enterprise AI architecture
As LLMs hit the limits of scale and cost, specialized SLMs are emerging as the faster, cheaper, and more private workhorse for the autonomous enterprise.
·infoworld.com·
Small language models: Rethinking enterprise AI architecture
Local AI
Local AI
Running LLMs on your own hardware
·oreilly.com·
Local AI
Trying Pair Programming With An LLM Chatbot
Trying Pair Programming With An LLM Chatbot
When it comes to software developers, there are a few distinct types. For example, the extroverted, chatty type, who is always going out there to share the latest and newest libraries and projects …
·hackaday.com·
Trying Pair Programming With An LLM Chatbot
undefined
undefined
The agent-led growth playbook: how to make AI agents discover, use, and pay for your developer tool, and defend against the ones you didn't invite. LLM discoverability, agent-first onboarding, agent payments, AX security.
·evilmartians.com·
undefined
How I use AI in 2026
How I use AI in 2026
How I use AI in my daily work as a maintainer and developer, from coding to triaging PRs and CI failures
·fedepaol.github.io·
How I use AI in 2026
My bets on open models, mid-2026
My bets on open models, mid-2026
What I expect to come next and why, focused on the open-closed gap.
·interconnects.ai·
My bets on open models, mid-2026
How to forget
How to forget
Most agent frameworks optimize for recall. Open-strix optimizes for forgetting — and that turns out to be the whole trick.
·timkellogg.me·
How to forget
undefined
undefined
Most AI SEO advice is unproven. We tested what ChatGPT, Claude, and Perplexity actually read on our own site. Six LLM visibility techniques that worked, eight that didn't, and the metrics to tell the difference.
·evilmartians.com·
undefined
Using a local LLM in OpenCode with llama.cpp – Aayush Garg
Using a local LLM in OpenCode with llama.cpp – Aayush Garg
Step-by-step setup for running a quantized Qwen3.5-27B model on a remote GPU via llama.cpp, exposing it over Tailscale and using it as a provider in OpenCode (optionally with Codex).
·aayushgarg.dev·
Using a local LLM in OpenCode with llama.cpp – Aayush Garg
StrongDM Software Factory
StrongDM Software Factory
StrongDM's field notes on non-interactive agentic development: specs + scenarios, validation harnesses, feedback loops, and the supporting components.
·factory.strongdm.ai·
StrongDM Software Factory
Running LLaMA Locally with Llama.cpp: A Complete Guide
Running LLaMA Locally with Llama.cpp: A Complete Guide
Llama.cpp is a powerful and efficient inference framework for running LLaMA models locally on your machine. Unlike other tools such as…
·medium.com·
Running LLaMA Locally with Llama.cpp: A Complete Guide
The Complete Developer's Guide to Running LLMs Locally
The Complete Developer's Guide to Running LLMs Locally
A comprehensive guide covering the local LLM stack from hardware requirements to production deployment. Compare Ollama, LM Studio, llama.cpp and build your first local AI application.
·sitepoint.com·
The Complete Developer's Guide to Running LLMs Locally
LM Studio Developer Docs | LM Studio Docs
LM Studio Developer Docs | LM Studio Docs
Build with LM Studio's local APIs and SDKs — TypeScript, Python, REST, and OpenAI and Anthropic-compatible endpoints.
·lmstudio.ai·
LM Studio Developer Docs | LM Studio Docs
Your GPUs Just Got 6x More Valuable. No New Hardware Required.
Your GPUs Just Got 6x More Valuable. No New Hardware Required.
Watch now | The variable that decides who wins the AI infrastructure war isn’t a faster chip or a better model. It’s a compression algorithm.
·natesnewsletter.substack.com·
Your GPUs Just Got 6x More Valuable. No New Hardware Required.
Who’s the Admin, Me or Claude?
Who’s the Admin, Me or Claude?
Credit: Museums Victoria / Unsplash There’s a lot of conversation right now about “context engineering” for dev work; structuring what you feed an LLM so it can do useful things. …
·cate.blog·
Who’s the Admin, Me or Claude?
Mastering Caching Methods in Large Language Models (LLMs)
Mastering Caching Methods in Large Language Models (LLMs)
Large Language Models (LLMs) like OpenAI’s GPT-4 have transformed natural language processing, enabling applications ranging from chatbots…
·masteringllm.medium.com·
Mastering Caching Methods in Large Language Models (LLMs)
How to Implement Effective LLM Caching
How to Implement Effective LLM Caching
A deep dive into effective caching strategies for building scalable and cost-efficient LLM applications, covering exact key vs. semantic caching, architectural patterns, and practical implementation tips.
·helicone.ai·
How to Implement Effective LLM Caching