It's been feeling like the local LLM space has really slowed over the past 6 months, curious if it's just me or others have felt it too. If you're using local models, which ones and for what?
The moat right now in #ai is the model, the moat in a year will be memory (how well the system remembers YOU and your queries and picks the right prompt and model, aka we are headed back to recsys real soon.)
Thereβs a reason ChatGPT started storing user session data although itβs risky and annoying from an engineering and legal perspective, and why Meta is putting semantic and hybrid search in all its products even though it is orthogonal with a good search experience in those interfaces.