How did we get from the first AI models to today's cutting-edge language models, and what’s next for AI infrastructure?
In this episode, I sit down with Alex to explore the history of AI, from early perceptrons to GPT-4, and the often overlooked hardware and system engineering challenges behind modern AI.
We discuss why building and training AI models today is far more than just code and data, it’s about scaling infrastructure, managing distributed GPUs, and creating robust pipelines for fine-tuning and domain-specific AI.
Key highlights include:
If you’re curious about where AI is headed, and what it really takes to build models beyond the hype, this is a conversation you don’t want to miss!