
ARCEE.AI small language models, open source and cost efficient AI | AWS for AI podcast
Failed to add items
Add to basket failed.
Add to wishlist failed.
Remove from wishlist failed.
Adding to library failed
Follow podcast failed
Unfollow podcast failed
-
Narrated by:
-
By:
About this listen
Join us for an enlightening conversation with Julien Simon, VP and Chief Evangelist at ARCEE.AI , as he shares deep insights on building practical and cost-efficient AI solutions. From his extensive experience at AWS, Hugging Face, and now ARCEE.AI, Julien discusses why "small is beautiful" when it comes to language models, revealing how 10B parameter models can now match the performance of much larger 72B models from just months ago. Learn about innovative techniques like model merging, the importance of proper infrastructure choices, and practical advice for organizations starting their AI journey. This episode covers critical topics including:
- Why small language models are the future of enterprise AI
- How to optimize costs while maintaining performance
- The role of CPU vs GPU inference
- Essential architecture considerations for AI workloads
- Best practices for building production-ready AI systems
Whether you're a startup, enterprise, or public sector organization, this episode offers invaluable guidance on building scalable, efficient, and practical AI solutions in today's rapidly evolving landscape.
Julien Simon Youtube channel : https://www.youtube.com/@juliensimonfr
to learn more about ARCEE.AI : https://www.arcee.ai/