Artificial neural networks were designed to emulate the human brain - and their insane performance on a wide range of tasks is pretty good evidence to support the comparison. Well, it's a bit more complicated than that, at least according to my guest Mark Ungless, former neuroscience lecturer at Imperial and Oxford and current Director of AI at the UK's Mental Health Innovations. Mark and I have collaborated on research for 5+ years, and I've long enjoyed his thoughts on the biology of the brain...
Feb 26, 2025•1 hr 12 min•Ep 4•Transcript available on Metacast Is the AI revolution we're experiencing going to push us into a future we can't imagine? Or will the pace of progress enable us to adjust along the way? Dan Shipper spends his time thinking and writing on these topics (and many others) as the founder and CEO of Every Media, a technology-focused publication trying to understand the future. Dan is also a lifelong coder and entrepreneur with a background in philosophy and its intersection with tech. Dan and I share a ton in common (beyond just our ...
Jan 23, 2025•51 min•Ep 3•Transcript available on Metacast GPT-3 didn't have much of a splash outside of the AI community, but it foreshadowed the AI explosion to come. Is o1 OpenAI's second GPT-3 moment? Machine Learning Researchers Guilherme Freire and Luka Smyth discuss OpenAI o1, it's impact, and it's potential. We discuss early impressions of o1, why inference-time compute and reinforcement learning matter in the LLM story, and the path from o1 to AI beginning to fulfill its potential. 00:00 Introduction and Welcome 00:22 Exploring O1: Initial Impr...
Oct 18, 2024•52 min•Ep 2•Transcript available on Metacast Dev Rishi is the founder and CEO of Predibase, the company behind Ludwig and LoRAX. Predibase just released LoRA Land , a technical report showing 310 models that can outcompete GPT-4 on specific tasks through fine-tuning. In this episode, Dev tries (pretty successfully) to convince me that fine-tuning is the future, while answering a bunch of interesting questions, like: Is fine-tuning hard? If LoRAX is a competitive advantage for you, why open-source it? Is model hosting becoming commoditized?...
May 24, 2024•52 min•Transcript available on Metacast Talfan Evans is a research engineer at DeepMind, where he focuses on data curation and foundational research for pre-training LLMs and multimodal models like Gemini. I ask Talfan: Will one model rule them all? What does "high quality data" actually mean in the context of LLM training? Is language model pre-training becoming commoditized? Are companies like Google and OpenAI keeping their AI secrets to themselves? Does the startup or open source community stand a chance next to the giants? Also c...
May 18, 2024•38 min•Transcript available on Metacast "Understanding what's going on in a model is important to fine-tune it for specific tasks and to build trust." Bhavna Gopal is a PhD candidate at Duke, research intern at Slingshot with experience at Apple, Amazon and Vellum. We discuss How adversarial robustness research impacts the field of AI explainability. How do you evaluate a model's ability to generalize? What adversarial attacks should we be concerned about with LLMs?
May 08, 2024•44 min•Transcript available on Metacast Chris Gagne manages AI research at Hume, which just released an expressive text-to-speech model in a super impressive demo. Chris and Daniel discuss AI and emotional understanding: How does “prosody” add a dimension to human communication? What is Hume hoping to gain by adding it to Human-AI communication? Do we want to interact with AI like we interact with humans? Or should the interaction models be different? Are we entering the Uncanny Valley phase of emotionally intelligent AI? Do LLMs actu...
Apr 19, 2024•40 min•Transcript available on Metacast Former OpenAI Research Scientist Joel Lehman joins to discuss the non-linear nature of technological progress and the present day implications of his book, Why Greatness Cannot Be Planned . Joel co-authored the book with Kenneth Stanley back in 2015. The two did ML research at OpenAI, Uber, and the University of Central Florida and wrote the book based on insights from their work. We discuss: AI time horizons, and the Implications for investors, researchers, and entrepreneurs Exploration vs expl...
Mar 22, 2024•47 min•Transcript available on Metacast “Where are the good AI products?” asks Varun Shenoy, ML engineer in his latest blog post . Varun and I talk through: What are the cool applications that exist? Why aren't there more of them? What do (the few) good AI application companies have in common? What technological or societal leaps are blocking the existence of more AI apps that matter? The optimist case and the pessimist case for the near future of AI. As Varun puts it, what if the emperor has no clothes? We'd love to hear what you thi...
Mar 12, 2024•47 min•Transcript available on Metacast ML Engineer and tech writer Donato Riccio wrote an article entitled "The End of RAG?" discussing what might replace Retrieval Augmented Generation in the near future. The article was received as highly controversial within the AI echo chamber, so I brought Donato on the podcast to discuss RAG, why people are so obsessed with vector databases, and the upcoming research in AI that might replace it. Takeaways: RAG is necessary due to LLMs' limited context window and scalability issues, and the need...
Feb 09, 2024•40 min•Transcript available on Metacast What’s going on with GPUs? We talk through the GPU bottleneck/supply gut, Meta’s apparent 600,000 H100-equivalents and the future of the GPU cloud. Neel Master is the CEO and founder of Cedana , enabling pause/migrate/resume for compute jobs. Neel is a serial entrepreneur, former founder of Engooden and angel investor. He started his career in ML research at MIT's CSAIL. Topics from this podcast include: Cedana's real-time save, migrate and resume for compute technology, enabling the migration o...
Feb 02, 2024•43 min•Transcript available on Metacast Founders of Lingopal, Deven Orie and Casey Schneider, join to talk about their startup story, developing real-time translation software for enterprises. Topics include: Why is translation so hard? How are enterprise and consumer AI products different (e.g. Google Translate vs Lingopal)? Should AI product companies be doing AI research? Is it safe to rely on open source? Share your thoughts with us at hello@slingshot.xyz or tweet us @slingshot_ai....
Jan 22, 2024•29 min•Transcript available on Metacast Founder of the SafeLlama community, Enoch Kan joins us today, to talk about safety in open source and medical AI. Enoch previously worked in AI for radiology, focused on mammography at Kheiron Medical. Enoch is an open source contributor, and his substack is called Cross Validated. Key topics they discuss include: New jailbreaks for LLMs appear every day. Does it matter? How do internet firewalls compare to AI “firewalls”? Why do human radiologists still exist? Would it be safe to replace them a...
Jan 12, 2024•37 min•Transcript available on Metacast Join Daniel Cahn on another SlingTalk episode with Kristian Freed (ex-CTO at Pariti and Elder), discussing the past, present and future of AI-assisted or AI-driven software. They talked about: The Evolution of Coding Tools: From basic text editors to advanced IDEs and the integration of AI tools like Co-Pilot. The Impact of AI on Software Development Practices: How AI is reshaping the way code is written and the process of software development. AI-Generated Code and Its Potential: Exploring the ...
Jan 05, 2024•33 min•Transcript available on Metacast In 1950, Alan Turing asked, “Can machines think?” He suggested the Imitation Game as a test to evaluate whether a machine can think, more commonly called the “Turing Test.” Today we ask, is the Turing Test outdated? Joining Slingtalks this week are Kristian Freed & Guilherme Freire, founding engineers at Slingshot. Guilherme argues against the Turing Test, Kristian argues in favor. Key topics they discuss include: A recent paper claims that GPT-4 comes close to passing the Turing Test. Is th...
Dec 15, 2023•45 min•Transcript available on Metacast Join Daniel Cahn on SlingTalks as he welcomes Jonathan Pedoeem (Founder of PromptLayer) to talk through Prompt Engineering. This episode offers an in-depth look into the past, present, and future of prompt engineering and the intricacies of crafting effective AI prompts. Key topics they discuss include: Is prompt engineering more art or more science? The role of “prompt engineer” and whether prompt engineering is a highly specialized skill or a skill as common as Googling Approaches to evaluatin...
Dec 08, 2023•39 min•Transcript available on Metacast Adam Kirsh (Head of Product & Engineering, Stealth Startup) joins Slingshot to talk about how AI is transforming investment due diligence. Beyond AI in diligence, we discuss: “Horizontal” and “vertical” business models, that start from a point solution Building products vs. building relationships, and on being an AI partner for the enterprise AI-native startups and the reinvention of business models for the AI age Making incremental progress as a startup vs. visionary top-down redesigns of e...
Nov 24, 2023•34 min•Transcript available on Metacast Ex-Datadog Founding PM, Ayush Kapur, joins Daniel Cahn on SlingTalks to talk through the overloaded term, "Human in the Loop". They hone in on the impact of both, emotional and philosophical aspects of human interactions, for instance, your interaction with a doctor, and how those services can be considered irreplaceable by AI. Key topics include: Human-in-the-loop as human review vs. partially automated processes Use cases where human-in-the-loop makes automation useless because humans have to ...
Nov 15, 2023•36 min•Transcript available on Metacast AI is increasingly doing the heavy lifting in our communications and content generation. On this episode, Guilherme Freire, Founding ML Engineer at Slingshot, joins the podcast to discuss the impact of AI-generated content. Some of the topics discussed: “Proof of Work” for humans, when AI makes personalization and connection too easy Potential for subpar AI-generated to put high-quality content creation out of business Hyper-personalized AI leading to greater divisiveness and bias; or on the fli...
Nov 08, 2023•40 min•Transcript available on Metacast Daniel hosts our machine learning research intern and Cambridge Masters student, Andy Lo, to talk about the present and future of ML programming. Topics include: PyTorch vs. TensorFlow vs. Jax vs MoJo No-code, low-code and pro-code for ML engineers The (frustrating) world of debugging ML code Have thoughts? We'd love to hear them! Drop an email at hello@slingshot.xyz or reach out on Twitter: @slingshot_ai....
Oct 30, 2023•33 min•Transcript available on Metacast In this episode, Daniel shares his perspective on the opportunities for the next wave of AI-native startups. Machine Learning isn’t just about sentiment classification, churn prediction, and revenue forecasting anymore. Generative models can simulate real intelligence. But hard problems continue to require hard solutions, and prompt engineering with retrieval augmented generation won’t be nearly enough. Have thoughts? We'd love to hear them! Drop an email at hello@slingshot.xyz or reach out on T...
Oct 25, 2023•8 min•Transcript available on Metacast Daniel hosts our Founding Engineer, Edwin Zhang to unravel the balance in Design Driven Developments. Key things they cover: The conundrums faced when balancing user wants with real, valuable needs - showcasing our stance on "Doing what people need, not just what they want." A peek into the futuristic vision of browsers like Arc and how we regard the Browser Company. The harmonious dance between engineering-driven, customer-driven, data-driven, and design-driven developments. The delicate art of...
Oct 16, 2023•25 min•Transcript available on Metacast In this tech talk, we dive deep into the technical specifics around LLM inference. The big question is: Why are LLMs slow? How can they be faster? And might slow inference affect UX in the next generation of AI-powered software? We jump into: Is fast model inference the real moat for LLM companies? What are the implications of slow model inference on the future of decentralized and edge model inference? As demand rises, what will the latency/throughput tradeoff look like? What innovations on the...
Oct 06, 2023•40 min•Transcript available on Metacast This episode delves into the ongoing debate of the competitiveness between open-source and closed-source models and the reasons behind Meta's decision to publish Llama2 with a permissive open-source license We cover: How much bigger can closed-sourced models be, compared to open-source? Are new competitor foundation models doomed, if Meta enters the game? What does “second place” look like, in open source? Can open-source datasets compete? Will very large open-source models soon become illegal?...
Sep 22, 2023•28 min•Transcript available on Metacast Ever seen a piece of work and thought, "Wow, a machine did that?" In our very first Slingtalk episode, we unravel the broad world of AI and where creativity plays a part in the process. We cover: - AI models and consciousness - Poetry in language models - Algorithms, novelty, and where inspiration comes from - Creativity and the part randomness plays in the process Journey through this fusion of code, model agents, and creativity! Have thoughts? We'd love to hear them! Drop an email at hello@sli...
Sep 12, 2023•14 min•Transcript available on Metacast