- One Cerebral
- Posts
- What is the Future of Open Source AI
What is the Future of Open Source AI
Co-Founders at Nous Research: Bowen Peng and Jeffrey Quesnelle
Credit and Thanks:
Based on insights from a16z.
Today’s Podcast Host: Anjney Midha (a16z)
Title
The Quest for Community-Trained Open Source AI Models
Guest
Bowen Peng & Jeffrey Quesnelle
Guest Credentials
Bowen Peng is a co-founder of Nous Research and a former MS student at the Université de Montréal, focusing on machine learning, computer graphics, and generative models. He has made significant contributions to the field of AI, including co-authoring the YaRN paper on efficient context window extension for large language models, which has been cited over 160 times. Peng's work on NTK-aware scaled RoPE has enabled LLaMA models to extend their context size without fine-tuning, demonstrating his expertise in advancing AI capabilities.
Jeffrey Quesnelle is a co-founder of Nous Research, working alongside Bowen Peng to accelerate open-source AI research and development. He played a key role in the creation of DisTrO, a project that demonstrated the possibility of training AI models across the public internet much faster than previously thought possible. Quesnelle has also been instrumental in developing the popular Hermes family of "neutral" and guardrail-free language models, showcasing his commitment to pushing the boundaries of AI technology.
Podcast Duration
1:16:33
This Newsletter Read Time
Approx. 5 mins
Brief Summary
Bowen Peng and Jeffrey Quesnelle discussed their innovative approach to artificial intelligence, emphasizing the importance of open-source collaboration and the potential for decentralized training models. They highlighted their project, DisTrO, which aims to democratize AI by enabling individuals to contribute to model training using standard internet connections, thus breaking away from the traditional reliance on centralized data centers. The conversation underscored the transformative potential of AI when accessible to a broader community, fostering creativity and innovation.
Deep Dive
At Nous Research, the duo has been instrumental in developing the Hermes series of AI models, which are designed to be neutrally aligned and allow users to instruct the model to adopt various personas. This approach contrasts with many existing AI models that are often limited by strict guidelines and predefined roles. The team’s decision-making process regarding what to work on is driven by a desire to tackle fundamental research questions that can yield significant advancements with minimal resources. They focus on identifying bottlenecks in AI development, such as the slow and expensive process of human data collection, which they addressed by pioneering the use of synthetic data generated by AI models themselves. This innovative approach has since become widely accepted in the field.
The catalyst for their latest initiative, DisTrO, arose from concerns about the future of open-source AI. As the landscape of AI development becomes increasingly dominated by large corporations, there is a risk that the pace of innovation could slow down. DisTrO aims to counter this trend by enabling the training of AI models over the internet, allowing individuals and smaller organizations to contribute their computing power to the process. This decentralized approach not only democratizes access to AI but also fosters a collaborative environment where diverse contributions can lead to groundbreaking advancements.
One of the most striking achievements of the DisTrO research has been the demonstration of an 857x reduction in bandwidth requirements for training AI models. This significant breakthrough allows for the training of complex models using standard internet connections, rather than relying on the high-speed interconnects typically found in centralized data centers. This shift has the potential to revolutionize the AI landscape, making it accessible to a broader range of contributors.
However, skepticism remains regarding the feasibility of such a decentralized model. Bowen and Jeffrey actively encourage replication of their findings, understanding that validation from the community is crucial for building credibility. They acknowledge that the initial reaction to their results has been one of disbelief, but they remain committed to transparency and collaboration, inviting others to test and verify their methods.
The concept of a SETI@home for AI, where individuals can contribute their computing resources to train models, raises questions about its impact on the GPU market. If successful, this model could shift the demand from high-end GPUs, typically used in centralized data centers, to more accessible consumer-grade hardware. This democratization of AI training could lead to a more equitable distribution of resources and opportunities in the field.
As they develop this "product" organically, Bowen and Jeffrey emphasize the importance of community involvement. They envision a future where anyone can participate in training AI models, contributing their computing power and expertise. This collaborative spirit is at the heart of their mission, as they seek to create a platform that empowers individuals to engage with AI technology.
Unexpected discoveries during the DisTrO research have also shaped their approach. For instance, they found that rather than requiring all GPUs to synchronize and share their models, they could allow each GPU to train independently while still sharing insights. This decentralized training method not only enhances efficiency but also opens up new avenues for exploration and innovation.
Orchestrating distributed training runs presents its own set of challenges, particularly in terms of communication and coordination among the various nodes. The team is exploring ways to optimize this process, ensuring that each participant can contribute effectively without overwhelming the network. They recognize that while backpropagation remains a critical component of training, the flexibility of their approach allows for a more dynamic and responsive system.
Looking ahead, the potential for ASICs (Application-Specific Integrated Circuits) to play a role in inference is an exciting prospect. As the technology evolves, there may be opportunities to leverage specialized hardware for both training and inference, further enhancing the efficiency and accessibility of AI development.
Key Takeaways
The DisTrO project aims to enable decentralized AI training using standard internet connections, reducing reliance on centralized data centers.
Community engagement is crucial for fostering innovation in AI, allowing diverse perspectives to contribute to advancements.
The shift from synchronized training to independent model training can enhance efficiency and flexibility in AI development.
Actionable Insights
Individuals interested in AI can participate in open-source projects like DisTrO, contributing their computing power to help train models.
Organizations can explore decentralized training methodologies to reduce costs and increase accessibility to AI technologies.
Thought leaders should advocate for policies that support open-source collaboration in AI, ensuring that advancements benefit a wider audience.
Why it’s Important
The discussion highlights the critical need for democratizing access to AI technologies, which can lead to more equitable innovation. By breaking down barriers to entry, the potential for diverse contributions can drive advancements that reflect a broader range of human experiences and needs. This shift is essential in ensuring that AI serves the interests of society as a whole rather than a select few.
What it Means for Thought Leaders
For thought leaders, the insights underscore the importance of fostering inclusive environments in technology development. They must recognize the value of community-driven initiatives and advocate for frameworks that support open-source collaboration. This approach not only enhances innovation but also ensures that the benefits of AI are distributed more equitably across society.
Mind Map

Key Quote
"We're trying to make AI that is representative of the whole world and that is personable to you."
Future Trends & Predictions
As the landscape of AI continues to evolve, we can expect a growing trend towards decentralized training models that leverage community resources. This shift may lead to the emergence of new platforms that facilitate collaborative AI development, allowing individuals to contribute their computing power and expertise. Additionally, as more organizations adopt these methodologies, we may see a democratization of AI technologies, resulting in a more diverse range of applications and innovations that cater to various societal needs. The future of AI could very well be defined by the collective efforts of a global community rather than a few dominant players.
Check out the podcast here:
Latest in AI
1. UBTech and Foxconn have signed a "comprehensive long-term partnership" to deploy humanoid robots, specifically the Walker S1, in iPhone manufacturing facilities, with the robot already completing two months of training for logistics operations in Shenzhen's Longhua district. The Walker S1 humanoid robot is being trained to handle increasingly complex and delicate production tasks in consumer electronics, with UBTech aiming to launch an upgraded Walker S2 version later this year to further enhance robotic manufacturing capabilities.
2. The University of Bristol has unveiled Isambard-AI, a £225 million ($276 million) supercomputer that leverages artificial intelligence to revolutionize drug and vaccine development by simulating molecular interactions within the human body at an atomic level. Set to become the UK's most powerful AI-enabled supercomputer when fully operational this summer, Isambard-AI is already being used to develop treatments for conditions like Alzheimer's, heart disease, and various cancers, as well as to enhance melanoma detection across diverse skin tones.
3. ByteDance has launched Trae, an AI-powered integrated development environment (IDE) for macOS that enables developers to automate code generation and project development through natural language interactions, currently supporting both Simplified Chinese and English interfaces. Powered by GPT-4o and Claude-3.5-Sonnet models and distributed by ByteDance's Singapore subsidiary, Trae offers features like AI-assisted Q&A, code completion, and a unique Builder Mode that can construct entire projects from scratch, positioning itself as a competitor to established tools like Cursor and Visual Studio Code.
Startup World
1. Eximius Ventures launched its second fund (Fund II) with a target corpus of $30 million to invest in early-stage startups across fintech, AI/SaaS, frontier tech, and consumer tech. The VC firm aims to invest in 25-30 companies with an initial $500,000 cheque per company, reserving half of the corpus for follow-on investments. Fund II has already invested in four companies in consumer tech and AI/SaaS, with Eximius Ventures leading or co-leading these investments alongside institutional investors.
2. Cohere Capital Partners, a Boston-based private equity firm, closed a $215 million fund in partnership with TPG NEXT. The firm focuses on investing in leading tech-enabled lower middle market companies, although specific details about the investments were not provided.
3. CyGO Entrepreneurs, a Lyon-based startup empowering European founders in data analytics and cybersecurity, raised $5 million in seed funding. The company, which has 24 employees, received investment from BNP Paribas Développement, although complete details of other investors were not disclosed.
Analogy
Building DisTrO is like turning a walled garden into a community farm. Traditionally, only large corporations with vast resources could tend the "AI garden," but Bowen and Jeffrey's decentralized approach invites anyone with a shovel—be it a modest GPU or spare time—to contribute. By solving bandwidth bottlenecks and fostering collaboration, they’re planting the seeds for a democratized AI ecosystem, where breakthroughs sprout from collective effort rather than exclusive ownership.
Thanks for reading, have a lovely day!
Jiten-One Cerebral
All summaries are based on publicly available content from podcasts. One Cerebral provides complementary insights and encourages readers to support the original creators by engaging directly with their work; by listening, liking, commenting or subscribing.
Reply