In a groundbreaking move, Silo AI has unveiled Poro, the first model in a planned family of open-source models intended to cover all 24 official European Union languages. This innovative development promises to revolutionize multilingual artificial intelligence (AI) and provide a transparent and ethical alternative to proprietary systems from major tech companies.
Porosity Model Details
The Poro 34B model takes center stage, boasting an impressive 34.2 billion parameters. Named after the Finnish word for “reindeer,” this model utilizes a cutting-edge BLOOM transformer architecture with ALiBi embeddings. With its vast size and sophisticated architecture, Poro aims to achieve state-of-the-art performance in natural language processing tasks across multiple languages.
Transparency and Documentation
SiloGen, the driving force behind Poro, is committed to transparency. To that end, they have introduced the Poro Research Checkpoints program, offering documentation of the model’s training progress. The initial checkpoint covers the first 30% of training, and benchmarks released by Silo AI demonstrate that Poro is already achieving state-of-the-art results even at this early stage.
Multilingual Capabilities and Language Diversity
One of the key strengths of Poro is its ability to leverage shared patterns across related languages. This advantage allows the model to excel even in languages with limited training data available. Poro’s multilingual capabilities have not come at the expense of its prowess in English, making it a truly versatile and powerful tool for natural language processing across diverse linguistic contexts.
The future of AI
The CEO of Silo AI, Peter Sarlin, firmly believes that open-source models like Poro represent the future of AI. These models provide a transparent and ethical alternative to closed models from major tech companies, fostering greater trust and inclusivity within the AI community. By embracing an open approach, Poro sets a new standard for AI development built on collaboration, fairness, and societal impact.
Poro’s Expansion and Releases
Silo AI has ambitious plans to further develop the Poro family of models. Regular Poro checkpoints will be released throughout the training process, ultimately covering all European languages. This iterative approach ensures continuous improvement and maintains Poro’s relevance and effectiveness as new language data becomes available.
Democratizing Access to Multilingual Models
Poro’s promise lies in its potential to democratize access to performant multilingual models. By providing Europe with a homegrown alternative to systems from dominant US tech companies, Poro aims to level the playing field and foster innovation within the European AI community. This shift in power dynamics could have far-reaching consequences, offering greater control and ownership over AI technologies.
Collaboration with the University and Research
Silo AI’s partnership with the University brings together Silo AI’s expertise in applied AI and computational resources with the University’s leadership in multilingual language modeling research. This collaboration ensures that Poro’s development is grounded in academic rigor and real-world applications, combining theoretical advancements with practical implementation.
Poro represents a significant milestone in the quest for open-source, multilingual AI models. With its transformative capabilities, robust architecture, and commitment to transparency, Poro is poised to disrupt the AI landscape. By democratizing access to high-performing multilingual models, it has the potential to shape a future where AI fosters collaboration, inclusivity, and innovation. As Poro expands and paves the way for other open-source models, Europe’s AI community gains a powerful tool in its pursuit of linguistic diversity and AI excellence.