{"id":1405,"date":"2025-04-24T18:12:00","date_gmt":"2025-04-24T18:12:00","guid":{"rendered":"https:\/\/violethoward.com\/new\/ethically-trained-ai-startup-pleias-releases-new-small-reasoning-models-optimized-for-rag-with-built-in-citations\/"},"modified":"2025-04-24T18:12:00","modified_gmt":"2025-04-24T18:12:00","slug":"ethically-trained-ai-startup-pleias-releases-new-small-reasoning-models-optimized-for-rag-with-built-in-citations","status":"publish","type":"post","link":"https:\/\/violethoward.com\/new\/ethically-trained-ai-startup-pleias-releases-new-small-reasoning-models-optimized-for-rag-with-built-in-citations\/","title":{"rendered":"Ethically trained AI startup Pleias releases new small reasoning models optimized for RAG with built-in citations"},"content":{"rendered":" \r\n
\n\t\t\t\t
\n

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More<\/em><\/p>\n\n\n\n


\n<\/div>

French AI startup Pleias made waves late last year with the launch of its ethically trained Pleias 1.0 family of small language models \u2014 among the first and only to date to be built entirely on scraping \u201copen\u201d data, that is, data explicitly labeled as public domain, open source, or unlicensed and not copyrighted. <\/p>\n\n\n\n

Now the company has announced the release of two open source small-scale reasoning models designed specifically for retrieval-augmented generation (RAG), citation synthesis, and structured multilingual output. <\/p>\n\n\n\n

The launch includes two core models \u2014 Pleias-RAG-350M and Pleias-RAG-1B \u2014 each also available in CPU-optimized GGUF format, making a total of four deployment-ready variants. <\/p>\n\n\n\n

They are all based on Pleias 1.0, and can be used independently or in conjunction with other LLMs that the organization may already or plan to deploy. All appear to be available under a permissive Apache 2.0 open source license, meaning they are<\/em> eligible for organizations to take, modify and deploy for commercial use cases.<\/p>\n\n\n\n

RAG, as you\u2019ll recall, is the widely-used technique that enterprises and organizations can deploy to hook an AI large language model (LLM) such as OpenAI\u2019s GPT-4o, Google\u2019s Gemini 2.5 Flash, Anthropic\u2019s Claude Sonnet 3.7 or Cohere\u2019s Command-A, or open source alternatives like Llama 4 and DeepSeek V3 to external knowledge bases, such as enterprise documents and cloud storages. <\/p>\n\n\n\n

This is often necessary for enterprises that want to build chatbots and other AI applications that reference their internal policies or product catalogs (an alternative, prompting a long context LLM with all the information necessary, may not be suitable for enterprise use cases where security and per-token transmission costs are concerns). <\/p>\n\n\n\n

The Pleias-RAG model family is the latest effort to bridge the gap between accuracy and efficiency in small language models.<\/p>\n\n\n\n

These models are aimed at enterprises, developers, and researchers looking for cost-effective alternatives to large-scale language models without compromising traceability, multilingual capabilities, or structured reasoning workflows.<\/p>\n\n\n\n

The target userbase is actually Pleias\u2019s home continent of Europe, as co-founder Alexander Doria told VentureBeat via direct message on the social network X:<\/p>\n\n\n\n

\u201cA primary motivation has been the difficulty of scaling RAG applications in Europe. Most private organization have little GPUs (it may have changed but not long ago less than 2% of all [Nvidia] H100 [GPUs] were in Europe). And yet simultaneously there are strong incentive to self-host for regulated reasons, including GDPR.<\/em><\/p>\n\n\n\n

\u201cSLMs have progressed significantly over the past year, yet they are too often conceived as \u2018mini-chatbots\u2019 and we have observed a significant drop of performance in non-English languages, both in terms of source understanding and quality of text generation. So we have been satisfied to hit most of our objectives: <\/em><\/p>\n\n\n\n