More efficient, streamlined, and useful AI: at the heart of CRI’s research

Digital transformation Research Decoding
Published on 20 February 2026
From large language models capable of analyzing hundreds of thousands of pages to supercomputers that consume as much electricity as a small town, artificial intelligence (AI) today faces a critical equation: how to continue to improve performance without causing computing, memory, and energy costs to skyrocket? At the Center for Research in Computer Science (CRI) at Mines Paris – PSL, researchers are tackling this fundamental challenge in an international context that promotes AI for the benefit of people, the planet, and progress, as exemplified by theIndia AI Impact Summit 2026 to be held in New Delhi from February 16 to 20.
Led by researcher Corinne Ancourt, the CRI conducts research combining AI, high-performance computing (HPC), and software optimization. Presented at the AI Workshop held at Mines Paris – PSL on December 10, 2025, this work has a clear objective: to design methods and tools that enable the efficient, frugal, and secure deployment of applications, particularly artificial intelligence, for the benefit of science, industry, and society.

Performance, memory, energy

AI under pressure

AI has continued to evolve over the decades. Initially based on expert systems relying on hand-coded rules, it underwent its first revolution with machine learning, which requires large amounts of data and parallel computing architectures. The rise of deep learning, followed by large language models (LLMs) and neural networks now with hundreds of billions of parameters, has profoundly changed the scale of AI.

This evolution highlights a central tension: the more powerful the models, the more memory, computing power, and energy they require.

However, these resources are neither infinite nor neutral. The energy cost of AI has become a major scientific, industrial, and environmental issue. It is precisely in this area that CRI’s research is focused.

At CRI, doing better with the same models

Contrary to popular belief, innovating in AI does not always mean creating new models. Much of CRI’s work consists of optimizing existing models:

  • Choosing better models
  • Configuring them better
  • Deploying them better on complex computing architectures (CPUs, GPUs, clusters, supercomputers)
  • Without compromising the quality of the results

The research is organized around three complementary areas

  1. Developing AI methods at the heart of experimental science

Advances in AI are feeding into a large number of scientific sectors.

A prime example is the work carried out in collaboration with CERN (European Organization for Nuclear Research) as part of high-energy irradiation experiments. These facilities test the resistance of materials and electronic components to radiation, a key issue for space, nuclear, and particle physics.

CRI researchers are using neural networks capable of learning a compact and relevant representation of complex data.

In concrete terms, these models make it possible to:

  • automatically analyze thousands of technical documents (using automatic natural language processing)
  • monitor the quality of particle beams in real time
  • detect anomalies invisible to the human eye

This work has a direct impact: more reliable experiments that are quicker to analyze and better documented, integrated into CERN’s operational tools.

  1. Automatically choosing the right AI models

Another major challenge in modern AI is choosing the right model. With a multitude of algorithms and parameters to choose from, this step is often costly, empirical, and reserved for experts.

At CRI, researchers are developing meta-learning approaches, i.e., methods capable of automatically recommending:

  • the most suitable type of model for a given dataset,
  • the most relevant settings (hyperparameters).

Notable results include:

  • large-scale comparisons of ensemble methods (combinations of several models), which identify strategies that are both effective and robust;
  • innovative work showing that LLMs themselves can serve as advisors, capable of proposing effective models based on a summary description of a dataset, without a costly research phase.

The result is considerable time savings for researchers and engineers, and more accessible AI, even for non-specialists.

  1. Accelerating models… without modifying them

Large AI models, particularly LLMs, rely on a key mechanism: attention. This is what allows the model to determine which information in a text is relevant for producing a response. But this mechanism is extremely costly when texts become very long.

CRI researchers exploit a key property of these models: attention matrices are in practice very sparse, meaning that the majority of their values contribute little or nothing to the final result. By exploiting this property, they have developed sparse calculation methods for attention, capable of:

  • reducing the calculations required for texts with more than 250,000 tokens by up to 98%
  • without any measurable loss of quality in the responses produced by the model

Other work focuses on:

  • optimizing neural networks for embedded systems with very limited memory resources
  • automatically reorganizing calculations via the compilation chain, drawing on CRI’s long-standing expertise in compilation and optimization

This research has a concrete impact, with faster models capable of processing massive documents and deployable on a wider variety of infrastructures.

Measuring, predicting, and optimizing energy for more energy-efficient AI

One of the most concrete results concerns the reduction of energy consumption in computing infrastructures. By carefully analyzing the functioning of target parallel architectures, CRI researchers have shown that energy efficiency can be significantly improved through appropriate scheduling of calculations at all levels, from instructions and operations to tasks executed on accelerators.

By developing:

  • tools for accurately measuring consumption
  • energy/performance predictive models
  • and then a new intelligent scheduler

they have managed to reduce the overall energy consumption of a computing cluster by more than 10% without slowing down scientific production. A decisive step towards more sustainable AI!

The CAMELIA PEPR, for sovereign AI

This research is being strategically extended through the Priority Research Program and Equipment (PEPR) on AI components, co-led by the https://www.cnrs.fr/frCEA and Inria, which aims to accelerate the development of artificial intelligence in France. Funded by France 2030 and led by the Agence nationale de la recherche (ANR), the objective of the PEPR CAMELIA (AI Components) is to design a complete hardware and software environment enabling the efficient execution of AI applications on hardware targets, developed as part of the project, constituting a sovereign alternative to the solutions currently available, most of which are foreign.

The CRI plays a key role in this PEPR by co-piloting the lot dedicated to the design and development of the software components essential for exploiting the targeted architectures and guaranteeing performance, portability, and energy efficiency.

This is a scientific, industrial, and technological sovereignty issue.

A workshop to bring together an AI community

This work was highlighted during the AI Workshop held in December 2025 at Mines Paris – PSL. Designed as an opportunity for internal exchange, the event allowed faculty, doctoral students, and engineers to present their projects, tools, and platforms through oral presentations and posters.

Beyond the diversity of topics, the workshop highlighted a common dynamic: building AI that is grounded in reality, capable of interacting with humans and integrating into complex systems.

AI at the CRI

Research that makes sense

At the center, AI is not only more powerful, it is also the subject of in-depth understanding, methodical optimization, and thoughtful integration into major scientific, industrial, and societal challenges. Through this research, the Computer Science Research Center (CRI) at Mines Paris – PSL affirms a clear vision: AI that calculates “less, but better,” and whose impact extends beyond the laboratory to contribute to both research advances and concrete, high-performance, and more frugal applications.

 


To go further:

  • Youssef Attia El Hili, Albert Thomas, Malik Tiomoko, Abdelhakim Benechehab, Corentin Leger, et al. In-Context Meta-Learning with Large Language Models for Automated Model and Hyperparameter Selection. Evaluating the Evolving LLM Lifecycle Workshop in Neural Information Processing Systems (NeurIPS 2025), Dec 2025, San Diego (CA), United States. ⟨hal-05462938⟩
  • Gianpietro Consolaro, Zhen Zhang, Harenome Razanajato, Nelson Lossing, Nassim Tchoulak, Adilla Susungi, Artur Cesar Araujo Alves, Renwei Zhang, Denis Barthou, Corinne Ancourt, and Cedric Bastoul. PolyTOPS: Reconfigurable and Flexible Polyhedral Scheduler. In 2024 IEEE/ACM International Symposium on Code Generation and Optimization (CGO), pages 28 – 40, Edinburgh, France, March 2024.
  • Roblex Nana Tchakoute and Claude Tadonki. EAS-Sim: A Framework and its Methodology for the Co-Design of Multi-Objective, Energy-Aware Schedulers for AI Clusters. In SC ’25 Workshops of the International Conference for High Performance Computing, Networking, Storage and Analysis, pages 2041–2050, St Louis MO USA, United States, November 2025
  • Quentin R Petit, Chong Li, Nahid Emad, Jack Dongarra. Efficient embedding initialization via dominant eigenvector projections. SC Workshops ’25 – Workshops of the International Conference for High Performance Computing, Networking, Storage and Analysis, Nov 2025, Saint Louis, United States. pp.1790-1799, ⟨10.1145/3731599.3767541⟩. ⟨hal-05369672⟩
  • Jaroslaw Szumega, Lamine Bougueroua, Blerina Gkotse, Pierre Jouvelot, Nicola Minafra, et al.. Characterization of an IRRAD beam profile monitor at the CERN T8 beamline and possible improvements via cross-analysis with multiwire proportional chamber. 16th International Particle Accelerator Conference, Jun 2025, Taiwan, Taiwan. pp.2921-2923, ⟨10.18429/JACoW-IPAC2025-THPM111⟩. ⟨hal-05334451⟩
  • Arthur Viens, Corinne Ancourt, and Jean-Louis Dufour. Depth-First Fusion and Tiling for CNN Memory Footprint Reduction in TVM. In AccML: Accelerated Machine Learning Workshop of HIPEAC 2026 conference. Krakow. January 2026.

Also to be discovered