Study LLM4Gov

Call for solution type:
SpeedUP
participation dates:
from 15-10-2024 to 19-11-2024
StatuS:
Open for applications
Deadline for QUESTIONS:
06-11-2024
CONTACT:
datascience@digital.etat.lu
Descriptif FR (Pdf, 129 Kb) Description EN (Pdf, 123 Kb) Project specifications LLM4Gov (Pdf, 250 Kb)

Name

Study LLM4Gov

Ordering party

Ministry for Digitalisation

Types

Study of feasibility, benchmark of the existing, proposition of a concept

Objectives

The Ministry for Digitalisation has received of a multitude of requests on the provision of a ‘Large Language Model’ (LLM) dedicated to the public sector’s needs. So, the main objective behind the integration of such a model is to automate a variety of tasks. These tasks predominantly include the generation of text-based content and summaries. The model is also expected to be capable of generating coherent responses to questions regarding specific business needs. As a first step prior to a potential implementation of one or more LLM’s, a study to identify the best solution is to be done with this call for solution. 

Challenge details

A study to identify the model(s) best suited to the needs of the Luxembourgish government is the first step prior to a potential implementation of one or more LLM’s. To identify the needs of public sector organisations regarding the use of Large language models, a survey has been sent out to gather their input. The results of the survey will be shared with the selected economic operator. Use cases such as analysing documents, an alternative research tool and content creation will be in the focus.

In addition to the survey results, the chosen economic partner will also be informed about several use cases that have already been identified for an LLM model within the government. This project aims at producing a study for the Ministry for Digitalisation ranking existing LLMs on relevant criteria, along with a plan to implement the identified use cases based on selected LLMs.

 

The following list of criteria are a non-exhaustive list of criteria that must at a minimum be considered in the study, and therefore also included in the final report:

Models’ selection:
  • The leading state-of-the-art LLMs should at minima be considered (both proprietary and Open-Source).
  • LLMs with different numbers of parameters should be considered (lightweight to flagship).
  • LLMs using different technologies must be considered (for instance transformers, decoder-only, encoder only, mixture of experts)
Capabilities and features:
  • General capabilities: the overall capabilities of candidate LLMs should be provided.
  • Use-case specific capabilities: the ability for each candidate LLM to provide a relevant answer to each use case should be evaluated.

Versatility:

  • The ability for each LLM to cover multiple use case with high-quality solutions should be outlined.
  • The ability for each LLM to onboard new use cases easily.

More information on possible use cases is going to be provided at a later stage (survey results).

Language Support:

  • It should be evaluated whether the model has to support English (EN), French (FR), German (DE) and Luxembourgish (LU).
  • LLMs should also be capable of understanding and generating code in various coding languages.

Benchmarking:

  • The capabilities and versatility of different LLMs should be supported by relevant benchmarks, as described above.
  • LLMs should be benchmarked on various use cases, including risk of hallucination, speed, and accuracy.
  • Characteristics of the benchmarking environment should be given, along with all relevant metrics to help ensuring that the obtained results are reproducible.
Integration

Integration with existing Models:

  • The feasibility of integrating existing state models (like speech to texts models, an employment specific chatbot etc.) should be analysed. More information on those models will be provided to the chosen economic partner.
  • API Availability: The model should provide an accessible and reliable API for integration with existing tools.

Hosting and Infrastructure:

  • The pros and cons of cloud vs on-premises hosting, licensed vs non-licensed solutions, and open-source vs proprietary solutions should be evaluated.
  • The ability to deal with sensitive data (personal or others) in a sovereign and secured environment should be evaluated.
  • The ability for the LLM to be deployed on an OpenShift container platform should be evaluated.
  • The ability for the LLM to be deployed on air-gapped platforms such as sovereign disconnected cloud platforms should be evaluated.

The Integration in a complete environment should be evaluated. This includes:

  • The ease of integration with tools such as, but not limited to:

- vector DBs,

- caches,

- monitoring, logging, and tracing tools,

- orchestration tools,

- agents.

  • The model's REST API should support OAuth2.0 and OIDC.
  • The possibility to implement the LLMs on different environments (Build and run environment).
  • The existence of a fully-fledged suite integrating the LLM.

Maintenance and Integration:

Considerations should include versioning, model updates, ensuring reproducibility of generated answers, maintenance efforts, hardware requirements (CPU, GPU, memory, storage), and energy consumption.

  • This should be done in general and specifically for the provided use cases.
Cost
  • Cost Analysis: A comprehensive cost analysis should be conducted, including the estimate cost of hosting, maintenance, potential license fees or human resources. In order to compare the LLMs the costs should be indicated in TCO (total cost of ownership) for a period of 5 years.

Other criteria might arise from answers given to a survey that will be shared at project launch.

Compliance
  • Data Protection:

- The model should be GDPR compliant.

- The exposure of the LLM to extraction attacks should be provided.

 

For complete information regarding the requirements, the deliverables, and the type of offer to be submitted, please refer to the project specifications linked on the website.