Introduction
Documentation for the 3musketeers in the "Specializing Large Language Models for Telecom Networks by ITU AI/ML in 5G" challenge
Last updated
Documentation for the 3musketeers in the "Specializing Large Language Models for Telecom Networks by ITU AI/ML in 5G" challenge
Last updated
Authors:
Tewodros Idris ()
Alex Gichamba ()
Brian Ebiyau ()
This documentation describes our solutions for both tracks in the "Specializing Large Language Models for Telecom Networks by ITU AI/ML in 5G" challenge.
A brief summary of each solution:
We develop a RAG pipeline based on ColBERT V2 ( and ) and provide as many retrieved chunks as the context window can support.
We also append the full forms of abbreviations that appear in the question and options.
We finetune the LLM with context to align the LLM's generated to our desired output format, and to improve the "usable" context window.
We adopt the same RAG pipeline and abbreviations expansion as for Phi-2.
Following the challenge guidelines, we could not finetune the model. We find that it responds very poorly when options are provided in the prompt. So, we simply do not provide the options in the prompt.
We allow the LLM to freely generate a response conditioned on the question and context, but not the options. We then develop a scoring system that attempts to find the most likely option given the response.
If you'd like, you can quickly jump to reproducing results: REPRODUCING RESULTS