/LLY-GLLM

LLY-GLLM is the precursor to the interpretative Large Language Model of the LILY Project and is designed to understand relationships between words. It is based on parameterized quantum gates.

Primary LanguagePythonMIT LicenseMIT

GLLM-removebg-preview

Introduction: LLY-GLLM

LLY-GLLM, which stands for Generative Large Language Model, is derived from the LILY Project and serves as a precursor to a Quantum Large Language Model. It is based on the LLY-GP model and primarily uses quantum circuits to represent words, combining them to form connections. For example, Circuit(King) + Circuit(Woman) = State(Queen). Similar to LLY-GP, these circuits consist of a series of L-gates that are filled with tokens.

LLY-GLLM is available on the LILY QML platform, making it accessible for researchers and developers.

For inquiries or further information, please contact: info@lilyqml.de.

Contributors

Role Name Links
Project Lead Leon Kaiser ORCID, GitHub
Inquiries and Management Raul Nieli Email
Supporting Contributors Eileen Kühn GitHub, KIT Profile
Supporting Contributors Max Kühn GitHub

Functionality

The LLY-GLLM model is based on the property of representing a multi-qubit system using a Q-Sphere. A Q-Sphere is a sphere that contains all possible states of a multi-qubit system on its surface, making the principle easier to illustrate.

By pre-configuring words for each state on this sphere, such as "King" and "Woman," these states are represented by a series of quantum circuits that use the tokens as input phases.

When these circuits are placed in sequence, the tuning phases can be optimized to indicate a specific state. As demonstrated in this example, the vector of "King" plus the vector of "Woman" would result in the vector "Queen." Technically, this is the circuit with the input token "King" followed by the circuit with the input token "Woman," which together represent the state associated with "Queen."

thumbnail_QM-105__1_-removebg-preview (2) thumbnail_QM-106-removebg-preview

Workflow

A word is essentially represented as a token, which is then placed into the input phase of the gate. The tuning phases in this case are pre-trained sets. When we have multiple circuits representing tokens arranged in sequence, we refer to these as layers. Each layer has its own tuning phases, which have been previously trained. In practical applications, the pre-trained tuning phases are utilized.

Outlook

Another point is the prospect of training an additional model, LLY-LLM. This is intended to be the large language model that emerges from the LILY project.