Congratulations to the winners of QuantMinds-CompatibL Hackathon 2023!
The online application at hackathon.compatibl.com has been updated with the leaderboard and award-winning prompts.
- Click on a leaderboard line or select a competition entry to run the code.
- Source code for the coding entries has been merged to the main branch of this repo.
- 1st place - Emanuele Cagliero and Chiara Borsani
- 2nd place - Shih-Hau Tan
- 3rd place - Joao Fonseca De Araujo
- 1st place - Shih-Hau Tan
- 2nd place - Giuseppe Mascolo and Miren Gutierrez Rodriguez
- 3rd place - Marko Petrov
- 1st place - Peter Hufton, Paolo Tripoli and Maksymilian Wojtas
- 1st place - Peter Hufton, Paolo Tripoli and Maksymilian Wojtas
- Shih-Hau Tan
- Samples: data/TermSheets-Hackathon.csv
- Default prompt (includes parameters for scoring): prompts/TermSheets.txt
- Samples: data/PricingModels-Hackathon.csv
- Default prompt (includes parameters for scoring): prompts/PricingModels.txt
For technical support before, during and after the hackathon, email support@compatibl-hackathon.freshdesk.com
- Both individual and team participation is permitted (one entry per individual)
- Each individual or team can participate in one or more of the four categories
- Participants can provide an alias to be shown in scoreboard for individual and team entries. Real names will not be disclosed without permission.
Most importantly, participation in the hackathon without entering the competition is welcome and can provide valuable experience with practical applications of AI in a collaborative environment.
- With models hosted online, a basic laptop is all you need.
- For non-coding entries, you can use an online playground provided by the organizers or any other playground (however see the note below about third-party playgrounds).
- For coding entries, you can clone this GitHub repo and work using Python 3.9 or later, or use your own tools.
- The cost of model hosting service use during the hackathon is expected to be less than GBP 50.
Use the links below for getting API keys for online model hosting services:
- OpenAI: https://platform.openai.com/api-keys
- Fireworks.ai https://readme.fireworks.ai/docs/quickstart
- Replicate.com https://replicate.com/account/api-tokens
- The open-source software package that will be used for scoring can be downloaded from GitHub at this URL .
- The same package will be available online during the hackathon at the URL that will be posted at the start of the competition.
- The same package can be used to develop and test your entry for both coding and non-coding submissions. Its use is not required and is a matter of convenience. Any other playground or Python package can be used instead to submit the entry.
- Ensure your computer has Python 3.9 or later
- Checkout from GitHub: https://github.com/compatibl/hackathon-quantminds-2023
- Create and activate venv or conda environment
- Run
pip install -r requirements.txt
- Run
python -m hackathon
- Click on the URL in text output to open the UI
This year the competition will be held in two streams, with each stream further subdivided by model type (GPT or LLAMA) for the total of four categories.
- Pricing Model Source Code Comprehension Stream (Source-GPT and Source-LLAMA categories)
- Trade Confirmation Comprehension Stream (Text-GPT and Text-LLAMA categories)
- At the beginning of the hackathon, examples for each category will be randomly divided into two equal parts through a public drawing.
- Participants will receive the first half of the examples to develop their competition entries.
- When the hackathon concludes, the second half of the examples will be utilized to score the entries.
- Non-coding and coding entries will compete alongside one another in each category.
- Coding participants are welcome to use and modify pre- and post-processing code for the non-coding entries included in this GitHub repo.
- Using advanced methods including chains is permitted for coding entries.
- Using your own tools instead of cloning this repo is permitted as long as both the entry itself and the tools are licensed under Apache or MIT license. Contact the organizers for information on other licenses.
Participants can choose between non-code and code submission formats in each of the four categories, to be scored together in each category.
- The competition entry is a text prompt submitted by email guides the model to produce the desired output in JSON format.
- The prompt can be developed using any chat application or playground.
- An official playground application will be provided by the organizers for use with participants' own keys.
- The entry is scored by parsing JSON in the model output.
- Any text before the opening or after the closing curly brackets is ignored.
- Participants can submit their own Python code via GitHub.
- The code should read a .csv file with input data and produce a .csv file with output fields.
- A sample open-source environment will be provided by the organizers in this package, however its use is optional and the entry can use any other open-source toolset in Python (MIT or Apache license).
- The competition entry itself must be provided under an open source license (MIT or Apache) as keeping the entries confidential is not practical in competition format.
- For each sample, the model output must provide category (multiple choice) and parameters.
- Incorrectly identifying the category among the available choices leads to zero score for the sample.
- For each sample where category is identified correctly, score is the number of correctly identified parameter values including for those parameters whose correct value is empty.
- For code submissions, the output must comply with the .csv format provided. Any additional text will be disregarded.
- For non-code submissions, the data will be extracted from JSON in the model output. Any text outside JSON will be disregarded. Any additional fields in JSON will be disregarded. If more than one JSON is provided in the output, only the first one will be used.
- When using GPT models, GPT-4-Turbo (model identifier gpt-4-1106-preview), GPT-4 and GPT-3.5-Turbo models are permitted. Due to its higher rate limits, GPT-4-Turbo model is preferable to GPT-4.
- When using LLAMA models, any model from the LLAMA 2 family is permitted, including LLAMA-2-70B and CodeLlama-34B. The list of available models varies by provider.
- Model settings such as seed, temperature, top-p, top-k and others may be specified with the entry.
- Among the two LLAMA model hosting providers, Fireworks offers higher performance than Replicate and is recommended.
- Using ChatGPT to prepare a submission is not recommended. Reason: ChatGPT and OpenAI API have significant differences, which will likely cause reduced performance when the entry is tuned using ChatGPT and then scored using OpenAI API.
- Similar differences may exist between some of the LLAMA-based third party playgrounds and the LLAMA 2 API.
- When using a third-party playground, ensure that no hidden pre- or post-processing is present and the model parameters are known and can be specified with your competition entry.
- Both on-code and code submissions will be scored using model hosting service specified by the participant.
- For non-coding entries, the options are OpenAI (GPT categories only), Replicate, or Fireworks.ai.
- For coding entries, the same options are available or any other where the organizers can obtain a key online without a lengthy submission process.
- Coding entries can also use local models. The organizers will provide GPU hardware to score them.
- The cost of scoring using hosted models or GPU hardware will be borne by the organizers.