MAGICS-LAB/DNABERT_2
[ICLR 2024] DNABERT-2: Efficient Foundation Model and Benchmark for Multi-Species Genome
ShellApache-2.0
Issues
- 1
- 2
Unable to reproduce covid results
#103 opened by anihab - 0
Results obtained in the original paper
#127 opened by Zehui127 - 0
Shapley plot
#126 opened by Anshullllllll - 1
finetune results seems not very stable
#125 opened by maris205 - 2
When will the code for pre-training model and training BPE tokenizer be available?
#74 opened by a-green-hand-jack - 4
Fine-tune for continuous labels
#79 opened by buwanim - 3
Is there a way to turn off the setting to use flash attention/triton library?
#104 opened by vivektreddy - 1
finetune error
#123 opened by maris205 - 2
- 4
The shape of embedding issue
#121 opened by WhiteNan - 3
Specific processes for pre-training
#118 opened by yuyadanyadan - 0
readme for GUE datasets?
#120 opened by maris205 - 0
Logits and Labels are different shapes
#119 opened by SamuelTWu - 2
- 2
Language Modelling Head
#116 opened by tdsone - 1
DNA Representation
#117 opened by ChangShaole - 2
Suggest better error message than 'assert q.is_cuda and k.is_cuda and v.is_cuda'
#112 opened by richelbilderbeek - 1
- 0
How to use fine tuned model for prediction?
#111 opened by Anshullllllll - 0
Where are your Species Classification datasets?
#110 opened by andreyfsch - 3
- 0
- 1
Whether huggingface released model has been further pretrained on GUE benchmark
#106 opened by Mamingqian - 3
GUE+ datasets?
#87 opened by leannmlindsey - 1
- 3
- 1
Data distribution in pretraining dataset
#101 opened by leannmlindsey - 1
About the pretrain data
#100 opened by wyhsleep - 1
Instability in reproducing GUE dataset result
#102 opened by Mamingqian - 2
- 3
EPI datasets, not getting published results
#99 opened by leannmlindsey - 3
About the motif prediction function
#94 opened by josecar24 - 1
Attention error
#96 opened by HelloWorldLTY - 1
Random issues still come up in use
#95 opened by josecar24 - 7
- 8
CUDA out of memory
#90 opened by chenruipu - 0
GUE labels
#93 opened by nehmea - 1
- 2
Getting embedding of a sequence
#89 opened by CorvusVaine - 1
- 1
Discuss a question about k-mer
#73 opened by xueleecs - 2
Pretraining, Pretraining, Pretraining!!!
#76 opened by multydoffer - 1
Special token treatment.
#81 opened by prwoolley - 1
Cannot Reproduce DNA-BERT2‘s Result
#86 opened by KatarinaYuan - 1
- 1
How do I output the attention from the model?
#80 opened by jkb-ag - 3
Unable to Retrieve ' hidden_states ' Despite ' Setting return_dict=True ' and ' output_hidden_states=True '
#85 opened by biglittleme - 0
splice site predictions
#83 opened by amitpande74 - 1
hidden_states = model(inputs)[0] # [1, sequence_length, 768]-- Is the second dimension really the sequence length?
#72 opened by xueleecs