Skip to content
Snippets Groups Projects
Commit 6fe5e31c authored by Jacques Fize's avatar Jacques Fize
Browse files

Change Readme (add Bert)

parent 10945496
No related branches found
No related tags found
No related merge requests found
...@@ -154,4 +154,6 @@ subset* ...@@ -154,4 +154,6 @@ subset*
time* time*
/data* /data*
\ No newline at end of file
output_bert_allcooc_adjsampling3radius20km_batch32_epoch10
\ No newline at end of file
...@@ -93,3 +93,14 @@ grid.run() ...@@ -93,3 +93,14 @@ grid.run()
| -e,--epochs | number of epochs | | -e,--epochs | number of epochs |
| -d,--dimension | size of the ngram embeddings | | -d,--dimension | size of the ngram embeddings |
| --admin_code_1 | (Optional) If you wish to train the network on a specific region | | --admin_code_1 | (Optional) If you wish to train the network on a specific region |
# New model based on BERT embeddings
In the recent years, BERT architecture proposed by Google researches enables to outperform state-of-art methods for differents tasks in NLP (POS, NER, Classification). To verify if BERT embeddings would permit to increase the performance of our approach, we code a script to use bert with our data. In our previous model, the model returned two values each on between [0,1]. Using Bert, the task has shifted to classification (softmax) where each class correspond to a cell on the glob. We use the hierarchical projection model : Healpix. Other projections model like S2geometry can be considered : https://s2geometry.io/about/overview.
In order, to run this model training, run the `bert.py` script :
python3 bert.py <train_dataset> <test_dataset>
The train and test dataset are table data composed of two columns: sentence and label.
\ No newline at end of file
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment