The repository contains the HUE dataset and Hanja language models, and the code baselines to implement them. See our paper for more details about HUE and Hanja PLMs.
HUE is composed of 4 tasks:
- Chronological Attribution (CA)
- Topic Classification (TC)
- Named Entity Recognition (NER)
- Summary Retrieval (SR)
HUE aims to encourage training Hanja language models that help to analyze the Korean historical documents written in Hanja which is an extinct language.
You can download all 12 here, or individually from the table below:
Model Name | Size |
---|---|
AnchiBERT + AJD/DRS | 379.2 MB |
mBERT + AJD/DRS | 379.2 MB |
Make sure you have installed the packages listed in environment.yml. If you use conda, you can create an environment from this package with the following command:
conda env create -f environment.yml
Codes, data, and models should be placed in this directory tree.
HUE
├── code
│ ├── HUE_fine-tuning_Chronological_Attribution.ipynb
│ ├── HUE_fine-tuning_Named_Entity_Recognition.ipynb
│ ├── HUE_fine-tuning_Summary_Retrieval.ipynb
│ └── HUE_fine-tuning_Topic_Classification.ipynb
├── dataset
│ ├── HUE_Chronological_Attribution
│ │ ├── HUE_Chronological_Attribution.csv
│ │ ├── HUE_Chronological_Attribution_dev.csv
│ │ ├── HUE_Chronological_Attribution_test.csv
│ │ └── HUE_Chronological_Attribution_train.csv
│ ├── HUE_Named_Entity_Recognition
│ │ ├── HUE_Named_Entity_Recognition_dev.csv
│ │ ├── HUE_Named_Entity_Recognition_test.csv
│ │ └── HUE_Named_Entity_Recognition_train.csv
│ ├── HUE_Summary_Retrieval
│ │ ├── HUE_Summary_Retrieval_dev.csv
│ │ ├── HUE_Summary_Retrieval_test.csv
│ │ └── HUE_Summary_Retrieval_train.csv
│ └── HUE_Topic_Classification
│ ├── HUE_Topic_Classification.csv
│ ├── HUE_Topic_Classification_dev.csv
│ ├── HUE_Topic_Classification_test.csv
│ └── HUE_Topic_Classification_train.csv
├── model
│ ├── AnchiBERT+AJD-DRS
│ │ ├── config.json
│ │ ├── pytorch_model.bin
│ │ ├── special_tokens_map.json
│ │ ├── tokenizer_config.json
│ │ └── vocab.txt
│ └── mBERT+AJD-DRS
│ ├── config.json
│ ├── pytorch_model.bin
│ ├── special_tokens_map.json
│ ├── tokenizer_config.json
│ └── vocab.txt
└── tokenizer
├── AnchiBERT+AJD-DRS
│ ├── special_tokens_map.json
│ ├── tokenizer_config.json
│ └── vocab.txt
└── mBERT+AJD-DRS
├── special_tokens_map.json
├── tokenizer_config.json
└── vocab.txt