Skip to content

[Paper][WWW2023] Structure Pre-training and Prompt Tuning for Knowledge Graph Transfer

Notifications You must be signed in to change notification settings

zjukg/KGTransformer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

5080111 · May 26, 2023

History

24 Commits
Jan 26, 2023
Mar 27, 2023
Jan 26, 2023
Jan 26, 2023
Jan 26, 2023
Jan 26, 2023
May 26, 2023
Jan 26, 2023
Apr 14, 2023
Feb 1, 2023
Jan 26, 2023
Mar 27, 2023
Jan 26, 2023
Jan 26, 2023
Feb 1, 2023
Jan 26, 2023
Mar 27, 2023
Mar 27, 2023
Mar 27, 2023
Mar 27, 2023
Mar 27, 2023
Mar 27, 2023
Jan 26, 2023

Repository files navigation

KGTransformer

Code and datasets for paper "Structure Pre-training and Prompt Tuning for Knowledge Graph Transfer" accepted by WWW'23.

Model Architecture

Subgraph Pre-training

Task prompt tuning

Requirements

Run the following commands to create an environment (assuming CUDA10.1):

(pytorch:1.8PAI-gpu-py36-cu101-ubuntu18.04)

pip install torch==1.8.0+cu101 -f https://download.pytorch.org/whl/torch_stable.html
pip install transformers==2.0.0
pip install tokenizers==0.11.1
pip install scipy==1.5.3
pip install scikit-learn==0.24.2
pip install packaging==21.3
pip install huggingface_hub==0.4.0

How to Run

Pretrain

python run_pretrain.py --pretrain_dataset BIG --dataset_name BIG --num_hidden_layers 4 --train_bs 16 --lr 1e-4 --epochs 10 

Downstream Task: Triple Classification

python run_down_triplecls.py --dataset_name WN18RR --pretrain_dataset BIG --down_task down_triplecls --train_bs 16 --test_bs 128 --epochs 50 --fixedT 1

Downstream Task: Zero-shot Image Classification

python run_down_zsl.py --dataset_name down_zsl --pretrain_dataset down_zsl --down_task down_zsl --train_bs 32 --test_bs 8 --epochs 10 --fixedT 1 --lr 1e-4 --test_epoch 1 --multi_pic 15

Downstream Task: Question Answering

pip install transformers==2.0.0
python run_down_qa.py --dataset_name down_qa --pretrain_dataset BIG --down_task down_qa --token_types 5 --train_bs 16 --big_bs 64 --train_split 1 --test_bs 64 --epochs 20 --encoder_lr 2e-5 --decoder_lr 1e-4 --fixedT 1 

About

[Paper][WWW2023] Structure Pre-training and Prompt Tuning for Knowledge Graph Transfer

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages