├── README.md ├── best_model_weight ├── ABGCN_14semeval_laptop.pth └── ABGCN_14semeval_rest.pth ├── dataset ├── 14semeval_laptop │ ├── test.txt │ └── train.txt ├── 14semeval_rest │ ├── test.txt │ └── train.txt └── Twitter │ ├── test.txt │ └── train.txt ├── dataset_npy ├── dataset_14semeval_laptop.npz ├── dataset_14semeval_rest.npz ├── dataset_Twitter.npz ├── vocab_14semeval_laptop.npy ├── vocab_14semeval_rest.npy └── vocab_Twitter.npy ├── embeddings ├── 14semeval_laptop_840B.pkl ├── 14semeval_rest_840B.pkl └── Twitter_840B.pkl ├── layers ├── __init__.py ├── __pycache__ │ ├── __init__.cpython-36.pyc │ ├── __init__.cpython-37.pyc │ ├── attention.cpython-36.pyc │ ├── attention.cpython-37.pyc │ ├── dynamic_rnn.cpython-36.pyc │ ├── dynamic_rnn.cpython-37.pyc │ ├── squeeze_embedding.cpython-36.pyc │ └── squeeze_embedding.cpython-37.pyc ├── attention.py ├── dynamic_rnn.py ├── module │ ├── attention │ │ ├── __pycache__ │ │ │ ├── attention.cpython-36.pyc │ │ │ ├── attention.cpython-37.pyc │ │ │ ├── bilinear_attention.cpython-36.pyc │ │ │ ├── concat_attention.cpython-36.pyc │ │ │ ├── dot_attention.cpython-36.pyc │ │ │ ├── dot_attention.cpython-37.pyc │ │ │ ├── mlp_attention.cpython-36.pyc │ │ │ ├── scaled_dot_attention.cpython-36.pyc │ │ │ ├── tanh_bilinear_attention.cpython-36.pyc │ │ │ └── tanh_concat_attention.cpython-36.pyc │ │ ├── attention.py │ │ ├── bilinear_attention.py │ │ ├── concat_attention.py │ │ ├── dot_attention.py │ │ ├── mlp_attention.py │ │ ├── multi_head_attention.py │ │ ├── no_query_attention.py │ │ ├── scaled_dot_attention.py │ │ ├── tanh_bilinear_attention.py │ │ └── tanh_concat_attention.py │ └── utils │ │ ├── __pycache__ │ │ ├── constants.cpython-36.pyc │ │ ├── constants.cpython-37.pyc │ │ ├── loss.cpython-36.pyc │ │ ├── sentence_clip.cpython-36.pyc │ │ └── squash.cpython-36.pyc │ │ ├── constants.py │ │ ├── loss.py │ │ ├── sentence_clip.py │ │ └── squash.py ├── point_wise_feed_forward.py └── squeeze_embedding.py ├── main.py ├── model ├── ABGCN.py ├── ATAE_LSTM.py ├── GCAE.py └── __pycache__ │ ├── ABGCN.cpython-37.pyc │ ├── ATAE_LSTM.cpython-37.pyc │ └── GCAE.cpython-37.pyc ├── mydataset.py ├── process_data.py ├── stages_saved_model └── first_stage │ └── ABGCN_Amazon_review_pre.pth └── utils.py /README.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/README.md -------------------------------------------------------------------------------- /best_model_weight/ABGCN_14semeval_laptop.pth: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/best_model_weight/ABGCN_14semeval_laptop.pth -------------------------------------------------------------------------------- /best_model_weight/ABGCN_14semeval_rest.pth: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/best_model_weight/ABGCN_14semeval_rest.pth -------------------------------------------------------------------------------- /dataset/14semeval_laptop/test.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/dataset/14semeval_laptop/test.txt -------------------------------------------------------------------------------- /dataset/14semeval_laptop/train.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/dataset/14semeval_laptop/train.txt -------------------------------------------------------------------------------- /dataset/14semeval_rest/test.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/dataset/14semeval_rest/test.txt -------------------------------------------------------------------------------- /dataset/14semeval_rest/train.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/dataset/14semeval_rest/train.txt -------------------------------------------------------------------------------- /dataset/Twitter/test.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/dataset/Twitter/test.txt -------------------------------------------------------------------------------- /dataset/Twitter/train.txt: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/dataset/Twitter/train.txt -------------------------------------------------------------------------------- /dataset_npy/dataset_14semeval_laptop.npz: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/dataset_npy/dataset_14semeval_laptop.npz -------------------------------------------------------------------------------- /dataset_npy/dataset_14semeval_rest.npz: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/dataset_npy/dataset_14semeval_rest.npz -------------------------------------------------------------------------------- /dataset_npy/dataset_Twitter.npz: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/dataset_npy/dataset_Twitter.npz -------------------------------------------------------------------------------- /dataset_npy/vocab_14semeval_laptop.npy: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/dataset_npy/vocab_14semeval_laptop.npy -------------------------------------------------------------------------------- /dataset_npy/vocab_14semeval_rest.npy: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/dataset_npy/vocab_14semeval_rest.npy -------------------------------------------------------------------------------- /dataset_npy/vocab_Twitter.npy: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/dataset_npy/vocab_Twitter.npy -------------------------------------------------------------------------------- /embeddings/14semeval_laptop_840B.pkl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/embeddings/14semeval_laptop_840B.pkl -------------------------------------------------------------------------------- /embeddings/14semeval_rest_840B.pkl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/embeddings/14semeval_rest_840B.pkl -------------------------------------------------------------------------------- /embeddings/Twitter_840B.pkl: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/embeddings/Twitter_840B.pkl -------------------------------------------------------------------------------- /layers/__init__.py: -------------------------------------------------------------------------------- 1 | # -*- coding: utf-8 -*- 2 | -------------------------------------------------------------------------------- /layers/__pycache__/__init__.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/__pycache__/__init__.cpython-36.pyc -------------------------------------------------------------------------------- /layers/__pycache__/__init__.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/__pycache__/__init__.cpython-37.pyc -------------------------------------------------------------------------------- /layers/__pycache__/attention.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/__pycache__/attention.cpython-36.pyc -------------------------------------------------------------------------------- /layers/__pycache__/attention.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/__pycache__/attention.cpython-37.pyc -------------------------------------------------------------------------------- /layers/__pycache__/dynamic_rnn.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/__pycache__/dynamic_rnn.cpython-36.pyc -------------------------------------------------------------------------------- /layers/__pycache__/dynamic_rnn.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/__pycache__/dynamic_rnn.cpython-37.pyc -------------------------------------------------------------------------------- /layers/__pycache__/squeeze_embedding.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/__pycache__/squeeze_embedding.cpython-36.pyc -------------------------------------------------------------------------------- /layers/__pycache__/squeeze_embedding.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/__pycache__/squeeze_embedding.cpython-37.pyc -------------------------------------------------------------------------------- /layers/attention.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/attention.py -------------------------------------------------------------------------------- /layers/dynamic_rnn.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/dynamic_rnn.py -------------------------------------------------------------------------------- /layers/module/attention/__pycache__/attention.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/__pycache__/attention.cpython-36.pyc -------------------------------------------------------------------------------- /layers/module/attention/__pycache__/attention.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/__pycache__/attention.cpython-37.pyc -------------------------------------------------------------------------------- /layers/module/attention/__pycache__/bilinear_attention.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/__pycache__/bilinear_attention.cpython-36.pyc -------------------------------------------------------------------------------- /layers/module/attention/__pycache__/concat_attention.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/__pycache__/concat_attention.cpython-36.pyc -------------------------------------------------------------------------------- /layers/module/attention/__pycache__/dot_attention.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/__pycache__/dot_attention.cpython-36.pyc -------------------------------------------------------------------------------- /layers/module/attention/__pycache__/dot_attention.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/__pycache__/dot_attention.cpython-37.pyc -------------------------------------------------------------------------------- /layers/module/attention/__pycache__/mlp_attention.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/__pycache__/mlp_attention.cpython-36.pyc -------------------------------------------------------------------------------- /layers/module/attention/__pycache__/scaled_dot_attention.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/__pycache__/scaled_dot_attention.cpython-36.pyc -------------------------------------------------------------------------------- /layers/module/attention/__pycache__/tanh_bilinear_attention.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/__pycache__/tanh_bilinear_attention.cpython-36.pyc -------------------------------------------------------------------------------- /layers/module/attention/__pycache__/tanh_concat_attention.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/__pycache__/tanh_concat_attention.cpython-36.pyc -------------------------------------------------------------------------------- /layers/module/attention/attention.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/attention.py -------------------------------------------------------------------------------- /layers/module/attention/bilinear_attention.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/bilinear_attention.py -------------------------------------------------------------------------------- /layers/module/attention/concat_attention.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/concat_attention.py -------------------------------------------------------------------------------- /layers/module/attention/dot_attention.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/dot_attention.py -------------------------------------------------------------------------------- /layers/module/attention/mlp_attention.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/mlp_attention.py -------------------------------------------------------------------------------- /layers/module/attention/multi_head_attention.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/multi_head_attention.py -------------------------------------------------------------------------------- /layers/module/attention/no_query_attention.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/no_query_attention.py -------------------------------------------------------------------------------- /layers/module/attention/scaled_dot_attention.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/scaled_dot_attention.py -------------------------------------------------------------------------------- /layers/module/attention/tanh_bilinear_attention.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/tanh_bilinear_attention.py -------------------------------------------------------------------------------- /layers/module/attention/tanh_concat_attention.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/attention/tanh_concat_attention.py -------------------------------------------------------------------------------- /layers/module/utils/__pycache__/constants.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/utils/__pycache__/constants.cpython-36.pyc -------------------------------------------------------------------------------- /layers/module/utils/__pycache__/constants.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/utils/__pycache__/constants.cpython-37.pyc -------------------------------------------------------------------------------- /layers/module/utils/__pycache__/loss.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/utils/__pycache__/loss.cpython-36.pyc -------------------------------------------------------------------------------- /layers/module/utils/__pycache__/sentence_clip.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/utils/__pycache__/sentence_clip.cpython-36.pyc -------------------------------------------------------------------------------- /layers/module/utils/__pycache__/squash.cpython-36.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/utils/__pycache__/squash.cpython-36.pyc -------------------------------------------------------------------------------- /layers/module/utils/constants.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/utils/constants.py -------------------------------------------------------------------------------- /layers/module/utils/loss.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/utils/loss.py -------------------------------------------------------------------------------- /layers/module/utils/sentence_clip.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/utils/sentence_clip.py -------------------------------------------------------------------------------- /layers/module/utils/squash.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/module/utils/squash.py -------------------------------------------------------------------------------- /layers/point_wise_feed_forward.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/point_wise_feed_forward.py -------------------------------------------------------------------------------- /layers/squeeze_embedding.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/layers/squeeze_embedding.py -------------------------------------------------------------------------------- /main.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/main.py -------------------------------------------------------------------------------- /model/ABGCN.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/model/ABGCN.py -------------------------------------------------------------------------------- /model/ATAE_LSTM.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/model/ATAE_LSTM.py -------------------------------------------------------------------------------- /model/GCAE.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/model/GCAE.py -------------------------------------------------------------------------------- /model/__pycache__/ABGCN.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/model/__pycache__/ABGCN.cpython-37.pyc -------------------------------------------------------------------------------- /model/__pycache__/ATAE_LSTM.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/model/__pycache__/ATAE_LSTM.cpython-37.pyc -------------------------------------------------------------------------------- /model/__pycache__/GCAE.cpython-37.pyc: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/model/__pycache__/GCAE.cpython-37.pyc -------------------------------------------------------------------------------- /mydataset.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/mydataset.py -------------------------------------------------------------------------------- /process_data.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/process_data.py -------------------------------------------------------------------------------- /stages_saved_model/first_stage/ABGCN_Amazon_review_pre.pth: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/stages_saved_model/first_stage/ABGCN_Amazon_review_pre.pth -------------------------------------------------------------------------------- /utils.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/WHU-ZQH/UIKA/HEAD/utils.py --------------------------------------------------------------------------------