작성중...
- 한영 기계 번역(Korean-English Machine Translation) 모델 개발 스터디
PyTorch
,koNLPY
,NLPY
,Gensim
package 활용
- Python 3.6 (may work with other versions, but I used 3.6)
- PyTorch 1.1.0
- Gensim 3.8.0
- konlpy 0.5.1
- nltk 3.4.4
git clone https://github.com/dlcjfgmlnasa/NMT-Koean-To-English.git --recursive
pip install -r requirement.txt
- Sequence to Sequence (Seq2Seq)
- Sequence to Sequence with Attention (Seq2Seq with Attention)
- Convolution Sequence to Sequence
- ByteNet
- SliceNet
- Transformer
- Parameter List
import argparse
parser = argparse.ArgumentParser()
parser.add_argument('--rnn_sequence_size', default=30, type=int)
parser.add_argument('--min_count', default=3, type=int)
parser.add_argument('--max_count', default=10000, type=int)
parser.add_argument('--embedding_size', default=200, type=int)
parser.add_argument('--rnn_dim', default=200, type=int)
parser.add_argument('--rnn_layer', default=3, type=int)
parser.add_argument('--batch_size', default=128, type=int)
- loss
- translation
- Parameter List
import argparse
parser = argparse.ArgumentParser()
parser.add_argument('--rnn_sequence_size', default=30, type=int)
parser.add_argument('--min_count', default=3, type=int)
parser.add_argument('--max_count', default=100000, type=int)
parser.add_argument('--embedding_size', default=200, type=int)
parser.add_argument('--rnn_dim', default=123, type=int)
parser.add_argument('--rnn_layer', default=3, type=int)
parser.add_argument('--rnn_dropout_rate', default=0.5, type=float)
parser.add_argument('--use_residual', default=True, type=bool)
parser.add_argument('--attention_method', default='general', choices=['dot', 'general', 'concat'], type=str)
parser.add_argument('--batch_size', default=128, type=int)
- 구현 완료 (2019.08.05)