Semantic Message Extraction for Text Based Data With Deep Neural Nets
- The project aims to semantically encode text and extract meaning to aid in transmission as proposed in [1] and [2] and is built with pyTorch.
- Traditional communication systems have focussed on improving the transmission rates through physical channels. They are constrained by the Shannon Limit which defines the maximum data rate that can be achieved over a channel. While this theoretical limit has not been reached, the current systems are close and 6g communication research proposes alternate transmission methods.
- Semantic communication is one such model which extracts the semantic message embedding from the original message, thus only necessiating the transfer of essential information and discarding the trivial. It follows logically then, that this enables the transmission of more information. Current research in the field employs Deep Learning architectures to achieve this extraction of meaning. While this is relevant to different modalities of data, the focus of this project is text.
- Multiple architectures have been proposed in literature, from LSTMs to Transformers[1] and AutoEncoders[2]. The aim of the system is to obtain c = f(m) where "c" is the semantic message extracted from the original "m" and "f" is a deep learning model that facilitates this conversion. BLEU scores (sentence BLEU) are used as the performance metric.
Europarl: A Parallel Corpus for Statistical Machine Translation offers sentence-aligned text for machine translation. Single-language data is sufficient for our purposes, since our aim here is to merely recreate the original message on the reciever's end. To this end, the English section of the "bg-en" dataset was used.
- Data Pre-Processing & Dataset creation
- Data: EuroParl corpus(bg-en)
- Samples: Training - 233093, Testing - 25899, Batch size 128
- Model Training
- Indexing, Embedding
- Semantic Encoding, Semantic Decoding (Transformers), greedy decoding
- Seq2Txt - Tensor to Text from Vocab
- Loss - CE Loss - Cross Entropy
- Performance metric - BLEU score
- Sentence BLEU between encoder-decoder output and original text with 1 gram, 2 gram, 3 gram, 4 gram and an average with equal weights(*0.25) scores were calculated.
- Training Epochs - 12
- Testing Epochs - 2
- Test cases - 25899
[1] Huiqiang Xie et.al., ”Deep Learning Enabled Semantic Communication Systems,”IEEE Transactions on signal processing, vol 69, 2021.
[2] Xinlai Luo, “Autoencoder-based Semantic Communication Systems with Relay Channels,” arXiv:2111.10083v1 [cs.IT] 19 Nov 2021