Abstract
This letter proposes a semantic importance-aware communication (SIAC) scheme using pre-trained language models (e.g., ChatGPT, BERT, etc.). Specifically, we propose a cross-layer design with a pre-trained language model embedded in/connected by the cross-layer manager. The pre-trained language model is utilized to quantify the semantic importance of data frames. Based on the quantified semantic importance, we investigate semantic importance-aware power allocation. Unlike existing deep joint source-channel coding (Deep-JSCC)-based semantic communication schemes, SIAC can be directly embedded into current communication systems by only introducing a cross-layer manager. Our experimental results show that the proposed SIAC scheme can achieve lower semantic loss than existing equal-priority communications.
| Original language | English |
|---|---|
| Pages (from-to) | 2328-2332 |
| Number of pages | 5 |
| Journal | IEEE Communications Letters |
| Volume | 27 |
| Issue number | 9 |
| DOIs | |
| Publication status | Published - Sept 1 2023 |
Keywords
- Semantic communications
- data importance
- power allocation
- pre-trained language model
ASJC Scopus subject areas
- Modelling and Simulation
- Computer Science Applications
- Electrical and Electronic Engineering
Fingerprint
Dive into the research topics of 'Semantic Importance-Aware Communications Using Pre-Trained Language Models'. Together they form a unique fingerprint.Cite this
- APA
- Standard
- Harvard
- Vancouver
- Author
- BIBTEX
- RIS