SSPDF - Js Viewer
SSPDF - Js Viewer
Hao Zhou, Chengming Hu, Dun Yuan, Ye Yuan, Di Wu, Xi Chen, and Xue Liu are with McGill University, Canada;
Digital Object Identifier: 10.1109/MWC.001.2400384 Hina Tabassum is with York University, Canada.
Pre-training
1) Training LLMs from High Thermal Design A network-specic LLM
enables LLMs with Pre-training is extremely
scratch on hundreds of Power GPUs such NVIDIA indicates great benets,
comprehension, resource-intensive,
billions of tokens for the H100-80G or A100-80G but training network LLMs
Pre-training reasoning, and requiring long training
next token prediction. are required, e.g., training from scratch may be
LLMs instruction following times, and substantial
2) Pre-training is a Llama3.1-405B took inappropriate due to the
capabilities, which energy consumption
foundational step of LLM approximately 30.84M computational resource
can address various and GPUs or TPUs.
development and usage. GPU hours. requirements.
downstream tasks.
1) RAG is a more
1) RAG involves extra 1) The cost depends on
RAG combines LLMs practical approach for
integration complexity, the retrieval method and
with an external RAG can improve implementing LLMs,
as it requires ecient the size of the knowledge
Retrieval knowledge base, LLM’s response especially when existing
retrieval algorithms. base. Inference may take
augmented enabling the model accuracy by leveraging network-specic datasets
2) RAG highly depends seconds to minutes per
generation to retrieve relevant external knowledge are available.
on a well-structured query.
(RAG) information during sources, generating 2) It ensures the LLM has
knowledge base, 2) Requires a balance
inference and provide more up-to-date results. access to the most relevant
requiring dedicated between retrieval speed
more accurate responses. and up-to-date network
creation. and accuracy.
data.
Note that the objective of this table is to introduce the features of each technique, and they serve dierent purposes in LLM development and application.
These approaches can be combined to achieve better performance, e.g., ne-tuning network LLMs and then prompting.
TABLE 1. Summary of various LLM usage approaches for wireless networks.
On the other hand, prompt engineering also text learning, serving as critical references for LLMs
requires considerable expertise in crafting the to learn from. Therefore, they should be careful-
prompts, which will directly affect the quality of ly selected, formatted, and designed, which may
generated content. Although prompt engineer- require professional understanding and knowl-
ing can signicantly improve LLM’s performance, edge. In-context learning is a promising approach
it relies on the model’s inherent capabilities, for to applying LLMs to wireless networks, which can
example, achieved by pre-training or ne-tuning. take advantage of previous network solutions as
Therefore, given the opportunities and challenges, demonstrations to address unseen network issues.
it is crucial to explore the integration of prompt- Meanwhile, wireless networks are complicated
ing techniques and wireless networks, showing a systems, and one task may include several logic and
promising way to utilize generative AI. reasoning steps. For instance, many network opti-
mization problems include several elements, e.g.,
proMpt engIneerIng For wIreLess networks base stations, transmission power, bandwidth, and
This section presents several prompting tech- network users. Using simple input-output pairs as
niques and their network applications, that is, demonstrations cannot provide the logical reasons
in-context learning, chain-of-thought, prompt- behind these examples, and LLMs may have dicul-
based planning, and self-renement methods. ty in learning. To this end, chain-of-thought is pro-
Firstly, as shown in Fig. 1, in-context learning posed as an advanced strategy to enhance LLM’s
is a fundamental and crucial capability of LLMs. It performance in complex reasoning tasks. Specifi-
refers to the process of learning from formatted cally, chain-of-thought will explain each logic step in
natural language instructions and demonstrations, the prompt demonstrations: “Let’s think step by step.
and then improving the performance on target Given the current base station transmission power
tasks. Given contextual demonstration prompts, {BS_power}, using the Shannon capacity equation,
LLMs can address various downstream tasks by the transmission data rate is {user_rate}. Then, we
using existing knowledge from pre-trained data. compare the achieved rate with the target data rate,
Demonstrations are of great importance for in-con- and {user_rate} is lower. Therefore, we may need
to increase the transmission power.” This example tion proposes a novel iterative prompting scheme,
includes two steps, “using Shannon capacity equa- in which we prompt LLMs iteratively to improve
tion” and “compare the achieved rate with the target the performance on target network tasks.
data rate,” and then the nal decision is to “increase
the transmission power.” With these two-step expla- nAturAL LAnguAge-bAsed tAsk descrIptIon
nations, LLMs can easily capture the relationship Firstly, as shown in Fig. 2, the target task is
between transmission power and target data rate, described by formatted natural language, includ-
and generate better replies using the logical chain. ing task goals, definitions, and extra rules, pro-
In-context learning and chain-of-thought viding fundamental task information to the LLM.
are mainly designed for single tasks, which may For instance, the task goal describes the problem
have difficulty tackling more complicated tasks, as a “decision-making task,” and it introduces the
for example, collecting information from multiple set of control variables as “candidate decisions.”
sources and producing multi-step reasoning to Then, it highlights the importance of “environment
generate a comprehensive answer. Many complex states,” indicating that the LLM has to consider
network tasks need to be broken down into more specic environment variables for decision-mak-
manageable sub-tasks. For example, the network ing. After that, it introduces the current target task
conguration usually involves multiple interrelated by “here is a new case to solve, and the current
network elements, and network project develop- environment state is….” Moreover, the nal rules
ment may consist of a series of coding and testing specify that “select from {Decision_set} based on
tasks. In this case, prompt-based planning aims to above examples.”
generate a group of sub-tasks, which will be exe- With these formatted task descriptions, the
cuted one by one. This planning capability is cru- LLM can capture the key elements of the deci-
cial for handling many large-scale network tasks. sion-making task, focusing on hidden patterns
In addition, LLMs may generate incorrect between environment states and corresponding
answers at the initial attempts, and external feed- decision variables. Such a design can also lower
back is essential to improve its replies. As shown the understanding difficulties caused by profes-
in Fig. 1, self-refinement allows LLMs to self-im- sional network knowledge, since general domain
prove their previous outputs by generating feed- LLMs such as GPTs and Llama are not specically
back and refinement prompts. For instance, the pre-trained for network applications.
LLM’s initial transmission power decision may be
incorrect, and we can use the same LLM to eval- IterAtIve proMptIng scheMe
uate and provide feedback, and then feed the As illustrated in Fig. 2, the above task description
feedback and renement to the LLM, improving will become input prompts for the LLM at network
the initial power control decisions. In particular, edge cloud, and the LLM selects a specific Deci-
it indicates that LLMs can automatically improve sion_n from the set of candidate decisions {Deci-
their previous outputs without human interven- sion_set}. Then, the network management decisions
tion. Self-renement can signicantly save human will be sent to the wireless network environment for
eort in improving LLM’s initial replies, especially implementation such as resource allocation, user
considering the complexity of network tasks. association and transmission power level decisions.
Finally, Table 2 compares the above techniques The network operation results be collected as a
in terms of main features, advantages, potential new example, including environment states, select-
issues, and wireless network applications. Note ed decisions, and rewards that are related to optimi-
that these methods can be combined to address zation objectives and network metrics.
complex tasks comprehensively. For instance, After that, the examples from the network out-
using self-renement in prompt-based planning to put, that is, previous explorations, are collected in
save eort in the evaluation phase, and integrating an experience pool, recording all previous exam-
chain-of-thought to improve the multi-step logic ples. The examples in this pool will be further evalu-
reasoning in sub-task decomposition. ated to choose specic examples well-suited for the
target task. For instance, a straightforward strategy
IterAtIve proMptIng For is to select previous examples with similar environ-
ment states as current states, for example, similar
wIreLess network optIMIzAtIon user numbers and channel conditions. Then the pre-
Optimization techniques are of great importance vious decisions and rewards become useful refer-
for wireless network management, and this sec- ences for LLM decision-making. However, note that
Prompt-based planning aims LLM-based planning requires Wireless networks are complicated
LLMs can make complex tasks
to address more complicated dedicated analyses to decompose systems and LLM-enabled
Prompt- more manageable by using
scenarios such as multi-hop a complicated task into multiple automated task decomposition
based their planning capabilities,
question answering. It breaks sub-tasks. Zero-shot and can be extremely useful for long-
planning lowering the overall
complex tasks down into automated task decomposition is term network management and
diculties.
manageable sub-tasks. still the main diculty. project development.
• Overall performance: The same LLM evaluates prompting method designs task-specific demon-
the overall prediction performance, for exam- stration, feedback, and renement prompts, allow-
ple, Mean Absolution Error (MAE) between ing the LLM to iteratively enhance its predictions
ground truth and predictions. by adhering to the instructions in these prompts,
• Periodical performance: Considering that without requiring additional model ne-tuning or
time-series data can be represented through training. Compared to the dedicated training of
sine and cosine functions in the real domain long short-term memory (LSTM) networks, the
[12], the same LLM is tasked with projecting proposed self-rened prompting method is more
both ground truth and predicted traffic onto computationally ecient with the inference abili-
these functions, respectively. By matching the ties, particularly advantageous for on-device learn-
projection of ground truth, predictions are ing on resource-constrained devices in real-world
expected to capture the periodic nature of wireless networks. Given that iteratively append-
ground truth accurately, such as fluctuations ing all historical predictions can increase the
during peak and o-peak trac periods. length of demonstration prompts and potentially
• Format and completeness: The predictions lead to higher resource consumption, a future
should match the format of ground truth as direction for the self-refined prompting method
well as be complete for each timestamp, for will focus on effectively compressing extensive
example, 24 values for daily trac prediction. demonstration prompts, thereby further enhanc-
• Prediction method: The prediction method is ing resource eciency.
summarized in the feedback demonstration
prompt, which calls for adopting more advanced cAse studIes
and accurate methods to enhance performance
further. To avoid additional computational costs, sIMuLAtIon settIngs
while the LLM may suggest developing a neural This section presents two case studies on network
network as a potential improvement method, optimization and prediction problems.
this recommendation is disregarded. Network Optimization: This case study
Given the above instructions, the refinement involves a base station power control problem a
demonstration prompts aim to provide spe- fundamental optimization task in wireless network
cific actionable steps associated with feedback elds. The considered problem formulations aim
demonstration prompts. This allows the same to minimize the base station power consumption
LLM to self-rene previous predictions by adher- and meanwhile maintain an average data rate
ing to the detailed feedback outlined in feedback threshold for users.1 We consider three adjacent
demonstration prompts. Note that the process of base stations and the associated user numbers
feedback generation and prediction refinement dynamically change from 5 to 15 each, the aver-
is iteratively conducted until the prediction per- age data rate threshold is 1.5 Mb/s/per user, and
formance converges, whereas the inference is the channel gain applies 3GPP urban network
completed without engaging in the feedback gen- models. This case study includes:
eration and prediction renement. Additionally, to • Iterative prompting: We consider 2 LLM mod-
prevent the repetition of previous incorrect pre- els: Llama3-7b-instruct as a small-scale model,
dictions, the history of predicted trac, feedback, and Llama3-70b-instruct as a large-scale model.
and refinement demonstration prompts are also We deploy the proposed iterative prompting
included as inputs to the LLM at each iteration. technique as introduced above, and in Fig. 2,
In summary, LLMs excel at recognizing and in which the LLM will explore the environment,
performing new tasks through prompt engineer- accumulate experience, and learn iteratively.
ing techniques that utilize contextual information, • DRL baseline: DRL is included as a baseline,
including task descriptions and demonstration which has been widely used to address network
examples. Specically, the proposed self-rened optimization problems after dedicated model
c) d)
FIGURE 4. Performance comparison in network optimization and traffic prediction case studies: a) Optimization Case
Study: improved power consumption of Llamas; b) Optimization Case Study: service quality comparison with training;c)
Prediction Case Study: average MAE and MSE comparison. e self-rene method achieves comparable performance as
LSTM as a few-shot predictor, while LSTM has been delicately trained on the datase; d) Prediction Case Study: predicted
traffic comparison in 24 hours.
MAE reductions of 29.72 percent and 17.09 per- LLMs also show higher learning efficiency than
cent, which underscores more powerful generaliza- existing techniques, and they allow human lan-
tion capability on unseen datasets through iterative guage-based input, showing the possibility of nat-
feedback generation and prediction renement pro- ural language-based future network management.
cesses. GPT-3.5 shows a worse performance due to
its outdated architecture and designs, which have reAL-worLd chALLenges And LIMItAtIons
been observed in many existing studies. Although This section will further discuss the limitations
LSTM serves as an optimal baseline with superior and real-world challenges of prompt engineering
performance, it is important to note that LSTM is applications.
specifically trained on the target dataset, incurring Firstly, prompt engineering is inherently limited
additional computational costs, which presents signif- by the context window size of LLMs, which restricts
icant challenges in scenarios involving resource-con- the capabilities to process tasks requiring extensive
strained devices within wireless networks. Given the contextual information beyond the model’s input
dynamic evolution of traffic distribution in non-sta- and memory constraints. Recent advancements
tionary wireless networks, a well-trained LSTM may such as Llama3 with a context window size of
overt the historical trac used during the pre-train- 128k tokens, signicantly mitigate this limitation.
ing, resulting in degraded performance on new, Secondly, the eectiveness of prompt engineer-
unseen trac over time. Furthermore, similar results ing relies heavily on the selection of demonstrations
can be observed in Fig. 4d, which illustrates the one- shown to the model or the order of demonstrations.
day ground truth trac and the predicted trac of Poorly chosen or suboptimal examples can result
various methods. The self-rened prediction method in biased outputs. Our work considers iterative
more closely matches the ground truth trac, indi- prompting to select better demonstrations, but pro-
cating strong scalability of predictions across various fessional experiences are still needed in this process.
base stations. By contrast, GPT-4 shows an obvious Moreover, crafting eective prompts for com-
mismatch from 9:00 to 15:00. plicated tasks in network environments usually
In summary, the simulation results in Fig. 4 requires trial-and-error experimentation. This iter-
demonstrate that LLMs have great potential in ative process is inecient, especially when results
handling network management, that is, optimiza- are inconsistent or dicult to reproduce. Without
tion and prediction tasks. They present satisfied clear guidelines, engineers must rely on intuition
performance as conventional machine learning or extensive testing to achieve desired outcomes.
algorithms such as DRL and LSTM, while avoiding This can slow down the development for novel
extra model training and fine-tuning complexity. and complex use cases as in 6G.
networks, a well-trained concLusIon learning (RL) and large language models (LLMs) into the domain of
telecommunication. He also collaborates with the Crypto-Metaverse-
The progress of generative AI and LLMs brings Blockchain-Cloud (CMBC) research group, exploring innovative proj-
LSTM may overt the promising opportunities for next-generation wire- ects in Web 3.0, decentralized storage, and distributed computing
historical traffic used less networks. This work provides a comprehen- systems. He hold a Bachelor’s degree from Xi’an Jiaotong University
and a Master’s degree from Imperial College London.
during the pre-training, sive overview of prompt engineering for network
resulting in degraded applications. It presents in-depth analyses of dif- Ye Yuan is a Ph.D. student in Computer Science at Mila – Que-
performance on new, ferent prompting techniques such as in-context bec AI Institute and McGill University. His research focuses on
learning, chain-of-thought, and self-renement. In score-based generative models with applications on addressing
unseen traffic over addition, this work proposes two novel prompt-
out-of-distribution challenges in oline black-box optimization, as
well as developing foundational knowledge models using diffu-
time. ing schemes for network optimization and pre- sion-based and autoregressive language models. Ye has published
diction problems, and the case studies show that papers in leading venues such as NeurIPS, EMNLP, IEEE COMST,
the proposed methods can achieve satisfactory IEEE WCL. He is a recipient of the BMO Responsible AI Senior
Scholar award and holds a Bachelor of Science in Honours Com-
performance. In the future, we will explore more puter Science from McGill University.
complicated and advanced network tasks such as
recongurable intelligence surfaces and integrat- Di Wu is a research scientist at Canada AI Inc and an Adjunct Profes-
sor at McGill University. He was a senior sta research scientist, a team
ed sensing and communication, uncovering the leader at Samsung AI Center Montreal. He did postdoctoral research
full potential of LLMs for future wireless networks. at Montreal MILA and Stanford University. He received his Ph.D. and
M.Sc. from McGill University in 2018 and Peking University in 2013,
reFerences respectively. Di also holds Bachelor’s degrees in microelectronics and
economics. His research interests mainly lie in reinforcement learning,
[1] H. Zhou et al., “Large Language Model (LLM) for Telecommuni-
cations: A Comprehensive Survey on Principles, Key Techniques, transfer learning, meta-Learning, and multitask Learning. He is also
and Opportunities,” arXiv preprint arXiv:2405.10825, 2024. interested in leveraging such algorithms to improve real-world systems.
[2] C. Liang et al., “Generative AI-Driven Semantic Communica-
tion Networks: Architecture, Technologies and Applications,” Xi CHen is an adjunct professor at School of Computer Science,
arXiv preprint arXiv:2401.00124, 2023. McGill University, and also a lead AI Researcher at Proactive AI
[3] S. Javaid et al., “Leveraging Large Language Models for Integrated Lab, working on agentic AI. His experience and passion lie in a
Satellite-Aerial-Terrestrial Networks: Recent Advances and Future wide range of AI domains, including LLM agents, alignment, rea-
Directions,” arXiv preprint arXiv:2407.04581, 2024. soning, decision making foundation models, 5G/6G foundation
[4] P. Sahoo et al., “A Systematic Survey of Prompt Engineering models, AI for communications, Integrated Sensing and Commu-
in Large Language Models: Techniques and Applications,” nications (ISAC), autonomous driving, smart IoT, smart homes,
arXiv:2402.07927, 2024. smart systems, vehicle-to-everything, and so on. He achieved his
[5] H. Du et al., “Generative Al-Aided Joint Training-Free Secure PhD degree at School of Computer Science, McGill University. He
Semantic Communications via Multi-modal Prompts,” Proc. received both M.Eng. and B.S. degrees from Department of Elec-
ICASSP 2024-2024 IEEE Int’l. Conf. Acoustics, Speech and Sig- tronic Engineering, Shanghai Jiao Tong University.
nal Processing, 2024, pp. 12,896–900.
[6] W. X. Zhao et al., “A Survey of Large Language Models,” Hina Tabassum received the Ph.D. degree from the King Abdullah
arXiv:2303.18223, 2023. University of Science and Technology (KAUST). She is currently an
[7] Z. Lin et al., “Pushing Large Language Models to the 6g Associate Professor with the Lassonde School of Engineering, York
Edge: Vision, Challenges, and Opportunities,” arXiv preprint University, Canada, where she joined as an Assistant Professor, in
arXiv:2309.16739, 2023. 2018. She is appointed as a Visiting Faculty at University of Toronto
[8] M. Xu et al., “When Large Language Model Agents Meet 6g in 2024 and the York Research Chair of 5G/6G-enabled mobility
Networks: Perception, Grounding, and Alignment,” IEEE Wire- and sensing applications in 2023, for ve years. She is appointed
less Commun., 2024. as IEEE Comsoc Distinguished Lecturer for the term 2025-2026.
[9] Y. Liu et al., “Optimizing Mobile-Edge Ai-Generated Everything She is listed in the Stanford’s list of the World’s Top Two-Percent
(Aigx) Services by Prompt Engineering: Fundamental, Frame- Researchers in 2021-2024. She received the Lassonde Innovation
work, and Case Study,” IEEE Network, 2023. Early-Career Researcher Award in 2023 and the N2Women: Rising
[10] H. Zhou et al., “Large Language Model (LLM)-Enabled In-Con- Stars in Computer Networking and Communications in 2022.
text Learning for Wireless Network Optimization: A Case Study
of Power Control,” arXiv preprint arXiv:2408.00214, 2024. Xue Liu received PhD degree in computer science from the Uni-
[11] C. Hu et al., “Self-Rened Generative Foundation Models for Wire- versity of Illinois at Urbana–Champaign, USA. He is a William
less Trac Prediction,” arXiv preprint arXiv:2408.10390, 2024. Dawson Scholar (chair professor) and a full professor with the
[12] P. Esling and C. Agon, “Time-Series Data Mining,” ACM Com- School of Computer Science, McGill University, Montreal, Cana-
puting Surveys, vol. 45, no. 1, 2012, pp. 1–34. da. His research interests include computer and communication
[13] M. Chiang et al., “Power Control in Wireless Cellular Net- networks, real-time and embedded systems, cyber-physical sys-
works,” Foundations and Trends® in Networking, vol. 2, no. 4, tems and the IoT, green computing, and smart energy technolo-
2008, pp. 381–533. gies. He has published more than 200 research papers in major
[14] G. Barlacchi et al., “A Multi-Source Dataset of Urban Life peer-reviewed international journals and conference proceedings
in the City of Milan and the Province of Trentino,” Scientific in these areas and received several best paper awards.