forked from manbo/internal-docs
Related Work Completed
This commit is contained in:
@@ -27,7 +27,6 @@
|
||||
|
||||
\ifuniqueAffiliation % 标准作者块
|
||||
\author{
|
||||
% 这里去掉了 \includegraphics{orcid.pdf} 以防止报错
|
||||
David S.~Hippocampus \\
|
||||
Department of Computer Science\\
|
||||
Cranberry-Lemon University\\
|
||||
@@ -77,7 +76,13 @@ Here introduces the background, problem statement, and contribution.
|
||||
% 2. Related Work
|
||||
\section{Related Work}
|
||||
\label{sec:related}
|
||||
In this section, we review the existing literature.
|
||||
Early generation of network data oriented towards ``realism'' mostly remained at the packet/flow header level, either through replay or statistical synthesis based on single-point observations. Swing, in a closed-loop, network-responsive manner, extracts user/application/network distributions from single-point observations to reproduce burstiness and correlation across multiple time scales \citep{10.1145/1151659.1159928,10.1145/1159913.1159928}. Subsequently, a series of works advanced header synthesis to learning-based generation: the WGAN-based method added explicit verification of protocol field consistency to NetFlow/IPFIX \citep{Ring_2019}, NetShare reconstructed header modeling as flow-level time series and improved fidelity and scalability through domain encoding and parallel fine-tuning \citep{10.1145/3544216.3544251}, and DoppelGANger preserved the long-range structure and downstream sorting consistency of networked time series by decoupling attributes from sequences \citep{Lin_2020}. However, in industrial control system (ICS) scenarios, the original PCAP is usually not shareable, and public testbeds (such as SWaT, WADI) mostly provide process/monitoring telemetry and protocol interactions for security assessment, but public datasets emphasize operational variables rather than packet-level traces \citep{7469060,10.1145/3055366.3055375}. This makes ``synthesis at the feature/telemetry level, aware of protocol and semantics'' more feasible and necessary in practice: we are more concerned with reproducing high-level distributions and multi-scale temporal patterns according to operational semantics and physical constraints without relying on the original packets. From this perspective, the generation paradigm naturally shifts from ``packet syntax reproduction'' to ``modeling of high-level spatio-temporal distributions and uncertainties'', requiring stable training, strong distribution fitting, and interpretable uncertainty characterization.
|
||||
|
||||
Diffusion models exhibit good fit along this path: DDPM achieves high-quality sampling and stable optimization through efficient $\epsilon$ parameterization and weighted variational objectives \citep{NEURIPS2020_4c5bcfec}, the SDE perspective unifies score-based and diffusion, providing likelihood evaluation and prediction-correction sampling strategies based on probability flow ODEs \citep{song2021scorebasedgenerativemodelingstochastic}. For time series, TimeGrad replaces the constrained output distribution with conditional denoising, capturing high-dimensional correlations at each step \citep{rasul2021autoregressivedenoisingdiffusionmodels}; CSDI explicitly performs conditional diffusion and uses two-dimensional attention to simultaneously leverage temporal and cross-feature dependencies, suitable for conditioning and filling in missing values \citep{tashiro2021csdiconditionalscorebaseddiffusion}; in a more general spatio-temporal structure, DiffSTG generalizes diffusion to spatio-temporal graphs, combining TCN/GCN with denoising U-Net to improve CRPS and inference efficiency in a non-autoregressive manner \citep{wen2024diffstgprobabilisticspatiotemporalgraph}, and PriSTI further enhances conditional features and geographical relationships, maintaining robustness under high missing rates and sensor failures \citep{liu2023pristiconditionaldiffusionframework}; in long sequences and continuous domains, DiffWave verifies that diffusion can also match the quality of strong vocoders under non-autoregressive fast synthesis \citep{kong2021diffwaveversatilediffusionmodel}; studies on cellular communication traffic show that diffusion can recover spatio-temporal patterns and provide uncertainty characterization at the urban scale \citep{11087622}. These results overall point to a conclusion: when the research focus is on ``telemetry/high-level features'' rather than raw messages, diffusion models provide stable and fine-grained distribution fitting and uncertainty quantification, which is exactly in line with the requirements of ICS telemetry synthesis. Meanwhile, directly entrusting all structures to a ``monolithic diffusion'' is not advisable: long-range temporal skeletons and fine-grained marginal distributions often have optimization tensions, requiring explicit decoupling in modeling.
|
||||
|
||||
Looking further into the mechanism complexity of ICS: its channel types are inherently mixed, containing both continuous process trajectories and discrete supervision/status variables, and discrete channels must be ``legal'' under operational constraints. The aforementioned progress in time series diffusion has mainly occurred in continuous spaces, but discrete diffusion has also developed systematic methods: D3PM improves sampling quality and likelihood through absorption/masking and structured transitions in discrete state spaces \citep{austin2023structureddenoisingdiffusionmodels}, subsequent masked diffusion provides stable reconstruction on categorical data in a more simplified form \citep{Lin_2020}, multinomial diffusion directly defines diffusion on a finite vocabulary through mechanisms such as argmax flows \citep{hoogeboom2021argmaxflowsmultinomialdiffusion}, and Diffusion-LM demonstrates an effective path for controllable text generation by imposing gradient constraints in continuous latent spaces \citep{li2022diffusionlmimprovescontrollabletext}. From the perspectives of protocols and finite-state machines, coverage-guided fuzz testing emphasizes the criticality of ``sequence legality and state coverage'' \citep{meng2025aflnetyearslatercoverageguided,godefroid2017learnfuzzmachinelearninginput,she2019neuzzefficientfuzzingneural}, echoing the concept of ``legality by construction'' in discrete diffusion: preferentially adopting absorption/masking diffusion on discrete channels, supplemented by type-aware conditioning and sampling constraints, to avoid semantic invalidity and marginal distortion caused by post hoc thresholding.
|
||||
|
||||
From the perspective of high-level synthesis, the temporal structure is equally indispensable: ICS control often involves delay effects, phased operating conditions, and cross-channel coupling, requiring models to be able to characterize low-frequency, long-range dependencies while also overlaying multi-modal fine-grained fluctuations on them. The Transformer series has provided sufficient evidence in long-sequence time series tasks: Transformer-XL breaks through the fixed-length context limitation through a reusable memory mechanism and significantly enhances long-range dependency expression \citep{dai2019transformerxlattentivelanguagemodels}; Informer uses ProbSparse attention and efficient decoding to balance span and efficiency in long-sequence prediction \citep{zhou2021informerefficienttransformerlong}; Autoformer robustly models long-term seasonality and trends through autocorrelation and decomposition mechanisms \citep{wu2022autoformerdecompositiontransformersautocorrelation}; FEDformer further improves long-period prediction performance in frequency domain enhancement and decomposition \citep{zhou2022fedformerfrequencyenhanceddecomposed}; PatchTST enhances the stability and generalization of long-sequence multivariate prediction through local patch-based representation and channel-independent modeling \citep{2023}. Combining our previous positioning of diffusion, this chain of evidence points to a natural division of labor: using attention-based sequence models to first extract stable low-frequency trends/conditions (long-range skeletons), and then allowing diffusion to focus on margins and details in the residual space; meanwhile, discrete masking/absorbing diffusion is applied to supervised/pattern variables to ensure vocabulary legality by construction. This design not only inherits the advantages of time series diffusion in distribution fitting and uncertainty characterization \citep{rasul2021autoregressivedenoisingdiffusionmodels,tashiro2021csdiconditionalscorebaseddiffusion,wen2024diffstgprobabilisticspatiotemporalgraph,liu2023pristiconditionaldiffusionframework,kong2021diffwaveversatilediffusionmodel,11087622}, but also stabilizes the macroscopic temporal support through the long-range attention of Transformer, enabling the formation of an operational integrated generation pipeline under the mixed types and multi-scale dynamics of ICS.
|
||||
|
||||
% 3. Methodology
|
||||
\section{Methodology}
|
||||
@@ -100,8 +105,6 @@ In this section, we present the future work.
|
||||
In this section, we summarize our contributions and future directions.
|
||||
|
||||
% 参考文献
|
||||
% 确保你的目录下有 references.bib 文件,并且里面有内容
|
||||
% 否则请先注释掉下面两行,以免编译报错或显示空白
|
||||
\bibliographystyle{unsrtnat}
|
||||
\bibliography{references}
|
||||
|
||||
|
||||
@@ -115,3 +115,307 @@
|
||||
doi={10.6028/NIST.SP.800-82r3},
|
||||
url={https://csrc.nist.gov/pubs/sp/800/82/r3/final}
|
||||
}
|
||||
|
||||
@article{10.1145/1151659.1159928,
|
||||
author = {Vishwanath, Kashi Venkatesh and Vahdat, Amin},
|
||||
title = {Realistic and responsive network traffic generation},
|
||||
year = {2006},
|
||||
issue_date = {October 2006},
|
||||
publisher = {Association for Computing Machinery},
|
||||
address = {New York, NY, USA},
|
||||
volume = {36},
|
||||
number = {4},
|
||||
issn = {0146-4833},
|
||||
url = {https://doi.org/10.1145/1151659.1159928},
|
||||
doi = {10.1145/1151659.1159928},
|
||||
abstract = {This paper presents Swing, a closed-loop, network-responsive traffic generator that accurately captures the packet interactions of a range of applications using a simple structural model. Starting from observed traffic at a single point in the network, Swing automatically extracts distributions for user, application, and network behavior. It then generates live traffic corresponding to the underlying models in a network emulation environment running commodity network protocol stacks. We find that the generated traces are statistically similar to the original traces. Further, to the best of our knowledge, we are the first to reproduce burstiness in traffic across a range of timescales using a model applicable to a variety of network settings. An initial sensitivity analysis reveals the importance of capturing and recreating user, application, and network characteristics to accurately reproduce such burstiness. Finally, we explore Swing's ability to vary user characteristics, application properties, and wide-area network conditions to project traffic characteristics into alternate scenarios.},
|
||||
journal = {SIGCOMM Comput. Commun. Rev.},
|
||||
month = aug,
|
||||
pages = {111–122},
|
||||
numpages = {12},
|
||||
keywords = {burstiness, energy plot, generator, internet, modeling, structural model, traffic, wavelets}
|
||||
}
|
||||
|
||||
@inproceedings{10.1145/1159913.1159928,
|
||||
author = {Vishwanath, Kashi Venkatesh and Vahdat, Amin},
|
||||
title = {Realistic and responsive network traffic generation},
|
||||
year = {2006},
|
||||
isbn = {1595933085},
|
||||
publisher = {Association for Computing Machinery},
|
||||
address = {New York, NY, USA},
|
||||
url = {https://doi.org/10.1145/1159913.1159928},
|
||||
doi = {10.1145/1159913.1159928},
|
||||
abstract = {This paper presents Swing, a closed-loop, network-responsive traffic generator that accurately captures the packet interactions of a range of applications using a simple structural model. Starting from observed traffic at a single point in the network, Swing automatically extracts distributions for user, application, and network behavior. It then generates live traffic corresponding to the underlying models in a network emulation environment running commodity network protocol stacks. We find that the generated traces are statistically similar to the original traces. Further, to the best of our knowledge, we are the first to reproduce burstiness in traffic across a range of timescales using a model applicable to a variety of network settings. An initial sensitivity analysis reveals the importance of capturing and recreating user, application, and network characteristics to accurately reproduce such burstiness. Finally, we explore Swing's ability to vary user characteristics, application properties, and wide-area network conditions to project traffic characteristics into alternate scenarios.},
|
||||
booktitle = {Proceedings of the 2006 Conference on Applications, Technologies, Architectures, and Protocols for Computer Communications},
|
||||
pages = {111–122},
|
||||
numpages = {12},
|
||||
keywords = {burstiness, energy plot, generator, internet, modeling, structural model, traffic, wavelets},
|
||||
location = {Pisa, Italy},
|
||||
series = {SIGCOMM '06}
|
||||
}
|
||||
|
||||
@article{Ring_2019,
|
||||
title={Flow-based network traffic generation using Generative Adversarial Networks},
|
||||
volume={82},
|
||||
ISSN={0167-4048},
|
||||
url={http://dx.doi.org/10.1016/j.cose.2018.12.012},
|
||||
DOI={10.1016/j.cose.2018.12.012},
|
||||
journal={Computers & Security},
|
||||
publisher={Elsevier BV},
|
||||
author={Ring, Markus and Schlör, Daniel and Landes, Dieter and Hotho, Andreas},
|
||||
year={2019},
|
||||
month=may, pages={156–172} }
|
||||
|
||||
@inproceedings{10.1145/3544216.3544251,
|
||||
author = {Yin, Yucheng and Lin, Zinan and Jin, Minhao and Fanti, Giulia and Sekar, Vyas},
|
||||
title = {Practical GAN-based synthetic IP header trace generation using NetShare},
|
||||
year = {2022},
|
||||
isbn = {9781450394208},
|
||||
publisher = {Association for Computing Machinery},
|
||||
address = {New York, NY, USA},
|
||||
url = {https://doi.org/10.1145/3544216.3544251},
|
||||
doi = {10.1145/3544216.3544251},
|
||||
abstract = {We explore the feasibility of using Generative Adversarial Networks (GANs) to automatically learn generative models to generate synthetic packet- and flow header traces for networking tasks (e.g., telemetry, anomaly detection, provisioning). We identify key fidelity, scalability, and privacy challenges and tradeoffs in existing GAN-based approaches. By synthesizing domain-specific insights with recent advances in machine learning and privacy, we identify design choices to tackle these challenges. Building on these insights, we develop an end-to-end framework, NetShare. We evaluate NetShare on six diverse packet header traces and find that: (1) across all distributional metrics and traces, it achieves 46\% more accuracy than baselines and (2) it meets users' requirements of downstream tasks in evaluating accuracy and rank ordering of candidate approaches.},
|
||||
booktitle = {Proceedings of the ACM SIGCOMM 2022 Conference},
|
||||
pages = {458–472},
|
||||
numpages = {15},
|
||||
keywords = {synthetic data generation, privacy, network packets, network flows, generative adversarial networks},
|
||||
location = {Amsterdam, Netherlands},
|
||||
series = {SIGCOMM '22}
|
||||
}
|
||||
|
||||
@inproceedings{Lin_2020, series={IMC ’20},
|
||||
title={Using GANs for Sharing Networked Time Series Data: Challenges, Initial Promise, and Open Questions},
|
||||
url={http://dx.doi.org/10.1145/3419394.3423643},
|
||||
DOI={10.1145/3419394.3423643},
|
||||
booktitle={Proceedings of the ACM Internet Measurement Conference},
|
||||
publisher={ACM},
|
||||
author={Lin, Zinan and Jain, Alankar and Wang, Chen and Fanti, Giulia and Sekar, Vyas},
|
||||
year={2020},
|
||||
month=oct, pages={464–483},
|
||||
collection={IMC ’20} }
|
||||
|
||||
@INPROCEEDINGS{7469060,
|
||||
author={Mathur, Aditya P. and Tippenhauer, Nils Ole},
|
||||
booktitle={2016 International Workshop on Cyber-physical Systems for Smart Water Networks (CySWater)},
|
||||
title={SWaT: a water treatment testbed for research and training on ICS security},
|
||||
year={2016},
|
||||
volume={},
|
||||
number={},
|
||||
pages={31-36},
|
||||
keywords={Sensors;Actuators;Feeds;Process control;Chemicals;Chemical sensors;Security;Cyber Physical Systems;Industrial Control Systems;Cyber Attacks;Cyber Defense;Water Testbed},
|
||||
doi={10.1109/CySWater.2016.7469060}}
|
||||
|
||||
@inproceedings{10.1145/3055366.3055375,
|
||||
author = {Ahmed, Chuadhry Mujeeb and Palleti, Venkata Reddy and Mathur, Aditya P.},
|
||||
title = {WADI: a water distribution testbed for research in the design of secure cyber physical systems},
|
||||
year = {2017},
|
||||
isbn = {9781450349758},
|
||||
publisher = {Association for Computing Machinery},
|
||||
address = {New York, NY, USA},
|
||||
url = {https://doi.org/10.1145/3055366.3055375},
|
||||
doi = {10.1145/3055366.3055375},
|
||||
abstract = {The architecture of a water distribution testbed (WADI), and on-going research in the design of secure water distribution system is presented. WADI consists of three stages controlled by Programmable Logic Controllers (PLCs) and two stages controlled via Remote Terminal Units (RTUs). Each PLC and RTU uses sensors to estimate the system state and the actuators to effect control. WADI is currently used to (a) conduct security analysis for water distribution networks, (b) experimentally assess detection mechanisms for potential cyber and physical attacks, and (c) understand how the impact of an attack on one CPS could cascade to other connected CPSs. The cascading effects of attacks can be studied in WADI through its connection to two other testbeds, namely for water treatment and power generation and distribution.},
|
||||
booktitle = {Proceedings of the 3rd International Workshop on Cyber-Physical Systems for Smart Water Networks},
|
||||
pages = {25–28},
|
||||
numpages = {4},
|
||||
keywords = {attack detection, cyber physical systems, cyber security, industrial control systems, water distribution testbed},
|
||||
location = {Pittsburgh, Pennsylvania},
|
||||
series = {CySWATER '17}
|
||||
}
|
||||
|
||||
@inproceedings{NEURIPS2020_4c5bcfec,
|
||||
author = {Ho, Jonathan and Jain, Ajay and Abbeel, Pieter},
|
||||
booktitle = {Advances in Neural Information Processing Systems},
|
||||
editor = {H. Larochelle and M. Ranzato and R. Hadsell and M.F. Balcan and H. Lin},
|
||||
pages = {6840--6851},
|
||||
publisher = {Curran Associates, Inc.},
|
||||
title = {Denoising Diffusion Probabilistic Models},
|
||||
url = {https://proceedings.neurips.cc/paper_files/paper/2020/file/4c5bcfec8584af0d967f1ab10179ca4b-Paper.pdf},
|
||||
volume = {33},
|
||||
year = {2020}
|
||||
}
|
||||
|
||||
@misc{song2021scorebasedgenerativemodelingstochastic,
|
||||
title={Score-Based Generative Modeling through Stochastic Differential Equations},
|
||||
author={Yang Song and Jascha Sohl-Dickstein and Diederik P. Kingma and Abhishek Kumar and Stefano Ermon and Ben Poole},
|
||||
year={2021},
|
||||
eprint={2011.13456},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.LG},
|
||||
url={https://arxiv.org/abs/2011.13456},
|
||||
}
|
||||
|
||||
@misc{rasul2021autoregressivedenoisingdiffusionmodels,
|
||||
title={Autoregressive Denoising Diffusion Models for Multivariate Probabilistic Time Series Forecasting},
|
||||
author={Kashif Rasul and Calvin Seward and Ingmar Schuster and Roland Vollgraf},
|
||||
year={2021},
|
||||
eprint={2101.12072},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.LG},
|
||||
url={https://arxiv.org/abs/2101.12072},
|
||||
}
|
||||
|
||||
@misc{tashiro2021csdiconditionalscorebaseddiffusion,
|
||||
title={CSDI Conditional Score-based Diffusion Models for Probabilistic Time Series Imputation},
|
||||
author={Yusuke Tashiro and Jiaming Song and Yang Song and Stefano Ermon},
|
||||
year={2021},
|
||||
eprint={2107.03502},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.LG},
|
||||
url={httpsarxiv.orgabs2107.03502},
|
||||
}
|
||||
|
||||
@misc{wen2024diffstgprobabilisticspatiotemporalgraph,
|
||||
title={DiffSTG: Probabilistic Spatio-Temporal Graph Forecasting with Denoising Diffusion Models},
|
||||
author={Haomin Wen and Youfang Lin and Yutong Xia and Huaiyu Wan and Qingsong Wen and Roger Zimmermann and Yuxuan Liang},
|
||||
year={2024},
|
||||
eprint={2301.13629},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.LG},
|
||||
url={https://arxiv.org/abs/2301.13629},
|
||||
}
|
||||
|
||||
@misc{liu2023pristiconditionaldiffusionframework,
|
||||
title={PriSTI: A Conditional Diffusion Framework for Spatiotemporal Imputation},
|
||||
author={Mingzhe Liu and Han Huang and Hao Feng and Leilei Sun and Bowen Du and Yanjie Fu},
|
||||
year={2023},
|
||||
eprint={2302.09746},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.LG},
|
||||
url={https://arxiv.org/abs/2302.09746},
|
||||
}
|
||||
|
||||
@misc{kong2021diffwaveversatilediffusionmodel,
|
||||
title={DiffWave: A Versatile Diffusion Model for Audio Synthesis},
|
||||
author={Zhifeng Kong and Wei Ping and Jiaji Huang and Kexin Zhao and Bryan Catanzaro},
|
||||
year={2021},
|
||||
eprint={2009.09761},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={eess.AS},
|
||||
url={https://arxiv.org/abs/2009.09761},
|
||||
}
|
||||
|
||||
@ARTICLE{11087622,
|
||||
author={Liu, Xiaosi and Xu, Xiaowen and Liu, Zhidan and Li, Zhenjiang and Wu, Kaishun},
|
||||
journal={IEEE Transactions on Mobile Computing},
|
||||
title={Spatio-Temporal Diffusion Model for Cellular Traffic Generation},
|
||||
year={2026},
|
||||
volume={25},
|
||||
number={1},
|
||||
pages={257-271},
|
||||
keywords={Base stations;Diffusion models;Data models;Uncertainty;Predictive models;Generative adversarial networks;Knowledge graphs;Mobile computing;Telecommunication traffic;Semantics;Cellular traffic;data generation;diffusion model;spatio-temporal graph},
|
||||
doi={10.1109/TMC.2025.3591183}}
|
||||
|
||||
@misc{austin2023structureddenoisingdiffusionmodels,
|
||||
title={Structured Denoising Diffusion Models in Discrete State-Spaces},
|
||||
author={Jacob Austin and Daniel D. Johnson and Jonathan Ho and Daniel Tarlow and Rianne van den Berg},
|
||||
year={2023},
|
||||
eprint={2107.03006},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.LG},
|
||||
url={https://arxiv.org/abs/2107.03006},
|
||||
}
|
||||
|
||||
@misc{li2022diffusionlmimprovescontrollabletext,
|
||||
title={Diffusion-LM Improves Controllable Text Generation},
|
||||
author={Xiang Lisa Li and John Thickstun and Ishaan Gulrajani and Percy Liang and Tatsunori B. Hashimoto},
|
||||
year={2022},
|
||||
eprint={2205.14217},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.CL},
|
||||
url={httpsarxiv.orgabs2205.14217},
|
||||
}
|
||||
|
||||
@misc{meng2025aflnetyearslatercoverageguided,
|
||||
title={AFLNet Five Years Later: On Coverage-Guided Protocol Fuzzing},
|
||||
author={Ruijie Meng and Van-Thuan Pham and Marcel Böhme and Abhik Roychoudhury},
|
||||
year={2025},
|
||||
eprint={2412.20324},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.SE},
|
||||
url={https://arxiv.org/abs/2412.20324},
|
||||
}
|
||||
|
||||
@misc{godefroid2017learnfuzzmachinelearninginput,
|
||||
title={Learn&Fuzz: Machine Learning for Input Fuzzing},
|
||||
author={Patrice Godefroid and Hila Peleg and Rishabh Singh},
|
||||
year={2017},
|
||||
eprint={1701.07232},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.AI},
|
||||
url={https://arxiv.org/abs/1701.07232},
|
||||
}
|
||||
|
||||
@misc{she2019neuzzefficientfuzzingneural,
|
||||
title={NEUZZ: Efficient Fuzzing with Neural Program Smoothing},
|
||||
author={Dongdong She and Kexin Pei and Dave Epstein and Junfeng Yang and Baishakhi Ray and Suman Jana},
|
||||
year={2019},
|
||||
eprint={1807.05620},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.CR},
|
||||
url={https://arxiv.org/abs/1807.05620},
|
||||
}
|
||||
|
||||
@misc{hoogeboom2021argmaxflowsmultinomialdiffusion,
|
||||
title={Argmax Flows and Multinomial Diffusion: Learning Categorical Distributions},
|
||||
author={Emiel Hoogeboom and Didrik Nielsen and Priyank Jaini and Patrick Forré and Max Welling},
|
||||
year={2021},
|
||||
eprint={2102.05379},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={stat.ML},
|
||||
url={https://arxiv.org/abs/2102.05379},
|
||||
}
|
||||
|
||||
@misc{dai2019transformerxlattentivelanguagemodels,
|
||||
title={Transformer-XL: Attentive Language Models Beyond a Fixed-Length Context},
|
||||
author={Zihang Dai and Zhilin Yang and Yiming Yang and Jaime Carbonell and Quoc V. Le and Ruslan Salakhutdinov},
|
||||
year={2019},
|
||||
eprint={1901.02860},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.LG},
|
||||
url={https://arxiv.org/abs/1901.02860},
|
||||
}
|
||||
|
||||
@misc{zhou2021informerefficienttransformerlong,
|
||||
title={Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting},
|
||||
author={Haoyi Zhou and Shanghang Zhang and Jieqi Peng and Shuai Zhang and Jianxin Li and Hui Xiong and Wancai Zhang},
|
||||
year={2021},
|
||||
eprint={2012.07436},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.LG},
|
||||
url={https://arxiv.org/abs/2012.07436},
|
||||
}
|
||||
|
||||
@misc{wu2022autoformerdecompositiontransformersautocorrelation,
|
||||
title={Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting},
|
||||
author={Haixu Wu and Jiehui Xu and Jianmin Wang and Mingsheng Long},
|
||||
year={2022},
|
||||
eprint={2106.13008},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.LG},
|
||||
url={https://arxiv.org/abs/2106.13008},
|
||||
}
|
||||
|
||||
@misc{zhou2022fedformerfrequencyenhanceddecomposed,
|
||||
title={FEDformer: Frequency Enhanced Decomposed Transformer for Long-term Series Forecasting},
|
||||
author={Tian Zhou and Ziqing Ma and Qingsong Wen and Xue Wang and Liang Sun and Rong Jin},
|
||||
year={2022},
|
||||
eprint={2201.12740},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.LG},
|
||||
url={https://arxiv.org/abs/2201.12740},
|
||||
}
|
||||
|
||||
@article{2023,
|
||||
title={A Note on Extremal Sombor Indices of Trees with a Given Degree Sequence},
|
||||
volume={90},
|
||||
ISSN={0340-6253},
|
||||
url={http://dx.doi.org/10.46793/match.90-1.197D},
|
||||
DOI={10.46793/match.90-1.197d},
|
||||
number={1},
|
||||
journal={Match Communications in Mathematical and in Computer Chemistry},
|
||||
publisher={University Library in Kragujevac},
|
||||
author={Damjanović, Ivan and Milošević, Marko and Stevanović, Dragan},
|
||||
year={2023},
|
||||
pages={197–202} }
|
||||
|
||||
Reference in New Issue
Block a user