Search Results for author: Matin Mortaheb

Found 7 papers, 0 papers with code

Transformer-Aided Semantic Communications

no code implementations2 May 2024 Matin Mortaheb, Erciyes Karakaya, Mohammad A. Amir Khojastepour, Sennur Ulukus

The transformer structure employed in large language models (LLMs), as a specialized category of deep neural networks (DNNs) featuring attention mechanisms, stands out for their ability to identify and highlight the most relevant aspects of input data.

Deep Learning-Based Real-Time Quality Control of Standard Video Compression for Live Streaming

no code implementations21 Nov 2023 Matin Mortaheb, Mohammad A. Amir Khojastepour, Srimat T. Chakradhar, Sennur Ulukus

The encoded bitrate and the quality of the compressed video depend on encoder parameters, specifically, the quantization parameter (QP).

Quantization Video Compression

Semantic Multi-Resolution Communications

no code implementations22 Aug 2023 Matin Mortaheb, Mohammad A. Amir Khojastepour, Srimat T. Chakradhar, Sennur Ulukus

The experiment with both datasets illustrates that our proposed method is capable of surpassing the SSCC method in reconstructing data with different resolutions, enabling the extraction of semantic features with heightened confidence in successive layers.

Multi-Task Learning

Personalized Decentralized Multi-Task Learning Over Dynamic Communication Graphs

no code implementations21 Dec 2022 Matin Mortaheb, Sennur Ulukus

Our algorithm uses exchanged gradients to calculate the correlations among tasks automatically, and dynamically adjusts the communication graph to connect mutually beneficial tasks and isolate those that may negatively impact each other.

Federated Learning Multi-Task Learning

Hierarchical Over-the-Air FedGradNorm

no code implementations14 Dec 2022 Cemil Vahapoglu, Matin Mortaheb, Sennur Ulukus

MTL can be integrated into a federated learning (FL) setting if tasks are distributed across clients and clients have a single shared network, leading to personalized federated learning (PFL).

Multi-Task Learning Personalized Federated Learning

FedGradNorm: Personalized Federated Gradient-Normalized Multi-Task Learning

no code implementations24 Mar 2022 Matin Mortaheb, Cemil Vahapoglu, Sennur Ulukus

In federated settings, the statistical heterogeneity due to different task complexities and data heterogeneity due to non-iid nature of local datasets can both degrade the learning performance of the system.

Multi-Task Learning Personalized Federated Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.