Skip to main navigation Skip to search Skip to main content

The batch primary components transformer and auto-plasticity learning linear units architecture: synthetic image generation case

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

4 Citations (Scopus)

Abstract

Context tokenizing, which is popular in Large Language and Foundation Models (LLM, FM), leads to their excessive dimensionality inflation. Traditional Transformer models strive to reduce intractable excessive dimensionality at the among-token attention level, while we propose additional between-dimensions attention mechanism for dimensionality reduction. A novel Transformer-based architecture is presented, which aims at the individual dimension attention and, by doing so, performs the implicit relevant primary components' feature selection in artificial neural networks (ANN). As an additional mechanism allowing adaptive plasticity learning in ANN, a neuron-specific Learning Rectified Linear Unit layer is proposed for further feature selection via weight decay. The performance of the presented layers is tested on the encoder-decoder architecture applied for the synthetic image generation task for the benchmark MNIST data set.
Original languageEnglish
Title of host publicationProceedings - 2023 10th International Conference on Social Networks Analysis, Management and Security, SNAMS 2023
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9798350318906
ISBN (Print)9798350329957
DOIs
Publication statusPublished - 2 Jan 2024
Event2023 Tenth International Conference on Social Networks Analysis, Management and Security (SNAMS) - Abu Dhabi
Duration: 21 Nov 202324 Nov 2023

Conference

Conference2023 Tenth International Conference on Social Networks Analysis, Management and Security (SNAMS)
CityAbu Dhabi
Period21/11/2324/11/23
Other2023 Tenth International Conference on Social Networks Analysis, Management and Security (SNAMS) (21/11/2023-24/11/2023, Abu Dhabi)

Keywords

  • ANN plasticity
  • catastrophic forgetting
  • cosine distance
  • feature selection
  • learning ReLU
  • transformer
  • Transformer

ASJC Scopus subject areas

  • Management of Technology and Innovation
  • Computer Networks and Communications
  • Information Systems
  • Information Systems and Management
  • Safety, Risk, Reliability and Quality
  • Communication

Fingerprint

Dive into the research topics of 'The batch primary components transformer and auto-plasticity learning linear units architecture: synthetic image generation case'. Together they form a unique fingerprint.

Cite this