Transformer Deep Learning Architectures: Advances and Applications

A special issue of Applied Sciences (ISSN 2076-3417). This special issue belongs to the section "Computing and Artificial Intelligence".

Deadline for manuscript submissions: 25 August 2024 | Viewed by 1641

Special Issue Editor

Department of Information Science, Department of Computer Science and Engineering (Joint Appointment), University of North Texas, Denton, TX 76203, USA
Interests: big data/data science; machine/deep learning; software development; health informatics; sensor information extraction

Special Issue Information

Dear Colleagues,

This Special Issue spotlights the advancements in and applications of Transformer-based deep learning architectures. Transformers have significantly influenced artificial intelligence (AI), particularly natural language processing (NLP), with their innovative approach to handling sequential data. This Special Issue explores the core components of these architectures, including their self-attention mechanism and positional encoding, and discusses recent developments that enhance efficiency, interpretability, and scalability.

The Special Issue also delves into the broad spectrum of applications of Transformers, ranging from traditional tasks such as text summarization, machine translation, and sentiment analysis, to innovative utilizations in language generation and conversational AI, including chatbots and dialogue systems like ChatGPT. Beyond these conventional domains, the Special Issue also highlights breakthrough applications in emerging fields such as computer vision, bioinformatics, health informatics and climate modeling. It provides insight into how models such as BERT and GPT are changing paradigms across various sectors.

Moreover, this Special Issue tackles the existing challenges in utilizing Transformer models, giving readers a well-rounded view of this field. It outlines potential future directions, providing a roadmap for continued innovation. This comprehensive guide offers invaluable insights to researchers, students, and practitioners interested in the cutting edge of deep learning technology.

Dr. Ting Xiao
Guest Editor

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Applied Sciences is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2400 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • AI
  • NLP
  • transformer
  • self-attention
  • ChatGPT
  • BERT
  • GPT
  • deep learning

Published Papers (2 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

31 pages, 15142 KiB  
Article
Scaling Implicit Bias Analysis across Transformer-Based Language Models through Embedding Association Test and Prompt Engineering
by Ravi Varma Kumar Bevara, Nishith Reddy Mannuru, Sai Pranathi Karedla and Ting Xiao
Appl. Sci. 2024, 14(8), 3483; https://doi.org/10.3390/app14083483 - 20 Apr 2024
Viewed by 270
Abstract
In the evolving field of machine learning, deploying fair and transparent models remains a formidable challenge. This study builds on earlier research, demonstrating that neural architectures exhibit inherent biases by analyzing a broad spectrum of transformer-based language models from base to x-large configurations. [...] Read more.
In the evolving field of machine learning, deploying fair and transparent models remains a formidable challenge. This study builds on earlier research, demonstrating that neural architectures exhibit inherent biases by analyzing a broad spectrum of transformer-based language models from base to x-large configurations. This article investigates movie reviews for genre-based bias, which leverages the Word Embedding Association Test (WEAT), revealing that scaling models up tends to mitigate bias, with larger models showing up to a 29% reduction in prejudice. Alternatively, this study also underscores the effectiveness of prompt-based learning, a facet of prompt engineering, as a practical approach to bias mitigation, as this technique reduces genre bias in reviews by more than 37% on average. This suggests that the refinement of development practices should include the strategic use of prompts in shaping model outputs, highlighting the crucial role of ethical AI integration to weave fairness seamlessly into the core functionality of transformer models. Despite the basic nature of the prompts employed in this research, this highlights the possibility of embracing structured prompt engineering to create AI systems that are ethical, equitable, and more responsible for their actions. Full article
(This article belongs to the Special Issue Transformer Deep Learning Architectures: Advances and Applications)
Show Figures

Figure 1

20 pages, 6446 KiB  
Article
ChatGPT Translation of Program Code for Image Sketch Abstraction
by Yulia Kumar, Zachary Gordon, Oluwatunmise Alabi, Jenny Li, Kathryn Leonard, Linda Ness and Patricia Morreale
Appl. Sci. 2024, 14(3), 992; https://doi.org/10.3390/app14030992 - 24 Jan 2024
Viewed by 726
Abstract
In this comprehensive study, a novel MATLAB to Python (M-to-PY) conversion process is showcased, specifically tailored for an intricate image skeletonization project involving fifteen MATLAB files and a large dataset. The central innovation of this research is the adept use of ChatGPT-4 as [...] Read more.
In this comprehensive study, a novel MATLAB to Python (M-to-PY) conversion process is showcased, specifically tailored for an intricate image skeletonization project involving fifteen MATLAB files and a large dataset. The central innovation of this research is the adept use of ChatGPT-4 as an AI assistant, pivotal in crafting a prototype M-to-PY converter. This converter’s capabilities were thoroughly evaluated using a set of test cases generated by the Bard bot, ensuring a robust and effective tool. The culmination of this effort was the development of the Skeleton App, adept at image sketching and skeletonization. This live and publicly available app underscores the enormous potential of AI in enhancing the transition of scientific research from MATLAB to Python. The study highlights the blend of AI’s computational prowess and human ingenuity in computational research, making significant strides in AI-assisted scientific exploration and tool development. Full article
(This article belongs to the Special Issue Transformer Deep Learning Architectures: Advances and Applications)
Show Figures

Figure 1

Planned Papers

The below list represents only planned manuscripts. Some of these manuscripts have not been received by the Editorial Office yet. Papers submitted to MDPI journals are subject to peer-review.

Title: Predictive Analysis of X Post Privatization Using ChatGPT and Deep Learning Techniques
Author: Kumar
Highlights: Developed a unique approach using LLMs and NLC for estimating the value of "X" post privatization. Combined traditional financial valuation with news sentiment analysis for precision. Utilized ChatGPT for creating custom transformers tailored to financial predictions. Showcased deep learning's role in modernizing financial forecasting.

Title: Scaling Implicit Bias Analysis across Transformer-based Language Models through Embedding Association Test and Prompt Engineering
Authors: Ravi Varma Kumar Bevara; Nishith Reddy Mannuru; Sai Pranathi Karedla; Ting Xiao
Affiliation: Department of Information Science, Department of Computer Science and Engineering, University of North Texas
Abstract: In the evolving field of machine learning, the deployment of fair and transparent models remains a challenge. This study builds on earlier research showing that neural architectures have built-in biases by looking at a wide range of transformer-based language models, from basic to extra-large configurations, for analyzing movie genre-based bias in reviews. Utilizing the Word Embedding Association Test, it was found that scaling models up tends to mitigate bias, with larger models showing up to a 29% reduction in prejudice. Remarkably, custom prompting emerges as a superior strategy, reducing bias by more than 37% on average. A bias mitigation framework is proposed that harmonizes scale-adapted prompts with model biases, advocating for the refinement of development practices. This study highlights the potential of prompt engineering in shaping model outputs, akin to neural conditioning in biological systems, and paves the way for interdisciplinary approaches to embed fairness into the fabric of transformer models. Despite computational limitations, these findings illuminate the path toward equitable machine intelligence by leveraging the dynamic nature of neural networks and structured prompts.

Back to TopTop