The Science and Information (SAI) Organization
  • Home
  • About Us
  • Journals
  • Conferences
  • Contact Us

Publication Links

  • IJACSA
  • Author Guidelines
  • Publication Policies
  • Outstanding Reviewers

IJACSA

  • About the Journal
  • Call for Papers
  • Editorial Board
  • Author Guidelines
  • Submit your Paper
  • Current Issue
  • Archives
  • Indexing
  • Fees/ APC
  • Reviewers
  • Apply as a Reviewer

IJARAI

  • About the Journal
  • Archives
  • Indexing & Archiving

Special Issues

  • Home
  • Archives
  • Proposals
  • ICONS_BA 2025

Computer Vision Conference (CVC)

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact

Computing Conference

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact

Intelligent Systems Conference (IntelliSys)

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact

Future Technologies Conference (FTC)

  • Home
  • Call for Papers
  • Submit your Paper/Poster
  • Register
  • Venue
  • Contact
  • Home
  • Call for Papers
  • Editorial Board
  • Guidelines
  • Submit
  • Current Issue
  • Archives
  • Indexing
  • Fees
  • Reviewers
  • RSS Feed

DOI: 10.14569/IJACSA.2026.0170381
PDF

An Efficient Computational Framework for Scalable Learning in Complex Data Environments Using Deep Neural Networks

Author 1: Priyanto
Author 2: Heri Nurdiyanto

International Journal of Advanced Computer Science and Applications(IJACSA), Volume 17 Issue 3, 2026.

  • Abstract and Keywords
  • How to Cite this Article
  • {} BibTeX Source

Abstract: This study introduces an efficient computational framework designed to support scalable learning in complex data environments using deep neural networks. In many real-world settings, data are not only large in volume but also diverse in structure, noisy in quality, and constantly evolving. These conditions often make conventional deep learning pipelines difficult to scale and expensive to maintain, especially when computational resources are limited or when rapid model updates are required. To address these challenges, we propose a framework that integrates adaptive data preprocessing, modular neural network architectures, and resource-aware training strategies into a unified learning pipeline. The framework is built to balance learning performance with computational efficiency, allowing models to be trained and updated without excessive overhead. Experiments were conducted on multiple heterogeneous datasets representing different levels of data complexity and scale. The results show that the proposed approach consistently improves training stability and convergence speed while maintaining competitive predictive performance compared to standard deep learning setups. In addition, the framework demonstrates better adaptability when handling data distribution shifts, which are common in dynamic environments. These findings suggest that scalable learning does not necessarily require increasingly complex model designs, but rather thoughtful integration of computational strategies that align model behavior with data characteristics and system constraints. The proposed framework offers a practical pathway for deploying deep learning solutions in large-scale, real-world applications where efficiency, robustness, and scalability are equally important.

Keywords: Scalable learning; deep neural networks; computational framework; complex data environments; efficient training

Priyanto and Heri Nurdiyanto. “An Efficient Computational Framework for Scalable Learning in Complex Data Environments Using Deep Neural Networks”. International Journal of Advanced Computer Science and Applications (IJACSA) 17.3 (2026). http://dx.doi.org/10.14569/IJACSA.2026.0170381

@article{2026,
title = {An Efficient Computational Framework for Scalable Learning in Complex Data Environments Using Deep Neural Networks},
journal = {International Journal of Advanced Computer Science and Applications},
doi = {10.14569/IJACSA.2026.0170381},
url = {http://dx.doi.org/10.14569/IJACSA.2026.0170381},
year = {2026},
publisher = {The Science and Information Organization},
volume = {17},
number = {3},
author = {Priyanto and Heri Nurdiyanto}
}



Copyright Statement: This is an open access article licensed under a Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, even commercially as long as the original work is properly cited.

IJACSA

Upcoming Conferences

Computer Vision Conference (CVC) 2026

21-22 May 2026

  • Amsterdam, The Netherlands

Computing Conference 2026

9-10 July 2026

  • London, United Kingdom

Artificial Intelligence Conference 2026

3-4 September 2026

  • Amsterdam, The Netherlands

Future Technologies Conference (FTC) 2026

15-16 October 2026

  • Berlin, Germany
The Science and Information (SAI) Organization
BACK TO TOP

Computer Science Journal

  • About the Journal
  • Call for Papers
  • Submit Paper
  • Indexing

Our Conferences

  • Computer Vision Conference
  • Computing Conference
  • Intelligent Systems Conference
  • Future Technologies Conference

Help & Support

  • Contact Us
  • About Us
  • Terms and Conditions
  • Privacy Policy

The Science and Information (SAI) Organization Limited is a company registered in England and Wales under Company Number 8933205.