Skip to main content
Loading...
Scholar9 logo True scholar network
  • Article ▼
    • Article List
    • Deposit Article
  • Mentorship ▼
    • Overview
    • Sessions
  • Questions
  • Scholars
  • Institutions
  • Journals
  • Login/Sign up
Back to Top

Transparent Peer Review By Scholar9

A Comparative Study on AI/ML Optimization Strategies within DevOps Pipelines Deployed on Serverless Architectures in AWS Cloud Platforms

Abstract

The application of Artificial Intelligence (AI) and Machine Learning (ML) in modern DevOps pipelines is a rapidly growing trend, with organizations seeking efficient, scalable, and cost-effective solutions to integrate AI/ML models into production environments. AWS's serverless architecture, with its powerful cloud-native services such as AWS Lambda, Step Functions, and SageMaker, provides a flexible platform for deploying AI/ML workloads at scale. However, while the serverless paradigm offers considerable benefits in terms of scalability and resource management, it also presents unique challenges, including cold start latency, resource allocation, and computational efficiency. This research focuses on a comparative analysis of AI/ML optimization strategies deployed within DevOps pipelines on AWS's serverless architectures. The aim is to identify and evaluate the various optimization strategies available to enhance the performance of AI/ML models, mitigate existing challenges, and improve the efficiency and cost-effectiveness of cloud-based DevOps workflows. This paper reviews optimization techniques such as hyperparameter tuning, model compression, pruning, batch inference, and parallel processing, and their impact on the performance of ML models deployed within AWS Lambda and SageMaker environments. The study involves the empirical evaluation of real-world use cases, providing insights into the trade-offs between model accuracy, resource consumption, and execution time. Key findings suggest that while AWS serverless platforms provide excellent scalability and ease of use, careful management of resources and optimization of workflows is essential to maximize their potential. Furthermore, this paper contributes to the field by proposing recommendations for best practices in optimizing AI/ML workflows in serverless environments, while offering insights into future research directions.

Vinodkumar Surasani Reviewer

badge Review Request Accepted

Vinodkumar Surasani Reviewer

23 Apr 2025 11:44 AM

badge Approved

Relevance and Originality

Methodology

Validity & Reliability

Clarity and Structure

Results and Analysis

Relevance and Originality:

The research presents a timely and highly relevant investigation into the optimization of AI/ML workloads within DevOps pipelines using AWS's serverless architecture. As organizations increasingly rely on cloud-native services like AWS Lambda, Step Functions, and SageMaker for scalable and cost-efficient deployments, understanding how to enhance these workflows is critical. The paper distinguishes itself by focusing not only on the application of AI/ML, but also on addressing platform-specific challenges such as cold start latency and resource allocation—an area with growing importance and relatively limited academic coverage. The contribution is original in its comparative evaluation of optimization strategies specifically within the serverless paradigm.

Methodology:

The study employs an empirical approach centered on real-world use cases, which enhances the relevance and applicability of the findings. The focus on optimization techniques such as model pruning, batch inference, and hyperparameter tuning reflects an awareness of practical DevOps challenges in serverless environments. The comparative nature of the analysis provides a solid foundation for drawing performance-based insights. However, the methodology would benefit from a more explicit description of evaluation metrics and deployment configurations to improve transparency and reproducibility.

Validity & Reliability:

The findings are well-aligned with the objectives of the research, demonstrating clear trade-offs between model accuracy, resource consumption, and execution time. By evaluating performance across multiple optimization strategies, the study builds a robust foundation for its conclusions. However, while the results appear credible and grounded in practical scenarios, broader generalizability would be enhanced by including a greater diversity of model types and use cases across different AI/ML domains. Nonetheless, the consistency of observed trends supports the internal validity of the research.

Clarity and Structure:

The Research Article is well-organized and clearly structured, with a logical flow from problem statement through to optimization strategies and findings. Technical concepts are presented in a concise and digestible manner, ensuring accessibility to readers with varying levels of expertise in cloud computing and machine learning. The discussion of AWS services is integrated seamlessly with the AI/ML workflow analysis, making the narrative coherent and informative.

Result Analysis:

The analysis provides a comprehensive view of how various optimization strategies affect the deployment of AI/ML models in AWS serverless environments. It effectively discusses the balance between cost-efficiency, performance, and computational constraints, while offering actionable recommendations. The research contributes practical insights into managing serverless ML workflows, reinforcing the need for strategic planning and optimization to fully leverage cloud-native platforms.

avatar

IJ Publication Publisher

Respected Sir,

Thank you for your detailed review and thoughtful comments. We appreciate your recognition of the research's originality, particularly in optimizing AI/ML workloads within AWS’s serverless architecture. Your observations on the relevance of addressing platform-specific challenges such as cold start latency and resource allocation are highly valued.

We also acknowledge your point regarding the transparency of evaluation metrics and deployment configurations, and we will consider providing more explicit details in future iterations to improve reproducibility. Additionally, your suggestion to incorporate a broader diversity of model types and use cases to enhance generalizability is constructive, and we agree that this would strengthen the research further.

Thank you again for your insightful feedback, which will certainly guide future work in this area.

Publisher

User Profile

IJ Publication

Reviewer

User Profile

Vinodkumar Surasani

More Detail

User Profile

Paper Category

Cloud Computing

User Profile

Journal Name

TIJER - Technix International Journal for Engineering Research

User Profile

p-ISSN

User Profile

e-ISSN

2349-9249

Subscribe us to get updated

logo logo

Scholar9 is aiming to empower the research community around the world with the help of technology & innovation. Scholar9 provides the required platform to Scholar for visibility & credibility.

QUICKLINKS

  • What is Scholar9?
  • About Us
  • Mission Vision
  • Contact Us
  • Privacy Policy
  • Terms of Use
  • Blogs
  • FAQ

CONTACT US

  • logo +91 82003 85143
  • logo hello@scholar9.com
  • logo www.scholar9.com

© 2025 Sequence Research & Development Pvt Ltd. All Rights Reserved.

whatsapp