Skip to main content
Loading...
Scholar9 logo True scholar network
  • Login/Sign up
  • Scholar9
    Publications ▼
    Article List Deposit Article
    Mentorship ▼
    Overview Sessions
    Q&A Institutions Network Journals
    Publications ▼
    Article List Deposit Article
    Mentorship ▼
    Overview Sessions
    Q&A Institutions Scholars Journals
  • Login/Sign up
  • Back to Top

    Transparent Peer Review By Scholar9

    A Comparative Study on AI/ML Optimization Strategies within DevOps Pipelines Deployed on Serverless Architectures in AWS Cloud Platforms

    Abstract

    The application of Artificial Intelligence (AI) and Machine Learning (ML) in modern DevOps pipelines is a rapidly growing trend, with organizations seeking efficient, scalable, and cost-effective solutions to integrate AI/ML models into production environments. AWS's serverless architecture, with its powerful cloud-native services such as AWS Lambda, Step Functions, and SageMaker, provides a flexible platform for deploying AI/ML workloads at scale. However, while the serverless paradigm offers considerable benefits in terms of scalability and resource management, it also presents unique challenges, including cold start latency, resource allocation, and computational efficiency. This research focuses on a comparative analysis of AI/ML optimization strategies deployed within DevOps pipelines on AWS's serverless architectures. The aim is to identify and evaluate the various optimization strategies available to enhance the performance of AI/ML models, mitigate existing challenges, and improve the efficiency and cost-effectiveness of cloud-based DevOps workflows. This paper reviews optimization techniques such as hyperparameter tuning, model compression, pruning, batch inference, and parallel processing, and their impact on the performance of ML models deployed within AWS Lambda and SageMaker environments. The study involves the empirical evaluation of real-world use cases, providing insights into the trade-offs between model accuracy, resource consumption, and execution time. Key findings suggest that while AWS serverless platforms provide excellent scalability and ease of use, careful management of resources and optimization of workflows is essential to maximize their potential. Furthermore, this paper contributes to the field by proposing recommendations for best practices in optimizing AI/ML workflows in serverless environments, while offering insights into future research directions.

    User Profile
    User Profile

    Vinodkumar Surasani Reviewer

    badge Review Request Accepted

    Vinodkumar Surasani Reviewer

    badge Approved

    Relevance and Originality

    Methodology

    Validity & Reliability

    Clarity and Structure

    Results and Analysis

    Relevance and Originality:

    The research presents a timely and highly relevant investigation into the optimization of AI/ML workloads within DevOps pipelines using AWS's serverless architecture. As organizations increasingly rely on cloud-native services like AWS Lambda, Step Functions, and SageMaker for scalable and cost-efficient deployments, understanding how to enhance these workflows is critical. The paper distinguishes itself by focusing not only on the application of AI/ML, but also on addressing platform-specific challenges such as cold start latency and resource allocation—an area with growing importance and relatively limited academic coverage. The contribution is original in its comparative evaluation of optimization strategies specifically within the serverless paradigm.

    Methodology:

    The study employs an empirical approach centered on real-world use cases, which enhances the relevance and applicability of the findings. The focus on optimization techniques such as model pruning, batch inference, and hyperparameter tuning reflects an awareness of practical DevOps challenges in serverless environments. The comparative nature of the analysis provides a solid foundation for drawing performance-based insights. However, the methodology would benefit from a more explicit description of evaluation metrics and deployment configurations to improve transparency and reproducibility.

    Validity & Reliability:

    The findings are well-aligned with the objectives of the research, demonstrating clear trade-offs between model accuracy, resource consumption, and execution time. By evaluating performance across multiple optimization strategies, the study builds a robust foundation for its conclusions. However, while the results appear credible and grounded in practical scenarios, broader generalizability would be enhanced by including a greater diversity of model types and use cases across different AI/ML domains. Nonetheless, the consistency of observed trends supports the internal validity of the research.

    Clarity and Structure:

    The Research Article is well-organized and clearly structured, with a logical flow from problem statement through to optimization strategies and findings. Technical concepts are presented in a concise and digestible manner, ensuring accessibility to readers with varying levels of expertise in cloud computing and machine learning. The discussion of AWS services is integrated seamlessly with the AI/ML workflow analysis, making the narrative coherent and informative.

    Result Analysis:

    The analysis provides a comprehensive view of how various optimization strategies affect the deployment of AI/ML models in AWS serverless environments. It effectively discusses the balance between cost-efficiency, performance, and computational constraints, while offering actionable recommendations. The research contributes practical insights into managing serverless ML workflows, reinforcing the need for strategic planning and optimization to fully leverage cloud-native platforms.

    IJ Publication Publisher

    Respected Sir,

    Thank you for your detailed review and thoughtful comments. We appreciate your recognition of the research's originality, particularly in optimizing AI/ML workloads within AWS’s serverless architecture. Your observations on the relevance of addressing platform-specific challenges such as cold start latency and resource allocation are highly valued.

    We also acknowledge your point regarding the transparency of evaluation metrics and deployment configurations, and we will consider providing more explicit details in future iterations to improve reproducibility. Additionally, your suggestion to incorporate a broader diversity of model types and use cases to enhance generalizability is constructive, and we agree that this would strengthen the research further.

    Thank you again for your insightful feedback, which will certainly guide future work in this area.

    Publisher

    User Profile

    IJ Publication

    Reviewers

    User Profile

    Vinodkumar Surasani

    User Profile

    Ramesh Krishna Mahimalur

    More Detail

    User Profile

    Paper Category

    Cloud Computing

    User Profile

    Journal Name

    TIJER - Technix International Journal for Engineering Research

    User Profile

    p-ISSN

    User Profile

    e-ISSN

    2349-9249

    Subscribe us to get updated

    logo logo

    Scholar9 is aiming to empower the research community around the world with the help of technology & innovation. Scholar9 provides the required platform to Scholar for visibility & credibility.

    QUICKLINKS

    • What is Scholar9?
    • About Us
    • Mission Vision
    • Contact Us
    • Privacy Policy
    • Terms of Use
    • Blogs
    • FAQ

    CONTACT US

    • +91 82003 85143
    • hello@scholar9.com
    • www.scholar9.com

    © 2026 Sequence Research & Development Pvt Ltd. All Rights Reserved.

    whatsapp