IndoAI

Smarter at the Edge: Evaluating Decentralized AI Deployment Models in Federated,Hierarchical, Microservices and Serverless edge AI Architectures

Abstract: The advent of edge computing and artificial intelligence (AI) has spurred the development of decentralized architectures to address privacy, latency, scalability, and cost challenges in AI deployment. This paper provides an in-depth examination of four architectures—Federated Learning (FL), Edge-Cloud Hierarchical, Microservices-Based Edge AI, and Serverless Edge AI—focusing on their components, workflows, advantages, challenges, and use cases. A comparative analysis evaluates their suitability across application domains, technological requirements, usefulness, privacy, scalability, latency, and cost-effectiveness, supported by a comprehensive review of recent literature and industry implementations. The findings highlight FL’s primacy in privacy-sensitive training, Edge-Cloud’s efficacy in latency-critical tasks, Microservices’ modularity for complex systems, and Serverless’ cost-efficiency for event-driven applications. Integrating these architectures with visionLLMs/VLM is on the horizon to train and develop various AI Models.

Keywords: Federated Learning (FL), Edge-Cloud Hierarchical, Microservices-Based Edge AI, Serverless Edge AI, vision LLM, edge AI, AI Camera, AI Models

Smarter at the Edge Evaluating Decentralized AI Deployment Models

Excerpt:

6 Comparative Analysis

    The four architectures differ in their suitability for specific applications, technological requirements, usefulness, and additional factors (e.g., privacy, scalability, latency, cost). This section provides a rigorous comparison, supported by metrics and case studies.

    6.1 Application Domains

    6.2 Technological Requirements

    6.3 Usefulness

    6.4 Additional Factors

    6.5 Summary Table

    CriteriaFederated LearningEdge-Cloud HierarchicalMicroservices-BasedServerless Edge AI
    ApplicationPrivacy-sensitive trainingLatency-critical hybrid tasksModular multi-task AIEvent-driven tasks
    Tech StackTensorFlow Federated, SMPCTensorFlow Lite, MQTTDocker, KubernetesAWS Greengrass, FaaS
    PrivacyHighModerateModerateLow
    ScalabilityExcellentGoodExcellentGood
    LatencyN/ALow (<10ms)Moderate (10–50ms)Moderate(100–500ms)
    CostHigh setup, low operationModerateHighLow
    EaseComplexModerateComplexEasy
    Metrics30% faster convergence<10ms latency50% faster updates40% cost savings
    1. Discussion

    A 2018 study [104] concurs that (a) the core cloud-only system outperforms the edge-only system having low inter-edge bandwidth, (b) a distributed edge cloud selection scheme can approach the global optimal assignment when the edge has sufficient compute resources and high inter-edge bandwidth and (c) adding capacity to an existing edge network without increasing the inter-edge bandwidth contributes to networkwide congestion and can reduce system capacity.

    NVIDIA Clara Federated Learning [96] uses distributed training across multiple hospitals to develop robust AI models without sharing personal data. Federated Learning [76] excels in privacy-sensitive training, making it ideal for healthcare (e.g., like  Apollo Hospitals under DPDP Act 2023) and smartphones (e.g., Xiaomi’s vernacular models). Its scalability supports large-scale deployments, but it is not suited for real-time inference[113]. Edge-Cloud Hierarchical[97] is optimal for latency-critical applications, such as autonomous vehicles (e.g., Tata Motors) and smart cities (e.g., India’s Smart Cities Mission), balancing edge and cloud strengths. Microservices-Based Edge AI suits complex, modular systems, divides big apps into smaller services, like video analytics (e.g., Reliance Retail’s surveillance) and smart retail, but its resource demands limit edge applicability. Serverless Edge AI is best for event-driven[98], cost-sensitive tasks, such as smart homes (e.g., Jio IoT) and surveillance, though cold start latency is a constraint[99].

    In India, FL aligns with privacy regulations, Edge-Cloud supports infrastructure growth (e.g., 5G rollout), Microservices enables retail innovation[100] and Serverless offers cost-effective IoT solutions[91]. Stakeholders should select architectures based on application needs—privacy for FL, latency for Edge-Cloud, modularity for Microservices and cost for Serverless.

    1. Future Directions

    India Context: The edge ai market in India is expected to reach a projected revenue of US$ 3,656.6 million by 2030. These architectures can drive Digital India, Ayushman Bharat and smart city initiatives[101].

    Following full paper can be accessed from

    1. https://jscglobal.org/volume-14-issue-6-june-2025/
    2. https://www.researchgate.net/publication/392450561_Smarter_at_the_Edge_Evaluating_Decentralized_AI_Deployment_Models_in_Federated_Hierarchical_Microservices_and_Serverless_edge_AI_Architectures
    Exit mobile version