Praxis: A Hybrid Mobile AI Application with Offline LLM Inference and Cloud API Integration
  • Author(s): Sri Jaya Lingeswaran A; Keithick R; Niranjan R; Adrash S
  • Paper ID: 1714023
  • Page: 2555-2559
  • Published Date: 16-04-2026
  • Published In: Iconic Research And Engineering Journals
  • Publisher: IRE Journals
  • e-ISSN: 2456-8880
  • Volume/Issue: Volume 9 Issue 8 February-2026
Abstract

Large Language Models (LLMs) have revolutionized natural language processing and artificial intelligence applications. However, deploying these models on mobile devices presents significant challenges due to limited computational resources, memory constraints, and energy consumption. This paper introduces Praxis, a novel hybrid mobile application architecture that seamlessly integrates on-device LLM inference with cloud-based API services. Praxis enables users to run small language models locally on Android devices for enhanced privacy and offline capability while providing the flexibility to switch to more powerful cloud-based models when connectivity and computational demands permit. Our implementation leverages React Native, Expo, and integration with Hugging Face model repositories, combined with secure API key management for services like OpenAI, Anthropic Claude, and Google Gemini. Experimental results demonstrate that Praxis achieves a balance between performance, privacy, energy efficiency, and user experience, making advanced AI capabilities accessible on resourceconstrained mobile devices.

Keywords

Large Language Models, Mobile Computing, On-Device Inference, Edge Computing, Hybrid Architecture, Privacy-Preserving AI, React Native

Citations

IRE Journals:
Sri Jaya Lingeswaran A, Keithick R, Niranjan R, Adrash S "Praxis: A Hybrid Mobile AI Application with Offline LLM Inference and Cloud API Integration" Iconic Research And Engineering Journals Volume 9 Issue 8 2026 Page 2555-2559 https://doi.org/10.64388/IREV9I8-1714023

IEEE:
Sri Jaya Lingeswaran A, Keithick R, Niranjan R, Adrash S "Praxis: A Hybrid Mobile AI Application with Offline LLM Inference and Cloud API Integration" Iconic Research And Engineering Journals, 9(8) https://doi.org/10.64388/IREV9I8-1714023