The (IEEE) 2026 International Conference on Revolutionary Artificial Intelligence and Future Applications
(Rev-AI 2026)

3–5 June, 2026 | Varna, Bulgaria
(Technically Co-Sponsored by IEEE)

Welcome to The (IEEE) 2026 International Conference on Revolutionary Artificial Intelligence and Future Applications (REV-AI 2026)

Call for Papers

The Rev-AI 2026 Conference is the premier global event convening pioneering researchers, industry leaders, and visionary innovators to explore the frontier of artificial intelligence. This three-day conference, set in the dynamic and innovative city of Varna, will showcase paradigm-shifting research, transformative applications, and collaborative dialogues on the technologies that are redefining the future of intelligence. Attendees will have the unparalleled opportunity to engage with the minds spearheading the AI revolution, and network with a global community of experts. From foundational models that rethink cognition to AI-driven solutions for humanity's greatest challenges, REV-AI 2026 promises a comprehensive expedition into the next era of artificial intelligence.

We are thrilled to invite the global community to the 2026 International Conference on Revolutionary Artificial Intelligence and Future Applications, the definitive event for shaping the future of intelligent systems.

  • Why Attend?
  • Discover the Next Frontier: Explore paradigm-shifting research and foundational breakthroughs that move beyond incremental progress. Engage with Pioneers: Join visionary keynotes and thought-provoking panels led by the architects of the AI revolution. Network Globally: Connect with an international network of leading academics, industry researchers, and startup founders.. Shape the Future: Participate in interactive workshops and sessions that delve into the ethical, societal, and technical foundations of tomorrow's AI.
  • Highlights
  • Unveiling next-generation AI architectures that surpass current limitations. Exploring Human-AI collaboration and cognitive augmentation. Tackling the grand challenges of AI safety, governance, and alignment. Showcasing revolutionary applications in medicine, climate science, and space exploration.

  • Don't miss this unparalleled opportunity to be at the epicenter of the next AI revolution. We invite submissions on all topics related to revolutionary AI, with a special interest in, but not limited to:
    • Foundational & Revolutionary AI Models

    • Next-Generation Architectures (e.g., beyond Transformers, State Space Models, Neuro-Symbolic AI)
    • AI for Scientific Discovery (AI4Sci)
    • Causal AI & Reasoning
    • Embodied AI and Active Perception
    • Foundation Models for Multimodal Understanding
    • Theoretical Limits of AI
    • Parameter-Efficient Fine-Tuning (PEFT): Advanced methods including LoRA (Low-Rank Adaptation), Adapters, and Prompt Tuning for rapidly adapting massive models to new tasks with minimal computational overhead.
    • QLoRA and Quantized PEFT: Techniques for fine-tuning quantized models (e.g., 4-bit) without performance degradation, enabling the adaptation of billion-parameter models on a single GPU.
    • Conditional Computation & Mixture-of-Experts (MoE): Advanced routing algorithms, training stability for sparse models, and dynamic activation of model parts for extreme efficiency at scale.
    • Mathematical Reformulations for Efficiency: Replacing core operations (e.g., attention mechanisms) with more computationally efficient approximations without loss of performance.
    • State Space Models (SSMs) for Efficiency: Leveraging SSMs like Mamba for sequential data handling that is fundamentally faster and more memory-efficient than traditional Transformers.
    • Training & Learning Process Optimization

    • Efficient Fine-Tuning Paradigms: Novel approaches to LoRA (e.g., DoRA, VeRA) and fusion with other methods for greater efficiency and effectiveness.
    • Fast Convergence Techniques: Novel optimizers, learning rate schedules, and training curricula that reduce total training time and computational cost.
    • Gradient-Free & Few-Step Optimization: Exploring methods like Evolutionary Strategies for training where backpropagation is impractical.
    • Sparse Training: Techniques to train a model with a sparse architecture from the very beginning, avoiding the expensive "train-then-prune" cycle.
    • Inference & Deployment Optimization

    • Dynamic Neural Networks: Models that can adapt their inference path based on input complexity, slashing latency for "easier" tasks.
    • Speculative Decoding & Lookahead Reasoning: Using smaller, faster models to "draft" responses verified by a larger model, dramatically accelerating LLM inference.
    • Hardware-Aware Neural Compression: Co-designing compression techniques with specific hardware properties for maximal throughput.
    • Merging and Consolidating LoRA Adapters: Methods for efficiently merging multiple fine-tuned LoRA adapters into a single, powerful model without catastrophic interference.
    • Sustainable & Green AI

    • Energy-Aware Model Design: Techniques to directly model, measure, and minimize the energy consumption of AI models.
    • Carbon-Efficient Training: Scheduling and locating training jobs in data centers powered by renewable energy.
    • The Green Impact of PEFT: Quantifying the massive reduction in computational resources and energy enabled by methods like LoRA and QLoRA.
    • Unified Frameworks & Evaluation

    • Multi-Objective Optimization Frameworks: New tools that simultaneously optimize for accuracy, latency, memory, energy, and robustness.
    • Benchmarking Efficiency at Scale: Developing robust benchmarks for evaluating the optimization of massive models across diverse hardware.
    • The Efficiency-Ability Trade-off: Theoretical and empirical studies on the relationship between a model's computational budget and its emergent capabilities.
    • AI & Human Collaboration & Society

    • Human-AI Teaming and Cognitive Augmentation
    • AI for Creativity, Art, and Co-Creation
    • AI-Driven Education and Personalized Learning
    • The Future of Work and Economics in an AI-Dominant World
    • AI for Governance and Public Policy
    • Philosophical and Ethical Frameworks for AGI
    • Conversational AI platforms such as:

    • ChatGPT (OpenAI), Gemini (Google), Copilot (Microsoft), Claude (Anthropic), Meta AI (Meta), Grok (xAI) Leading Open-Weight Models: Llama (Meta), Mistral AI, Qwen (Alibaba)
    • Major Chinese Platforms such as:

    • DeepSeek, Ernie Bot (Baidu), Tongyi Qianwen (Alibaba), Zhipu AI
    • Specialized Coding Assistants such as:

    • GitHub Copilot, Replit CodeComplete, Amazon CodeWhisperer, Codium
    • More related AI topics

    • Next-Generation Model Architectures for Efficiency (e.g., Mixture-of-Experts, State Space Models)
    • Low-Rank Adaptation (LoRA)
    • Quantized Fine-Tuning and QLoRA
    • Speculative Decoding and Advanced Inference Acceleration
    • Conditional Computation and Dynamic Neural Networks
    • Green AI: Energy-Aware Training and Sustainable Model Design
    • Neural Architecture Search (NAS) and Automated Model Co-Design
    • Hardware-Aware Compression and Ultra-Low-Bit Quantization
    • Optimization for Edge-Cloud Hybrid Intelligence Systems
    • Fast Convergence and Efficient Learning Algorithms
    • Parameter-Efficient Fine-Tuning (PEFT)
    • AI, Machine Learning, Deep Learning, Federated Learning and Creativity
    • Innovations in AI Algorithm Development
    • Natural Language Processing (LNP)
    • Augmented Reality (AR), Virtual Reality (VR), and Extended reality (XR)
    • Developing Trustworthy and Reliable AI Systems
    • Intelligent Intrusion Detection Systems for Internet of Things (IoT) Applications
    • Network Forensics Leveraging Intelligent Systems and Data Analytics
    • Data Analytics Approaches for Privacy-by-Design in Smart Healthcare Systems
    • Datasets, Benchmarks, and Open-Source Tools for Cybersecurity Applications
    • Efficient Deep Learning Techniques for Resource-Constrained Environments
    • Adversarial Machine Learning and Mitigation of Backdoor Attacks
    • Blockchain Technologies for Strengthening Cybersecurity Frameworks
    • Advanced Intelligent Solutions and Data Analytics for Enhancing Cloud and Edge Security
    • Malware Detection and Vulnerability Assessment Using Intelligent Systems
    • Intelligent Approaches for Detecting and Mitigating Misinformation
    • Intelligent Systems for Detecting Cyber-Attacks Effectively

    Submission Guidelines:

    Paper format

    Submitted papers (.pdf format) must use the A4 IEEE Manuscript Templates for Conference Proceedings.

    Please make use of the appropriate IEEE template for conference proceedings to prepare your revised manuscript.
    Failure to do so may result in excluding your paper from the conference proceedings.
    DO NOT include headers/footers or page numbers in the final submission.

    Length

    Submitted papers may be up to 8 pages includes the references.

    Originality

    Papers submitted to Rev-AI 2026 must be the original work of the authors. The may not be simultaneously under review elsewhere. Publications that have been peer-reviewed and have appeared at other conferences or workshops may not be submitted to anywhere else. Authors should be aware that IEEE has a strict policy with regard to plagiarism https://www.ieee.org/publications/rights/plagiarism/plagiarism-faq.html The authors' prior work must be cited appropriately.

    Proofreading:

    Please proofread your submission carefully. It is essential that the language use in the paper is clear and correct so that it is easily understandable. (Either US English or UK English spelling conventions are acceptable.)

    Author list

    Please ensure that you submit your papers with the full and final list of authors in the correct order. The author list registered for each submission is not allowed to be changed in any way after the paper submission deadline.

    IMPORTANT: No-Show Policy

    At least one author of each accepted paper is required to register to the conference and present the paper.

    Publication:

    All papers that are accepted, registered, and presented in Rev-AI and the workshops co-located with it will be submitted to IEEEXplore for publication under the condition of acceptance from IEEE.