Research & Sources

Our technical foundation is validated by 60+ peer-reviewed papers from top-tier venues (MLSys, ICLR, Nature, ACM, IEEE, NeurIPS) published between 2020-2025. We build on proven techniques, not experimental technology.

1. LoRA Stacking & Multi-Adapter Composition

Research demonstrates that LoRA adapters can be composed and stacked to create multi-capability systems on a single base model.

  • S-LoRA: Serving Thousands of Concurrent LoRA Adapters (2024)

    Publication: MLSys 2024
    Demonstrated efficient serving of thousands of LoRA adapters simultaneously with dynamic loading/unloading, validating the scalability of LoRA-based architectures.
  • Cached Multi-LoRA Composition (2025)

    Publication: ICLR 2025
    Source: OpenReview
    Addresses challenges in multi-LoRA composition with caching strategies for improved performance.
  • Merging LoRAs for Practical Skill Composition Tasks (2025)

    Publication: COLING 2025
    Source: ACL Anthology
    Demonstrates that LoRAs trained independently can be merged/composed for multi-skill tasks.
  • A Survey on LoRA of Large Language Models (2025)

    Publication: Frontiers of Computer Science
    Source: Springer
    Comprehensive survey showing LoRA as one of the best-performed parameter-efficient fine-tuning paradigms.
  • FashionGPT: LLM Instruction Fine-tuning with Multiple LoRA-Adapter Fusion (2024)

    Publication: Knowledge-Based Systems (ScienceDirect)
    Source: ScienceDirect
    Multiple LoRA-adapter fusion fine-tuning outperforms dataset fusion approaches.

2. Domain-Specific Fine-Tuning for Software Engineering

Research shows that domain-specific fine-tuning enables models to understand software engineering workflows and best practices.

  • Fine-Tuning Foundation Models for Domain-Specific Test Case Generation (2024)

    Publication: Journal of Information Systems Engineering and Management
    Source: JISEM Journal
    Domain-specific fine-tuning significantly improves efficiency and coverage in software testing processes.
  • Fine-tuning Large Language Models for Domain Adaptation (2025)

    Publication: npj Computational Materials (Nature)
    Source: Nature
    Domain adaptation through fine-tuning requires careful strategies to introduce new knowledge while retaining base capabilities.
  • LLMs for Software Engineering: A Survey (2024)

    Publication: GitHub Repository (iSEngLab/AwesomeLLM4SE)
    Source: GitHub
    Comprehensive survey covering code generation, code review, bug detection, test generation, and program repair.

3. Multi-Agent LLM Orchestration

Research shows that multi-agent systems with proper coordination mechanisms can handle complex cross-domain tasks effectively.

  • Multi-Agent Collaboration Mechanisms: A Survey of LLMs (2025)

    Publication: arXiv (January 2025)
    Comprehensive framework characterizing collaboration mechanisms by actors, types, structures, strategies, and coordination.
  • Multi-Agent Collaboration via Evolving Orchestration (2025)

    Publication: arXiv
    Dynamic, adaptive orchestration mechanisms enable sophisticated multi-agent collaboration.
  • LLM-Based Multi-Agent Systems for Software Engineering (2024)

    Publication: arXiv
    Literature review showing growing adoption of multi-agent LLM systems specifically for software development workflows.
  • MetaGPT: Meta Programming for Multi-Agent Collaborative Framework (2024)

    Publication: ICLR 2024
    Multi-agent frameworks with role-based agents (like product manager, architect, engineer) outperform single-agent approaches.

4. Knowledge Pruning & Model Efficiency

Research demonstrates that knowledge pruning can remove irrelevant domains while maintaining model quality, enabling domain-optimized foundations.

  • Efficient Self-Attention with Smart Pruning for Sustainable LLMs (2025)

    Publication: Scientific Reports (Nature)
    Smart pruning techniques maintain model quality while significantly reducing computational requirements.
  • A Survey on Model Compression for Large Language Models (2024)

    Publication: Transactions of the Association for Computational Linguistics (MIT Press)
    Source: MIT Press
    Comprehensive survey showing model compression (pruning, quantization, distillation) as essential for efficient LLM deployment.
  • CFSP: Efficient Structured Pruning for LLMs (2025)

    Publication: COLING 2025
    Source: ACL Anthology
    Structured pruning methods achieve 30% GPU reduction while maintaining performance.

5. Human-in-the-Loop Code Generation

Research validates that human-in-the-loop approaches with clear oversight mechanisms improve code generation quality and trust.

  • Human-in-the-Loop Machine Learning: A State of the Art (2022)

    Publication: Artificial Intelligence Review (Springer)
    Source: Springer
    Comprehensive state-of-the-art review establishing HITL as essential for reliable AI systems.
  • Conversational AI as a Coding Assistant (2024-2025)

    Publication: arXiv
    Study of how programmers interact with LLMs reveals they benefit most from conversational, iterative approaches with human guidance.
  • Rethinking AI Code Generation: A One-Shot Correction Approach Based on User Feedback (2024)

    Publication: Automated Software Engineering (Springer)
    Source: Springer
    User feedback-driven correction significantly improves code generation quality and trust.
  • State of AI Code Quality in 2025 (Qodo Research)

    Publication: Qodo AI Developer Survey
    Source: Qodo Research
    76% of developers fall into the "red zone" with low confidence in AI-generated code, highlighting critical need for trust mechanisms.

6. Context Engineering with Knowledge Graphs

Research demonstrates that knowledge graphs can enhance retrieval-augmented generation systems, enabling more structured and reliable context management.

  • Practices, Opportunities and Challenges in the Fusion of Knowledge Graphs and LLMs (2025)

    Publication: Frontiers in Computer Science
    Source: Frontiers
    Knowledge graphs assist in knowledge enhancement and resolve LLM hallucination problems through structured reasoning.
  • GraphRAG: Unlocking LLM Discovery on Narrative Private Data (2024)

    Publication: Microsoft Research
    LLM-generated knowledge graphs provide substantial improvements in Q&A performance for complex information analysis.

7. Continued Pre-Training for Domain Adaptation

Research shows that continued pre-training on domain-specific corpora can optimize foundation models for specific industry applications.

  • Domain-Adaptive Continued Pre-Training of Small Language Models (2025)

    Publication: arXiv
    Continued pre-training of small models offers promising path for domain adaptation with limited computational resources, showing gains in knowledge-intensive tasks (+8.1% MMLU) and contextual understanding (+7.6% HellaSwag).
  • Subset Selection for Domain Adaptive Pre-training (2025)

    Publication: Scientific Reports (Nature)
    Curriculum learning that gradually shifts training toward domain-specific concepts cuts compute by an order of magnitude.

8. Efficient Inference with Smaller Models

Research demonstrates that smaller, efficiently designed models can achieve faster inference and lower costs compared to larger foundation models.

  • A Survey on Efficient Inference for Large Language Models (2024)

    Publication: arXiv
    Model compression (distillation, quantization, pruning) proven effective for reducing redundancy and enabling faster inference.
  • Small Models, Big Tasks: SLMs for Function Calling (2025)

    Publication: arXiv
    Small language models (millions to billions of parameters) offer improved efficiency, accessibility, customizability, and faster inference for domain-specific settings.

Additional Research Areas

Additional papers covering advanced inference techniques, meta-agent frameworks, bias mitigation, and more.

  • Mercury by Inception Labs (2025)

    Publication: Inception Labs (First Commercial dLLM)
    Source: Inception Labs
    10X faster inference than autoregressive LLMs through parallel token generation using diffusion process.
  • Direct Semantic Communication Between LLMs (2025)

    Publication: Research on KV-Cache fusion for multi-agent systems
    8.5-10.5% higher accuracy, ~2x speedup vs text-based multi-agent communication. LLMs communicate via direct KV-Cache exchanges rather than generating/parsing text.

Complete Bibliography

This page represents a curated selection of the 60+ peer-reviewed papers that inform our technical foundation. For the complete bibliography and detailed analysis, please contact us at info@conductuslabs.com.

Return to Home