ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.22352
47
0

Meta-LoRA: Meta-Learning LoRA Components for Domain-Aware ID Personalization

28 March 2025
Barış Batuhan Topal
Umut Özyurt
Zafer Doğan Budak
Ramazan Gokberk Cinbis
ArXivPDFHTML
Abstract

Recent advancements in text-to-image generative models, particularly latent diffusion models (LDMs), have demonstrated remarkable capabilities in synthesizing high-quality images from textual prompts. However, achieving identity personalization-ensuring that a model consistently generates subject-specific outputs from limited reference images-remains a fundamental challenge. To address this, we introduce Meta-Low-Rank Adaptation (Meta-LoRA), a novel framework that leverages meta-learning to encode domain-specific priors into LoRA-based identity personalization. Our method introduces a structured three-layer LoRA architecture that separates identity-agnostic knowledge from identity-specific adaptation. In the first stage, the LoRA Meta-Down layers are meta-trained across multiple subjects, learning a shared manifold that captures general identity-related features. In the second stage, only the LoRA-Mid and LoRA-Up layers are optimized to specialize on a given subject, significantly reducing adaptation time while improving identity fidelity. To evaluate our approach, we introduce Meta-PHD, a new benchmark dataset for identity personalization, and compare Meta-LoRA against state-of-the-art methods. Our results demonstrate that Meta-LoRA achieves superior identity retention, computational efficiency, and adaptability across diverse identity conditions. Our code, model weights, and dataset are released onthis http URL.

View on arXiv
@article{topal2025_2503.22352,
  title={ Meta-LoRA: Meta-Learning LoRA Components for Domain-Aware ID Personalization },
  author={ Barış Batuhan Topal and Umut Özyurt and Zafer Doğan Budak and Ramazan Gokberk Cinbis },
  journal={arXiv preprint arXiv:2503.22352},
  year={ 2025 }
}
Comments on this paper