<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>http://stablediffusionwiki.com/index.php?action=history&amp;feed=atom&amp;title=LoRA</id>
	<title>LoRA - Revision history</title>
	<link rel="self" type="application/atom+xml" href="http://stablediffusionwiki.com/index.php?action=history&amp;feed=atom&amp;title=LoRA"/>
	<link rel="alternate" type="text/html" href="http://stablediffusionwiki.com/index.php?title=LoRA&amp;action=history"/>
	<updated>2026-05-15T13:32:48Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.39.4</generator>
	<entry>
		<id>http://stablediffusionwiki.com/index.php?title=LoRA&amp;diff=169&amp;oldid=prev</id>
		<title>StableTiger3 at 18:23, 22 August 2023</title>
		<link rel="alternate" type="text/html" href="http://stablediffusionwiki.com/index.php?title=LoRA&amp;diff=169&amp;oldid=prev"/>
		<updated>2023-08-22T18:23:19Z</updated>

		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;table style=&quot;background-color: #fff; color: #202122;&quot; data-mw=&quot;interface&quot;&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;col class=&quot;diff-marker&quot; /&gt;
				&lt;col class=&quot;diff-content&quot; /&gt;
				&lt;tr class=&quot;diff-title&quot; lang=&quot;en&quot;&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;← Older revision&lt;/td&gt;
				&lt;td colspan=&quot;2&quot; style=&quot;background-color: #fff; color: #202122; text-align: center;&quot;&gt;Revision as of 14:23, 22 August 2023&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot; id=&quot;mw-diff-left-l1&quot;&gt;Line 1:&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;Line 1:&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Overview ==&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Overview ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;−&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #ffe49c; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Natural language processing includes large-scale pretraining and adaptation to specific tasks or domains. Full fine-tuning becomes impractical with large models.&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Natural language processing includes large-scale pretraining and adaptation to specific tasks or domains. Full fine-tuning becomes impractical with large models&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;.  So Low Rank Adaptation (LoRA) is a solution to this.  In the world of natural language processing, many applications depend on using a large, pre-trained language model and then adapting it for different specific tasks. Typically, this is done through fine-tuning, which changes all the parameters of the original model. This method becomes a problem with very large models, like GPT-3, as it keeps the same vast number of parameters, making deployment challenging.&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-deleted&quot;&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt; &lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-deleted&quot;&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;Some have tried to solve this by only changing a few parameters or adding external modules for new tasks. This makes the model more efficient but can slow it down or even reduce its effectiveness. These solutions often don&amp;#039;t perform as well as fine-tuning, so there&amp;#039;s a compromise between efficiency and quality.&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-deleted&quot;&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt; &lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-side-deleted&quot;&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot; data-marker=&quot;+&quot;&gt;&lt;/td&gt;&lt;td style=&quot;color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;The Low-Rank Adaptation (LoRA) approach proposes a new way to tackle this issue. Inspired by the idea that changes in weights during adaptation have a low &amp;quot;intrinsic rank,&amp;quot; LoRA optimizes specific parts of the neural network, keeping the pre-trained weights untouched. This method makes LoRA efficient in both storage and computation, even with extremely large models like GPT-3. It provides a promising balance between maintaining model quality and enhancing operational efficiency&lt;/ins&gt;.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;br/&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Definitions ==&lt;/div&gt;&lt;/td&gt;&lt;td class=&quot;diff-marker&quot;&gt;&lt;/td&gt;&lt;td style=&quot;background-color: #f8f9fa; color: #202122; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #eaecf0; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;== Definitions ==&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>StableTiger3</name></author>
	</entry>
	<entry>
		<id>http://stablediffusionwiki.com/index.php?title=LoRA&amp;diff=168&amp;oldid=prev</id>
		<title>StableTiger3: Created page with &quot;== Overview == Natural language processing includes large-scale pretraining and adaptation to specific tasks or domains. Full fine-tuning becomes impractical with large models.  == Definitions == === Fine-Tuning === Refers to slight adjustments to a pre-trained model for specific tasks. Less feasible for larger models due to high cost and parameter count.  === Low-Rank Adaptation (LoRA) === LoRA retains pre-trained weights and incorporates trainable rank decomposition ma...&quot;</title>
		<link rel="alternate" type="text/html" href="http://stablediffusionwiki.com/index.php?title=LoRA&amp;diff=168&amp;oldid=prev"/>
		<updated>2023-08-22T18:11:34Z</updated>

		<summary type="html">&lt;p&gt;Created page with &amp;quot;== Overview == Natural language processing includes large-scale pretraining and adaptation to specific tasks or domains. Full fine-tuning becomes impractical with large models.  == Definitions == === Fine-Tuning === Refers to slight adjustments to a pre-trained model for specific tasks. Less feasible for larger models due to high cost and parameter count.  === Low-Rank Adaptation (LoRA) === LoRA retains pre-trained weights and incorporates trainable rank decomposition ma...&amp;quot;&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;== Overview ==&lt;br /&gt;
Natural language processing includes large-scale pretraining and adaptation to specific tasks or domains. Full fine-tuning becomes impractical with large models.&lt;br /&gt;
&lt;br /&gt;
== Definitions ==&lt;br /&gt;
=== Fine-Tuning ===&lt;br /&gt;
Refers to slight adjustments to a pre-trained model for specific tasks. Less feasible for larger models due to high cost and parameter count.&lt;br /&gt;
&lt;br /&gt;
=== Low-Rank Adaptation (LoRA) ===&lt;br /&gt;
LoRA retains pre-trained weights and incorporates trainable rank decomposition matrices in the Transformer architecture, drastically reducing trainable parameters and GPU memory cost.&lt;br /&gt;
&lt;br /&gt;
=== Trainable Parameters ===&lt;br /&gt;
These are adjustable aspects of the model during training. LoRA reduces these by 10,000 times, enhancing efficiency.&lt;br /&gt;
&lt;br /&gt;
=== Model Quality ===&lt;br /&gt;
The accuracy or performance of a model. LoRA performs on par or better than traditional fine-tuning, even with fewer parameters.&lt;br /&gt;
&lt;br /&gt;
=== Rank-Decomposition Matrices ===&lt;br /&gt;
Used in LoRA to reduce complexity without compromising quality or adding additional inference latency.&lt;br /&gt;
&lt;br /&gt;
=== Inference Latency ===&lt;br /&gt;
Time taken for the model to respond. LoRA does not increase this latency.&lt;br /&gt;
&lt;br /&gt;
== Benefits of LoRA ==&lt;br /&gt;
* Significant reduction in trainable parameters and GPU memory requirements.&lt;br /&gt;
* Comparable or superior performance to fine-tuning on models like RoBERTa, DeBERTa, GPT-2, and GPT-3.&lt;br /&gt;
* Higher training throughput, no additional inference latency.&lt;br /&gt;
&lt;br /&gt;
== Empirical Investigation ==&lt;br /&gt;
Study into rank-deficiency in language model adaptation gives insights into LoRA&amp;#039;s efficacy.&lt;br /&gt;
&lt;br /&gt;
== Availability ==&lt;br /&gt;
Package released for integration with PyTorch, including implementations and checkpoints for RoBERTa, DeBERTa, and GPT-2.&lt;/div&gt;</summary>
		<author><name>StableTiger3</name></author>
	</entry>
</feed>