LoRa adaptors

Vibrant green numbers on a computer screen, showcasing binary code and data streams.

LoRa in the IOT community are the low power long range wireless standards devices that are utilized to send a signal over large distances. However, in the AI field LoRa implies Low rank adaptation – which is relatively more efficient compute way to fine tune pretrained models (LLM, vision transformers, diffusion models).

With a pretrained model is general and to improve its performance on a particular task a process called fine-tuning needs to be performed. Now, it is easier, to fine tune a model for a task rather than develop a brand new model from scratch. Not only easier but also computationally less intense especially since restart with all the data and resources are not required to make a new model for a new task.

Large models by definition are large and fine tuning on the new dataset is computation expensive, needs data and is harder to share. To accommodate fine tuning other methods have been tried – such as bottleneck adaptors, prefix/prompt tuning which involve modifying all model weights which take resources and can lead to issues like forgetting wherein the model loses the capability that existed before in the model.

LoRa (low rank adaptation) focuses on a select set of parameters to modify. This is called low-rank decomposition to adapt the large model but freeze the pretrained model weights intact. Low rank matrices called matrix A and matrix B. Matrix A is for the changes that are needed and matrix B helps in projecting matrix A changes back to the original parameter space. Thus, knowledge in the core model is unchanged but now the new model can adapt to new tasks.

These matrices are placed in attention layers in transformers or in cross-attention and self attention in diffusion models.

The workflow is simple – freeze the base model → inject LoRa modules into specific layer, train LoRa parameters on yoru dataset. Then you deploy either as merged with base model or as separate modules for multi-task adaptability.

As example, it can be used for modifying a LLM for domain specific knowledge, transfer a new style in image generation.

Similar Posts

  • | |

    Functional prediction of microbial sequences

    Even with E.coli and M.tuberculosis we only know ½ of them. Can you use ML model to define function: as natural language, or molecular interaction or chemical reaction. Function as molecular interaction: protein – protein interaction. Genomics: Learn association between genes (just like words). It is called gLM2. A multi-modal single residue resolution gLM. GLM2…

  • Language of Graphics

    Key IDEA: Chart types are great shortcuts but graphics are really sophisticated visual structures GoFish language for python using the Gestalt grouping principle. It can make graphics or chart that can display the messages that are intentional. .mark() – what shape to draw at each postion and what channels to encode part to whole relationships…

  • Detection of AI created content

    As AI has progressed generating code, writing, music or other intelligent language based skills through an LLM there is a parallel growth in detection of AI generated content.  Like everything AI, it is a probability game. Trying to estimate the combination of words/tokens and comparing with what is in the model. Using a standard model…

  • Error codes

    There is no error code in the answers that are provided by AI prompts. It will return an answer that is the best fit to the prompt or the question, but it does not tell you the probability that it is not correct or that it is low probability of answer. The conversational AI will…

  • |

    Open source protein models

    A company called Profluent (profluent.bio) has been developing protein models that can be used for designing new proteins (https://www.nature.com/articles/s41587-022-01618-2), modeling of new CRISPR-Cas sequences (https://www.nature.com/articles/s41586-025-09298-z) and developing LLM for protein generation (https://www.biorxiv.org/content/10.1101/2025.11.12.688125v1.article-info). What is amazing is that they have open sourced all their models and Profluent-E1 is available in GitHub to download and use. (https://github.com/Profluent-AI/E1)…