Skip to content

MergedLinear still available? #852

@DanielRoeder1

Description

@DanielRoeder1

Feature request

The original lora library as well as some issues of peft show the existence of a MergedLinear layer for models like GPT2 that used a single Linear layer for qkv projection.
The feature seemed to be available some time ago but I can't seem to figure out how to enable it / if it is still available. Initializing a GPT2 model with a LoraConfig now appears to create a normal Lora Lienar layer that does not separate the projections.

Please advise if this feature is still available.

Motivation

MergedLinear is used for Decoder with a shared qkv projection layer.

Your contribution

(This feature was available some time ago, so I am uncertain about the current state=

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions