Skip to content

LORA Continuos pre-training on 7B Instruct Model #3509

@sinchanabhat

Description

@sinchanabhat

Hi,

  1. I was wondering if it is possible (or even makes sense) to try continuous pre-training using LORA for a 7b instruct model? I only have huge raw text, my hypothesis is if I train on LORA weights using the raw text, we could achieve domain adaptation to some level without losing the instruct behaviour.
  2. Can I use the raw text as is in SFT Trainer and achieve this or should I use a larger LLM to convert these raw text into instructions?

Any suggestions would be really helpful.

Thanks

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions