by taikon on 10/7/24, 2:05 PM with 29 comments
by vessenes on 10/7/24, 4:45 PM
For what it's worth, the writing is .. bland. In the way that only an LLMs writing can be -- relatively grammatically sound, and totally soulless. I will never think of the love story of Elizabeth and Thomas again, despite having read the entire thing.
In early days of GPT-3, I experimented a lot with getting it respond as certain authors, and it was really quite excellent at that. This is one of the many things that seem likely to have been nerfed over time, I'd guess partly because human preference training just asks for bland responses, and partly because the injected prompts from OpenAI strongly discourage doing things related to real people, and those preferences are carried through, subtlely or not, into the augmented training data most open models tune on.
by mmaunder on 10/7/24, 5:11 PM
by ed on 10/7/24, 4:17 PM
The model is stock llama, fine tuned with a set of long documents to encourage longer outputs.
Most of the action seems to happen in an agent.
by danng87 on 10/7/24, 3:35 PM
Does anyone know how LongWriter handles maintaining coherence and structure in longer outputs? Also, are there specific strategies or parameters recommended for fine-tuning LLaMA 3.1 with this setup to maximize the quality of generated text?
by 8bitsrule on 10/8/24, 1:22 AM
by hshshshsvsv on 10/8/24, 2:22 PM
by wkat4242 on 10/7/24, 5:46 PM
by jw12 on 10/8/24, 1:09 PM
by fitsumbelay on 10/8/24, 12:30 AM
by alwinaugustin on 10/7/24, 3:26 PM