from Hacker News

Ask HN: How should I burn $8k for Google Gemini 1.5 Pro?

by hoerzu on 6/10/24, 10:31 PM with 10 comments

Was looking into applying LLMs onto entity extraction in earnings calls. (Around 400mb of text data). 1 Million tokes is around 1-3$.

Open for ideas and experiments.

  • by reneberlin on 6/11/24, 12:08 AM

    Just be careful with the "temperature" when dealing with numbers and financial data as well. You should check samples of the retrieved data by hand.

    Setting the temperature lower than 0.9 means reducing the "creativity" and making it less prone to hallucinate.

  • by reneberlin on 6/11/24, 12:02 AM

    That's not how you do it. You use a Vectordatabase and a Retriever. This way, not all tokens of a document are used with the prompt, just the relevant parts will end up in the conversation / prompts. This way you save a lot of money and you are NOT limited to use GoogleAI - you can use whatever AI you want.
  • by reneberlin on 6/11/24, 12:06 AM

    Use Flowise and a VectorDB like upstash or pinecone.

    Flowise Tutorial https://www.youtube.com/watch?v=V7uBy3VQJAc

  • by langcss on 6/11/24, 5:01 AM

    Can’t you pre-filter it with a vector lookup or something cheap first (gpt3.5?). Or even simply filter first by other things you are considering for the trading decision (price to earnings for example). That might save you a lot of cost.
  • by BOOSTERHIDROGEN on 6/11/24, 7:22 AM

    Could someone from hedge funds kindly share their current technology stack if they utilize LLMs?
  • by dissahc on 6/11/24, 2:57 AM

    are you sure you need a powerful model for this? llama3-8b is at least 10 times cheaper and might suffice for something like this
  • by infecto on 6/11/24, 1:08 AM

    What’s the goal?