pull down to refresh

I think you may be misunderstanding the pricing? The default invoice of 25 cents pays for many queries, not just one. The margin is actually very low currently compared to API.
Was this possibly what happened? I think maybe we need to make the UX much more clear on what 25 cents gets you because some other users have also voiced this. Let me know please.
No I understand the pricing very well. Your margins obviously depend a lot on the underlying model and the prompt.
Opus, GPT-4 etc are on the higher end and I can maybe see it being an OK margin there.
But you can find llama3-70b APIs out there for like $0.20/million tokens.
reply
We use openrouter as our supplier for Llama3-70B and their price is $0.27/M and we tack on a margin after that and round up to the nearest sat. Ultimately though llama3 queries our platform rarely exceed a few sats (2/10ths of a penny) so they aren't exactly breaking the bank lol.
I guess if you are plugging in hundreds of thousands of tokens of context it starts to matter but for 99% of normal users this is incredibly cheap for value you are getting from AI.
reply
Ah, I think I figured out the confusion. You saw the "25 cents pays for 8-10 queries" and thought that applied to Meta Llama 3. The 8-10 is in reference to GPT4 Turbo. You will get like over 100 queries with Meta Llama 3.
Yea we need to do better explaining these things to people.
reply
100 sats \ 2 replies \ @kevin 30 Apr
Ok, then it makes a lot more sense. I did select llama3 from the dropdown and still got $0.25. If it's different then that should be reflected in the UI.
reply
If you select Llama from the dropdown and run some queries, you will see in the "Account Activity" section the actual price you paid. Should be 1-3 sats usually. The 25 cents is just a one time deposit to buy a bunch of credits. After that payment you are then drawing down upon that 25 cents over time. You can set it to 5 cents too if you want and that should still buy you quite a few Llama queries.
reply
I understand your flow now. You came to website and changed from the default model of gpt4 turbo to llama before you submitted your first query. Then the payment modal came up with the 8-10 queries sentence which seemed expensive to you.
We will definitely be revamping the initial payment modal because it is very confusing to a lot of people.
Thanks for working through this with me.