DeepInfra(@DeepInfra) 's Twitter Profileg
DeepInfra

@DeepInfra

Fast ML inference. Run top AI models using a simple API.

ID:1623086169759318017

linkhttps://deepinfra.com calendar_today07-02-2023 22:28:09

157 Tweets

1,8K Followers

39 Following

DeepInfra(@DeepInfra) 's Twitter Profile Photo

At lunch today we were talking how it would make sense if
7B models were 7 cents,
13B models were 13 cents,
and of course 8B models were 8 cents (per 1M tokens).
It just makes sense, so we set our prices!
deepinfra.com/pricing

account_circle
DeepInfra(@DeepInfra) 's Twitter Profile Photo

Also we just dropped pricing on most of our 7b, 13b and 70b models to $0.10, $0.18, $0.64 per million input tokens. We will always have the best prices. deepinfra.com/pricing

account_circle
DeepInfra(@DeepInfra) 's Twitter Profile Photo

Official Mixtral-8x22b-Instruct model just got released and is now on DeepInfra. This is the best open LLM and we are hosting at the best price of $0.65 / 1M tokens. deepinfra.com/mistralai/Mixt…

account_circle
DeepInfra(@DeepInfra) 's Twitter Profile Photo

Just added the first instruction finetunned version of mixtral 8x22B deepinfra.com/HuggingFaceH4/… still not official from mixtral but not bad

account_circle
DeepInfra(@DeepInfra) 's Twitter Profile Photo

We set the bar when we launched the first mixtral, and we're going to do it again! The new mixtral, with 65K context, at ... 65c / 1 million tokens! This new model is almost 3 times larger. deepinfra.com/mistralai/Mixt…

account_circle
DeepInfra(@DeepInfra) 's Twitter Profile Photo

databricks/dbrx-instruct is now on @deepinfra. $0.6 per 1M tokens - the best price of all providers. Also up to 130 tps. Try it our here: deepinfra.com/databricks/dbr…

account_circle
DeepInfra(@DeepInfra) 's Twitter Profile Photo

You can now host your custom LLMs at DeepInfra. It's managed LLM hosing service. Pay-per GPU/h $2/A100, $4/H100. It's super simple. Read more here deepinfra.com/blog/custom-ll…

account_circle
DeepInfra(@DeepInfra) 's Twitter Profile Photo

Checkout the new Mistral 0.2 model. We are running it in fp8 on H100s and it does 160tps. deepinfra.com/mistralai/Mist…

account_circle
Eric Zelikman ✈️ ICLR(@ericzelikman) 's Twitter Profile Photo

A couple exciting updates! First, we quantitatively evaluated the improvement from combining Quiet-STaR with chain-of-thought (i.e. letting the model think before each CoT token). We found it improves zero-shot CoT accuracy on GSM8K by over 7%!

A couple exciting updates! First, we quantitatively evaluated the improvement from combining Quiet-STaR with chain-of-thought (i.e. letting the model think before each CoT token). We found it improves zero-shot CoT accuracy on GSM8K by over 7%!
account_circle
DeepInfra(@DeepInfra) 's Twitter Profile Photo

Also we just shipped the highly requested multimodal Llava1.5. Give it a try deepinfra.com/llava-hf/llava…

account_circle
DeepInfra(@DeepInfra) 's Twitter Profile Photo

Guided JSON response is now available DeepInfra API. Read more about it here
deepinfra.com/blog/json-mode. Restricting the output to JSON had almost no performance penalty and is FREE.

account_circle
DeepInfra(@DeepInfra) 's Twitter Profile Photo

We’re committed to building AI that improves lives & unlocks a better future for humanity. We are proud to sign the Open Letter: Build AI for a Better Future along with @SVAngel, OpenAI, Meta, Google, Microsoft & many others. openletter.svangel.com

account_circle