| # DeepInfra | |
| >[DeepInfra](https://deepinfra.com/docs) allows us to run the | |
| > [latest machine learning models](https://deepinfra.com/models) with ease. | |
| > DeepInfra takes care of all the heavy lifting related to running, scaling and monitoring | |
| > the models. Users can focus on your application and integrate the models with simple REST API calls. | |
| >DeepInfra provides [examples](https://deepinfra.com/docs/advanced/langchain) of integration with LangChain. | |
| This page covers how to use the `DeepInfra` ecosystem within `LangChain`. | |
| It is broken into two parts: installation and setup, and then references to specific DeepInfra wrappers. | |
| ## Installation and Setup | |
| - Get your DeepInfra api key from this link [here](https://deepinfra.com/). | |
| - Get an DeepInfra api key and set it as an environment variable (`DEEPINFRA_API_TOKEN`) | |
| ## Available Models | |
| DeepInfra provides a range of Open Source LLMs ready for deployment. | |
| You can see supported models for | |
| [text-generation](https://deepinfra.com/models?type=text-generation) and | |
| [embeddings](https://deepinfra.com/models?type=embeddings). | |
| You can view a [list of request and response parameters](https://deepinfra.com/meta-llama/Llama-2-70b-chat-hf/api). | |
| Chat models [follow openai api](https://deepinfra.com/meta-llama/Llama-2-70b-chat-hf/api?example=openai-http) | |
| ## LLM | |
| See a [usage example](/docs/integrations/llms/deepinfra). | |
| ```python | |
| from langchain_community.llms import DeepInfra | |
| ``` | |
| ## Embeddings | |
| See a [usage example](/docs/integrations/text_embedding/deepinfra). | |
| ```python | |
| from langchain_community.embeddings import DeepInfraEmbeddings | |
| ``` | |
| ## Chat Models | |
| See a [usage example](/docs/integrations/chat/deepinfra). | |
| ```python | |
| from langchain_community.chat_models import ChatDeepInfra | |
| ``` | |