Here you may find a list of models that are currently available for use in your Generative AI Assistants.
model name | container name | model link | open-source? | size (billion parameters) | GPU usage | max tokens (prompt + response) | description |
---|---|---|---|---|---|---|---|
BLOOMZ 7B | transformers-lm-bloomz7b | link | yes | 7.1B | 33GB | 2,048 tokens | An open-source multilingual instruction-based large language model (46 languages). NB: free of charge. This model is up and running on our servers and can be used for free. |
GPT-J 6B | transformers-lm-gptj | link | yes | 6B | 25GB | 2,048 tokens | An open-source English-only large language model which is NOT fine-tuned for instruction following and NOT capable of code generation. NB: free of charge. This model is up and running on our servers and can be used for free. |
GPT-3.5 | openai-api-davinci3 | link | no (paid access via API) | supposedly, 175B | - (cannot be run locally) | 4,097 tokens | A multilingual instruction-based large language model which is capable of code generation. Unlike ChatGPT, not optimised for chat. NB: paid. You must provide your OpenAI API key to use the model. Your OpenAI account will be charged according to your usage. |
ChatGPT | openai-api-chatgpt | link | no (paid access via API) | supposedly, 175B | - (cannot be run locally) | 4,096 tokens | Based on gpt-3.5-turbo – the most capable of the entire GPT-3/GPT-3.5 models family. Optimized for chat. Able to understand and generate code. NB: paid. You must provide your OpenAI API key to use the model. Your OpenAI account will be charged according to your usage. |
Open-Assistant SFT-1 12B | transformers-lm-oasst12b | link | yes | 12B | 26GB (half-precision) | 5,120 tokens | An open-source English-only instruction-based large language model which is NOT good at answering math and coding questions. NB: free of charge. This model is up and running on our servers and can be used for free. |
Vicuna 13B | transformers-lm-vicuna13b | link | yes, but only for non-commercial use | 13B | 29GB (half-precision) | 2,048 tokens | An instruction-based large language model fine-tuned on LLaMa that achieves more than 90%* quality of OpenAI ChatGPT and Google Bard. The model performs best in English and is NOT good at answering math, reasoning, and coding questions. NB-1: Free of charge. This model is up and running on our servers and can be used for free. NB-2: cannot be used for commercial purposes (license restriction). |
GPT-4 | openai-api-gpt4 | link | no (paid access via API) | supposedly, 175B | - (cannot be run locally) | 8,192 tokens | A multilingual instruction-based large language model which is capable of code generation and other complex tasks. More capable than any GPT-3.5 model, able to do more complex tasks, and optimized for chat. NB: paid. You must provide your OpenAI API key to use the model. Your OpenAI account will be charged according to your usage. |
GPT-4 32K | openai-api-gpt4-32k | link | no (paid access via API) | supposedly, 175B | - (cannot be run locally) | 32,768 tokens | A multilingual instruction-based large language model which is capable of code generation and other complex tasks. Same capabilities as the base gpt-4 mode but with 4x the context length. NB: paid. You must provide your OpenAI API key to use the model. Your OpenAI account will be charged according to your usage. |
GPT-JT 6B | transformers-lm-gptjt | link | yes | 6B | 26GB | 2,048 tokens | An open-source English-only large language model which was fine-tuned for instruction following but is NOT capable of code generation. NB: free of charge. This model is up and running on our servers and can be used for free. |
ChatGPT 16k | openai-api-chatgpt-16k | link | no (paid access via API) | supposedly, 175B | - (cannot be run locally) | 16,384 tokens | Same capabilities as the standard gpt-3.5-turbo model but with 4 times the context. NB: paid. You must provide your OpenAI API key to use the model. Your OpenAI account will be charged according to your usage. |
Anthropic Claude-v1 | anthropic-api-claude-v1 | link | no (paid access via API) | - (cannot be run locally) | 9,000 tokens | The largest model, ideal for a wide range of more complex tasks. NB: paid. You must provide your Anthropic API key to use the model. Your Anthropic API account will be charged according to your usage. | |
Anthropic Claude Instant v1 | anthropic-api-claude-instant-v1 | link | no (paid access via API) | - (cannot be run locally) | 9,000 tokens | A smaller model with far lower latency, sampling at roughly 40 words/sec! Its output quality is somewhat lower than the latest claude-1 model, particularly for complex tasks. However, it is much less expensive and blazing fast. NB: paid. You must provide your Anthropic API key to use the model. Your Anthropic API account will be charged according to your usage. |