NEW WEBSITE LAUNCH
Subscribe to our newsletter

Compare Models

  • Google

    FLAN-T5

    FREE
    If you already know T5, FLAN-T5 is just better at everything. For the same number of parameters, these models have been fine-tuned on more than 1,000 additional tasks covering more languages – the NLP is for English, German, French. It has Apache-2.0 license which is a permissive open source license that allows for commercial use. With appropriate prompting, it can perform zero-shot NLP tasks such as text summarization, common sense reasoning, natural language inference, question answering, sentence and sentiment classification, translation, and pronoun resolution.
  • Google

    Flan-UL2

    FREE
    Developed by Google, Flan-UL2, which is a more powerful version of the T5 model that has been trained using Flan, and it is downloadable from Hugging Face. It shows performance exceeding the ‘prior’ versions of Flan-T5. With the ability to reason for itself and generalize better than the previous models, Flan-UL2 is a great improvement. Flan-UL2 is a machine learning model that can generate textual descriptions of images and has the potential to be used for image search, video captioning, automated content generation, and visual question answering. Flan-UL2 has an Apache-2.0 license, which is a permissive open source license that allows for commercial use.
    If Flan-UL2’s 20B parameters are too much, consider the previous iteration of Flan-T5, which comes in five different sizes and might be more suitable for your needs.
  • Cohere

    Generate

    $0.015
    Cohere is a Canadian startup that provides high-performance and secure LLMs for the enterprise. Their models work on public, private, or hybrid clouds.
    Cohere Generate can be used for tasks such as copywriting, named entity recognition, paraphrasing, and summarization. It can be particularly useful for automating time-consuming and repetitive copywriting tasks and re-wording text to suit a specific reader or context.
    Cohere Generate is available as an API that can be integrated into various libraries using Python, Node, or Go software development kits (SDKs).
    We have shown the price of the Cohere Generate Default version, but a Cohere Generate Custom model is available but is double the price (0.030 per 1/k tokens). However, custom models can lead to some of the best-performing NLP models for many tasks.
  • OpenAI

    GPT-3.5-turbo 16k

    $0.004
    GPT-3.5-turbo 16k has the same capabilities as the standard gpt-3.5-turbo (4k model) but with 4 times the context but at twice the price. In general, a larger context window can be more powerful because it takes into account more information from the surrounding text, which can lead to better predictions
    GPT-3.5-turbo was designed to provide better performance and is well-known as the model that, by default, powers ChatGPT. However, paying customers who subscribe to ChatGPT Plus can change the model to GPT-4 before you start a chat.
    GPT-3.5-turbo is optimized for conversational formats and is superior to GPT-3 models, and the performance of GPT-3.5-turbo is on par with Instruct Davinci-003. GPT-3.5-turbo was trained on a massive corpus of text data, including books, articles, and web pages from across the internet and is used for tasks like content and code generation, question answering, translation, and more. Access is available through a request to OpenAI’s API or through the web application (try for free).
  • OpenAI

    GPT-3.5-turbo 4k

    $0.002
    GPT-3.5-turbo is an upgraded version of the GPT-3 model. It was designed to provide better performance and is well-known as the model that, by default, powers ChatGPT (however, paying customer who subscribe to ChatGPT Plus can change the model to GPT-4 before you start a chat).
    GPT-3.5-turbo is optimized for conversational formats and is superior to GPT-3 models, and the performance of GPT-3.5-turbo is on par with Instruct Davinci-003 (however is also ten times cheaper and has been seen to be three times faster). GPT-3.5-turbo was trained on a massive corpus of text data, including books, articles, and web pages from across the internet and is used for tasks like content and code generation, question answering, translation, and more. In some cases, GPT-3.5-turbo results can sometimes be too “chatty” or “creative”. Access is available through a request to OpenAI’s API or through the web application (try for free).

  • OpenAI

    GPT-4 32K context

    $0.12

    GPT-4 is OpenAI’s new design that incorporates additional improvements and advancements, including being multimodal so it can take both text and image inputs. With broad general knowledge and domain expertise, GPT-4 can follow complex instructions in natural language and solve difficult problems with accuracy. GPT-4 has a more diverse range of training data, incorporating additional languages and sources beyond just English. This means that the model will be able to process and generate text in multiple languages and better understand the nuances and subtleties of different languages and dialects. This is the extended 32k token context-length model, which is separate to the 8k model (and is more expensive).

    GPT-4 API access is now available.

     

    Note: At the time of writing, ChatGPT Plus subscribers can access Chat GPT-4 by logging into the web application.

  • OpenAI

    GPT-4 8K context

    $0.06

    GPT-4 is OpenAI’s new design that incorporates additional improvements and advancements, including being multimodal so it can take both text and image inputs. With broad general knowledge and domain expertise, GPT-4 can follow complex instructions in natural language and solve difficult problems with accuracy. GPT-4 has a more diverse range of training data, incorporating additional languages and sources beyond just English. This means that the model will be able to process and generate text in multiple languages and better understand the nuances and subtleties of different languages and dialects. There are a few different GPT-4 models to choose from. The standard GPT-4 model offers 8k tokens for the context. GPT-4 API access is now available.

    Note: For the ChatGPT web application, ChatGPT is powered by GPT-3.5 turbo by default. However, if you are a paying customer and subscribe to ChatGPT Plus, you can change the model to GPT-4 before you start a chat.

  • Google

    LaMDA

    OTHER
    LaMDA stands for Language Model for Dialogue Application. It is a conversational Large Language Model (LLM) built by Google as an underlying technology to power dialogue-based applications that can generate natural-sounding human language. LaMDA is built by fine-tuning a family of Transformer-based neural language models specialized for dialog and teaching the models to leverage external knowledge sources. The potential use cases for LaMDA are diverse, ranging from customer service and chatbots to personal assistants and beyond. LaMDA is not open source; currently, there are no APIs or downloads. However, Google is working on making LaMDA more accessible to researchers and developers. In the future, it is likely that LaMDA will be released as an open source project, and that APIs and downloads will be made available.
  • Google

    PaLM 2 chat-bison-001

    $0.0021535
    PaLM 2 has just launched (May 2023) and is Google’s next-generation Large Language Model, built on Google’s Pathways AI architecture. PaLM 2 was trained on a massive dataset of text and code, and it can handle many different tasks and learn new ones quickly. It is seen as a direct competitor to OpenAI’s GPT-4 model. It excels at advanced reasoning tasks, including code and math, classification and question answering, translation and multilingual proficiency (100 languages), and natural language generation better than our previous state-of-the-art LLMs, including its predecessor PaLM.
    PaLM 2 is the underlying model driving the PaLM API that can be accessed through Google’s Generative AI Studio. PaLM 2 has four submodels with different sizes. Bison is the best value in terms of capability and chat-bison-001 has been fine-tuned for multi-turn conversation use cases. If you want to see PaLM 2 capabilities, the simplest way to use it is through Google Bard (PaLM 2 is the technology that powers Google Bard).

     

    Watch Paige Bailey introducing PaLM 2: view here

  • ChatGLM

    PaLM 2 text-bison-001

    $0.004
    PaLM 2 has just launched (May 2023) and is Google’s next-generation Large Language Model, built on Google’s Pathways AI architecture. PaLM 2 was trained on a massive dataset of text and code, and it can handle many different tasks and learn new ones quickly. It is seen as a direct competitor to OpenAI’s GPT-4 model. It excels at advanced reasoning tasks, including code and math, classification, question answering, translation and multilingual proficiency (100 languages), and natural language generation better than our previous state-of-the-art LLMs, including its predecessor PaLM.

     

    PaLM 2 is the underlying model driving the PaLM API that can be accessed through Google’s Generative AI Studio. PaLM 2 has four submodels with different sizes. Bison is the best value in terms of capability and cost, and text-bison-001 can be fine-tuned to follow natural language instructions and is suitable for various language tasks such as classification, sentiment analysis, entity extraction, extractive question answering, summarization, re-writing text in a different style, and concept ideation.

     

    If you want to see PaLM 2 capabilities, the simplest way to use it is through Google Bard (PaLM 2 is the technology that powers Google Bard).

     

    Watch Paige Bailey introducing PaLM 2: view here

  • Google

    PaLM 2 textembedding-gecko-001

    $0.0004
    PaLM 2 has just launched (May 2023) and is Google’s next-generation Large Language Model, built on Google’s Pathways AI architecture. PaLM 2 was trained on a massive dataset of text and code, and it can handle many different tasks and learn new ones quickly. It is seen as a direct competitor to OpenAI’s GPT-4 model. It excels at advanced reasoning tasks, including code and math, classification and question answering, translation and multilingual proficiency (100 languages), and natural language generation better than our previous state-of-the-art LLMs, including its predecessor PaLM.
    PaLM 2 is the underlying model driving the PaLM API that can be accessed through Google’s Generative AI Studio. PaLM 2 has four submodels with different sizes: Unicorn (the largest), Bison, Otter, and Gecko (the smallest) and the different sizes of the submodels allow PaLM 2 to be more efficient and to perform different tasks. Gecko is the smallest and cheapest model for simple tasks and textembedding-gecko-001 returns model embeddings for text inputs.
    If you want to see PaLM 2 capabilities, the simplest way to use it is through Google Bard (PaLM 2 is the technology that powers Google Bard).

     

    Watch Paige Bailey introducing PaLM 2: view here

  • RedPajama

    RedPajama-INCITE-7B-Instruct

    FREE
    The RedPajama project aims to create a set of leading open source models. RedPajama-INCITE-7B-Instruct was developed by Together and leaders from the open source AI community. RedPajama-INCITE-7B-Instruct model represents the top-performing open source entry on the HELM benchmarks, surpassing other cutting-edge open models like LLaMA-7B, Falcon-7B, and MPT-7B. The instruct-tuned model is designed for versatility and shines when tasked with few-shot performance.

     

    The Instruct, Chat, Base Model, and ten interim checkpoints are now available on HuggingFace, and all the RedPajama LLMs come with commercial licenses under Apache 2.0.

     

    Play with the RedPajama chat model version here – https://lnkd.in/g3npSEbg
1 2 3 4

Ada (fine tuning) GPT-3
This website uses cookies to improve your experience. By using this website you agree to our Privacy Policy Policy.