Enterprise
AI models for your enterprise needs: A quick primer on making the right choice
Dec 19, 2024
5 min read
When I say “generative AI,” what’s the first thing that comes to mind? If you said ChatGPT, you’re not alone. One of the most common misconceptions about GenAI is that it’s one thing, a single model that can serve every need.
In reality, there are 255 large-scale models (those with training compute over 1023 floating point operations, essentially the scale of GPT-3 or higher). These can be categorized in dozens of ways, depending on what you need it for, how much you’re willing to spend, how secure you need it to be, etc.
In this blog post, we explore a few of these tangents.
Model scale
Large-scale vs. small-scale or single-purpose micromodels
Let’s begin at the beginning. The primary indicator of an AI model tends to be its scale. Here are models looked at through the lens of scale.
Foundation models
Foundation models tend to be the largest and are trained on enormous data sets. They are characterized by their ability to:
Perform a wide range of tasks, logical, creative, or problem-solving
Generate multimodal output across text, images, code, or video
Be fine-tuned for specific tasks
Examples: Open AI’s GPT, Google’s Gemini, Claude 3.5 Sonnet
Use case: The best use case of these models tends to be in generating content based on natural language input. Writing blog posts, documentation, text summaries, etc. are great for this type of model.
Smaller or single-purpose micromodels
Smaller models tend to be able to do specific tasks well. They are typically overtrained and overfit to perform a specific task. They are:
Trained on fewer than 10 billion parameters
Lightweight, balancing performance and efficiency
Great for on-device applications, like mobile assistants or smart appliances
Useful for domain-specific tasks
Examples: TinyLlama, Mistral Nemo
Use case: Small models are great for specific tasks like code generation, language translation, mobile applications, etc.
Domain-applicability
General purpose vs. domain-specific models
Many AI models are typically built to be general-purpose, as in, they are able to do a wide variety of tasks across use cases. For instance, with a foundation model, you might build a chatbot for customer service, sales, scenario modeling or recommendation engines.
On the other hand, organizations are realizing the need for increased accuracy in processing specific contexts and terminology related to each domain. This resulted in domain-specific models, which are:
Smaller, trained on niche data related to the industry
More accurate in interpreting domain-specific language
Personalizable with greater precision
Examples: Med-PaLM, BloombergGPT, ClimateBERT
Use case: In healthcare, domain-specific models are great for diagnosis. In law, it’s good for contract analysis, compliance review, etc.
Task complexity
General purpose vs. task-specific models
Another dimension similar to domain-specificity is task-specificity. Simply put, a model that is trained for code completion wouldn’t be able to generate videos as well. It’s specific to that particular task.
Task-specific models are characterized by:
Specialization is one particular task or closely related tasks
Excellent accuracy and reduced hallucination in performing that specific task(s)
Lower latency and faster processing
Higher interpretability
Examples: CodeLLama, Veo
Use cases: A great use case for task-specific LLM is in language translation. You can enable distributed teams to converse in different languages with real-time AI-based translation.
Device and deployment
Heavy-weight vs. lightweight models
When we speak of model deployment, we often think of scalable cloud or sprawling on-prem infrastructure. This is true of heavy-weight, powerful models. However, not all use cases lend themselves to having AI deployed this way.
Lightweight models are:
Efficient, using as little computing as possible
Usable in on-device, such as mobile phones or smart devices
Functional despite limited connectivity
Examples: Gemini Nano, Phi-2
Use cases: For low-latency use cases, lightweight models deployed on devices like smartphones are perfect.
Model licensing
Proprietary vs. open source models
In the server operating systems space, open source has over 90% market share. This is a trend likely to duplicate in the LLM space as well.
Proprietary models
A proprietary large language model is developed and maintained by a singular provider, who often charges a subscription fee to use it. Proprietary models often come with regular updates, higher reliability, better performance and support.
You can deploy it in your environment or integrate it through APIs. However, you will not have access to the source code, limiting customization. You might also have limitations on how you use the software.
Examples: Claude 2, GPT-4
Use cases: If you’re just integrating Generative AI into your workflows for a general productivity boost, proprietary models are easier. For instance, Microsoft Copilot is easy to get started if your enterprise is on the Microsoft platform.
Open-source models
In open-source models, the source code and architecture are available for the public to use or modify. These models are community-driven, depending on a wide range of users for development and maintenance. However, it is common for big organizations like Facebook, Google, and OpenAI to develop and maintain open-source LLMs as well.
Open-source models are generally free. The availability of source code helps teams customize, fine-tune, and adapt the code to their needs.
Examples: Meta’s LlaMA, Falcon 180B, TextBlob
Use cases: You can have generalized, domain-specific and task-specific models that are open-source. It’s best when you have the capacity to review, choose, fine-tune and implement open-source LLMs on your own.
With that foundation, here are some pointers on choosing the right GenAI model for your enterprise needs.
If your use case is generalized and performs a wide range of tasks, choose large language models
If your use case is specific and limited to a few similar tasks, choose domain-specific, task-specific or small language models
If you need greater flexibility and customizability, go open-source
If you’re cost-sensitive, choose smaller open-source models
If you don’t have the resources to implement models yourself, choose proprietary models with more straightforward integration and better support
If you’re having trouble making effective and sustainable decisions, speak to Tune AI’s experts. We’ve worked with all kinds of models for use cases across enterprise departments. We’re happy to help.
When I say “generative AI,” what’s the first thing that comes to mind? If you said ChatGPT, you’re not alone. One of the most common misconceptions about GenAI is that it’s one thing, a single model that can serve every need.
In reality, there are 255 large-scale models (those with training compute over 1023 floating point operations, essentially the scale of GPT-3 or higher). These can be categorized in dozens of ways, depending on what you need it for, how much you’re willing to spend, how secure you need it to be, etc.
In this blog post, we explore a few of these tangents.
Model scale
Large-scale vs. small-scale or single-purpose micromodels
Let’s begin at the beginning. The primary indicator of an AI model tends to be its scale. Here are models looked at through the lens of scale.
Foundation models
Foundation models tend to be the largest and are trained on enormous data sets. They are characterized by their ability to:
Perform a wide range of tasks, logical, creative, or problem-solving
Generate multimodal output across text, images, code, or video
Be fine-tuned for specific tasks
Examples: Open AI’s GPT, Google’s Gemini, Claude 3.5 Sonnet
Use case: The best use case of these models tends to be in generating content based on natural language input. Writing blog posts, documentation, text summaries, etc. are great for this type of model.
Smaller or single-purpose micromodels
Smaller models tend to be able to do specific tasks well. They are typically overtrained and overfit to perform a specific task. They are:
Trained on fewer than 10 billion parameters
Lightweight, balancing performance and efficiency
Great for on-device applications, like mobile assistants or smart appliances
Useful for domain-specific tasks
Examples: TinyLlama, Mistral Nemo
Use case: Small models are great for specific tasks like code generation, language translation, mobile applications, etc.
Domain-applicability
General purpose vs. domain-specific models
Many AI models are typically built to be general-purpose, as in, they are able to do a wide variety of tasks across use cases. For instance, with a foundation model, you might build a chatbot for customer service, sales, scenario modeling or recommendation engines.
On the other hand, organizations are realizing the need for increased accuracy in processing specific contexts and terminology related to each domain. This resulted in domain-specific models, which are:
Smaller, trained on niche data related to the industry
More accurate in interpreting domain-specific language
Personalizable with greater precision
Examples: Med-PaLM, BloombergGPT, ClimateBERT
Use case: In healthcare, domain-specific models are great for diagnosis. In law, it’s good for contract analysis, compliance review, etc.
Task complexity
General purpose vs. task-specific models
Another dimension similar to domain-specificity is task-specificity. Simply put, a model that is trained for code completion wouldn’t be able to generate videos as well. It’s specific to that particular task.
Task-specific models are characterized by:
Specialization is one particular task or closely related tasks
Excellent accuracy and reduced hallucination in performing that specific task(s)
Lower latency and faster processing
Higher interpretability
Examples: CodeLLama, Veo
Use cases: A great use case for task-specific LLM is in language translation. You can enable distributed teams to converse in different languages with real-time AI-based translation.
Device and deployment
Heavy-weight vs. lightweight models
When we speak of model deployment, we often think of scalable cloud or sprawling on-prem infrastructure. This is true of heavy-weight, powerful models. However, not all use cases lend themselves to having AI deployed this way.
Lightweight models are:
Efficient, using as little computing as possible
Usable in on-device, such as mobile phones or smart devices
Functional despite limited connectivity
Examples: Gemini Nano, Phi-2
Use cases: For low-latency use cases, lightweight models deployed on devices like smartphones are perfect.
Model licensing
Proprietary vs. open source models
In the server operating systems space, open source has over 90% market share. This is a trend likely to duplicate in the LLM space as well.
Proprietary models
A proprietary large language model is developed and maintained by a singular provider, who often charges a subscription fee to use it. Proprietary models often come with regular updates, higher reliability, better performance and support.
You can deploy it in your environment or integrate it through APIs. However, you will not have access to the source code, limiting customization. You might also have limitations on how you use the software.
Examples: Claude 2, GPT-4
Use cases: If you’re just integrating Generative AI into your workflows for a general productivity boost, proprietary models are easier. For instance, Microsoft Copilot is easy to get started if your enterprise is on the Microsoft platform.
Open-source models
In open-source models, the source code and architecture are available for the public to use or modify. These models are community-driven, depending on a wide range of users for development and maintenance. However, it is common for big organizations like Facebook, Google, and OpenAI to develop and maintain open-source LLMs as well.
Open-source models are generally free. The availability of source code helps teams customize, fine-tune, and adapt the code to their needs.
Examples: Meta’s LlaMA, Falcon 180B, TextBlob
Use cases: You can have generalized, domain-specific and task-specific models that are open-source. It’s best when you have the capacity to review, choose, fine-tune and implement open-source LLMs on your own.
With that foundation, here are some pointers on choosing the right GenAI model for your enterprise needs.
If your use case is generalized and performs a wide range of tasks, choose large language models
If your use case is specific and limited to a few similar tasks, choose domain-specific, task-specific or small language models
If you need greater flexibility and customizability, go open-source
If you’re cost-sensitive, choose smaller open-source models
If you don’t have the resources to implement models yourself, choose proprietary models with more straightforward integration and better support
If you’re having trouble making effective and sustainable decisions, speak to Tune AI’s experts. We’ve worked with all kinds of models for use cases across enterprise departments. We’re happy to help.
When I say “generative AI,” what’s the first thing that comes to mind? If you said ChatGPT, you’re not alone. One of the most common misconceptions about GenAI is that it’s one thing, a single model that can serve every need.
In reality, there are 255 large-scale models (those with training compute over 1023 floating point operations, essentially the scale of GPT-3 or higher). These can be categorized in dozens of ways, depending on what you need it for, how much you’re willing to spend, how secure you need it to be, etc.
In this blog post, we explore a few of these tangents.
Model scale
Large-scale vs. small-scale or single-purpose micromodels
Let’s begin at the beginning. The primary indicator of an AI model tends to be its scale. Here are models looked at through the lens of scale.
Foundation models
Foundation models tend to be the largest and are trained on enormous data sets. They are characterized by their ability to:
Perform a wide range of tasks, logical, creative, or problem-solving
Generate multimodal output across text, images, code, or video
Be fine-tuned for specific tasks
Examples: Open AI’s GPT, Google’s Gemini, Claude 3.5 Sonnet
Use case: The best use case of these models tends to be in generating content based on natural language input. Writing blog posts, documentation, text summaries, etc. are great for this type of model.
Smaller or single-purpose micromodels
Smaller models tend to be able to do specific tasks well. They are typically overtrained and overfit to perform a specific task. They are:
Trained on fewer than 10 billion parameters
Lightweight, balancing performance and efficiency
Great for on-device applications, like mobile assistants or smart appliances
Useful for domain-specific tasks
Examples: TinyLlama, Mistral Nemo
Use case: Small models are great for specific tasks like code generation, language translation, mobile applications, etc.
Domain-applicability
General purpose vs. domain-specific models
Many AI models are typically built to be general-purpose, as in, they are able to do a wide variety of tasks across use cases. For instance, with a foundation model, you might build a chatbot for customer service, sales, scenario modeling or recommendation engines.
On the other hand, organizations are realizing the need for increased accuracy in processing specific contexts and terminology related to each domain. This resulted in domain-specific models, which are:
Smaller, trained on niche data related to the industry
More accurate in interpreting domain-specific language
Personalizable with greater precision
Examples: Med-PaLM, BloombergGPT, ClimateBERT
Use case: In healthcare, domain-specific models are great for diagnosis. In law, it’s good for contract analysis, compliance review, etc.
Task complexity
General purpose vs. task-specific models
Another dimension similar to domain-specificity is task-specificity. Simply put, a model that is trained for code completion wouldn’t be able to generate videos as well. It’s specific to that particular task.
Task-specific models are characterized by:
Specialization is one particular task or closely related tasks
Excellent accuracy and reduced hallucination in performing that specific task(s)
Lower latency and faster processing
Higher interpretability
Examples: CodeLLama, Veo
Use cases: A great use case for task-specific LLM is in language translation. You can enable distributed teams to converse in different languages with real-time AI-based translation.
Device and deployment
Heavy-weight vs. lightweight models
When we speak of model deployment, we often think of scalable cloud or sprawling on-prem infrastructure. This is true of heavy-weight, powerful models. However, not all use cases lend themselves to having AI deployed this way.
Lightweight models are:
Efficient, using as little computing as possible
Usable in on-device, such as mobile phones or smart devices
Functional despite limited connectivity
Examples: Gemini Nano, Phi-2
Use cases: For low-latency use cases, lightweight models deployed on devices like smartphones are perfect.
Model licensing
Proprietary vs. open source models
In the server operating systems space, open source has over 90% market share. This is a trend likely to duplicate in the LLM space as well.
Proprietary models
A proprietary large language model is developed and maintained by a singular provider, who often charges a subscription fee to use it. Proprietary models often come with regular updates, higher reliability, better performance and support.
You can deploy it in your environment or integrate it through APIs. However, you will not have access to the source code, limiting customization. You might also have limitations on how you use the software.
Examples: Claude 2, GPT-4
Use cases: If you’re just integrating Generative AI into your workflows for a general productivity boost, proprietary models are easier. For instance, Microsoft Copilot is easy to get started if your enterprise is on the Microsoft platform.
Open-source models
In open-source models, the source code and architecture are available for the public to use or modify. These models are community-driven, depending on a wide range of users for development and maintenance. However, it is common for big organizations like Facebook, Google, and OpenAI to develop and maintain open-source LLMs as well.
Open-source models are generally free. The availability of source code helps teams customize, fine-tune, and adapt the code to their needs.
Examples: Meta’s LlaMA, Falcon 180B, TextBlob
Use cases: You can have generalized, domain-specific and task-specific models that are open-source. It’s best when you have the capacity to review, choose, fine-tune and implement open-source LLMs on your own.
With that foundation, here are some pointers on choosing the right GenAI model for your enterprise needs.
If your use case is generalized and performs a wide range of tasks, choose large language models
If your use case is specific and limited to a few similar tasks, choose domain-specific, task-specific or small language models
If you need greater flexibility and customizability, go open-source
If you’re cost-sensitive, choose smaller open-source models
If you don’t have the resources to implement models yourself, choose proprietary models with more straightforward integration and better support
If you’re having trouble making effective and sustainable decisions, speak to Tune AI’s experts. We’ve worked with all kinds of models for use cases across enterprise departments. We’re happy to help.
Written by
Anshuman Pandey
Co-founder and CEO