Chapter 2: I am an AI model
-
Unit 1 - Starting from scratch6 Topics|2 Quizzes
-
1.1 Recognizing Opportunities for Your Business
-
1.2 Driving Factors and Strategic Considerations for AI Adoption in SMEs
-
Activity: “Glossary”
-
1.3 Powering Business Efficiency with AI LLM Tools
-
1.4 Types of Contracts for Adopting AI Software and Services
-
1.5 Establishing Internal Policies for Responsible LLM Use
-
1.1 Recognizing Opportunities for Your Business
-
Unit 2 - Develop a personalised AI strategy for your business4 Topics|3 Quizzes
Quizzes
1.3 Powering Business Efficiency with AI LLM Tools
Having explored various AI applications and how to evaluate features, let’s now look ‘under the hood’ at a core technology driving many of these advancements: Large Language Models, often referred to as LLMs. This section focuses on helping you understand these foundational models better.
1.4.1 Understanding Popular AI Language Models for Business Usere
If you’re considering creating your own AI-driven solutions or integrating powerful language capabilities directly into your workflows, you’ll likely encounter well-known foundational LLMs such as OpenAI’s ChatGPT series, Anthropic’s Claude, or Google’s Gemini. Choosing which model or provider might work best for your specific needs can seem like a task reserved for AI experts, given the rapid developments and technical nuances. However, you can approach this at an introductory level to gain clarity on the key factors to consider for an initial assessment.
(While this provides a starting point, making strategic, long-term decisions about adopting or building on specific LLMs often requires deeper technical analysis and due diligence beyond this introductory overview.)
Initial Considerations When Choosing an LLM
Before diving deep into specific model comparisons, start by clarifying your own requirements. Thinking through these points will help you narrow down options and ask vendors the right questions:
- Identify Your Needs: What specific task(s) do you primarily want the AI to perform? (e.g., drafting marketing copy, summarizing reports, answering customer service queries, generating code, analyzing data?). Does your task require expertise or tuning specific to your industry?
- Ease of Use and Technical Resources: How easily can the model be implemented? Are you looking for ready-to-use interfaces (like the web versions of ChatGPT, Claude, Gemini), or do you need API access for integration? Consider platforms with strong vendor support or extensive documentation if your internal technical resources are limited.
- Customization and Fine-Tuning: Do you need the AI to learn your specific company voice, terminology, or knowledge base? Assess whether you need the capability to fine-tune a model on your own data, and what that would entail.
- Integration Capabilities: How easily can the model’s API or platform integrate with the software and systems you already use (e.g., CRM, email marketing tools, internal databases)? Check for compatibility and available integrations.
- Cost Structure: LLM costs can vary significantly. Understand the pricing model (e.g., subscription tiers, pay-per-use based on tokens, flat fees). Factor in potential setup costs, integration expenses, and ongoing operational costs, not just the base price.
- Data Privacy and Compliance: This is critical. Ensure any considered model or platform complies with data protection laws relevant to your business and customers (like GDPR). Understand clearly how your input data is used (e.g., for training the vendor’s model?), stored, and protected. Look for vendors transparent about their security and compliance measures.
- Scalability and Future Needs: Choose a model and provider that can potentially grow with your business needs. Consider how often the models are updated and improved. Is there a clear future roadmap that aligns with your potential long-term usage?
Comparing Specific Models
Understanding the fundamental differences between major LLMs is helpful. While the absolute “best” model changes rapidly with new releases, the types of differences often persist. Some models might excel in creative writing (like certain versions of ChatGPT), others prioritize safety and nuanced reasoning (often associated with Claude), while some lead in multimodal capabilities (handling text, images, audio like Gemini) or deep integration with specific ecosystems.
The following table presents an overview comparing some established LLM providers often discussed in a business context: OpenAI (ChatGPT models), Anthropic (Claude models), and Google (Gemini models) in 2025 (May).
Important Note: The AI landscape changes extremely fast! Specific model versions (e.g., GPT-4 vs. GPT-4o, different Claude 3 variants like Haiku/Sonnet/Opus, various Gemini versions like Pro/Ultra) have distinct capabilities, performance levels, context window sizes, and pricing. Always consult the providers’ current documentation for the latest, most accurate information before making any decisions. This table is intended only to illustrate the kinds of strategic differences and typical strengths associated with these major players, helping you understand what factors to compare.
Feature / Aspect | ChatGPT (OpenAI Models) | Claude (Anthropic Models) | Gemini (Google Models) |
General Focus / Strength | High Versatility, Strong Creativity, Conversational Ease, Broad Ecosystem | Emphasis on Safety, Ethics, Nuanced Reasoning, Handling Very Long Context | Strong Multimodality (Text, Image, Audio, Video), Deep Google Ecosystem Integration, Speed |
Key Characteristics | Excels at generating diverse creative text formats. | Designed with “Constitutional AI” for safety & ethical alignment. | Natively built to understand and combine multiple data types (text, images, etc.). |
Potential SME Uses | Marketing & Content Creation: Drafting emails, blog posts, social media content, product descriptions. | Compliance & HR: Analyzing dense legal documents or regulations, drafting internal policies, summarizing compliance materials. | Marketing & E-commerce: Generating descriptions based on product images, analyzing visual trends, creating multimodal ad content. |
Key Considerations / Trade-offs | May require more prompt engineering for highly nuanced or safety-critical tasks. | Conservatism might limit certain creative or exploratory uses. | Maximum benefit often realized within the Google Workspace ecosystem. |
As you’ve likely gathered, LLMs do not come in one size fits all! Choosing an LLM – or perhaps even more than one for different needs – is fundamentally connected to your intended use case, as each model is typically optimized for different ranges of tasks. Some models excel at creative generation, while others are engineered for more rigorous analysis or adherence to safety guidelines, and these represent just a few of the dimensions to consider!
1.4.2 Calculating Usage Costs for LLMs
When using Large Language Models (LLMs), especially when accessed via Application Programming Interfaces (APIs) from providers like OpenAI, Google, Anthropic, or through platforms like Microsoft Azure AI, understanding the cost structure is essential. Usage costs are typically calculated based on the amount of text processed, measured in units called “tokens.”
Token-Based Billing Explained
What is a Token?
In the context of AI language models, a token isn’t necessarily a whole word. It’s the basic unit of text the model processes. Depending on the model’s specific “tokenizer” (the tool that breaks down text), a token could be a whole word, a part of a word (subword), a single character, or even punctuation. The process of breaking text down this way is called tokenization.
Estimation for English
While the exact count varies, a common rule of thumb for English text is that one token represents roughly 4 characters or about 0.75 words. Conversely, one word is approximately 1.33 tokens. (Note: This ratio can differ significantly for other languages and for text containing lots of punctuation or code. Most providers offer online tools to calculate the precise token count for your specific text based on their tokenizer.)
Input vs. Output Tokens
Costs are calculated based on both the text you send to the model (your input tokens or “prompt”) and the text the model generates for you (the output tokens or “completion”/”response”).
Pricing Structure
Crucially, providers often charge different rates per token for input versus output. Typically, output tokens are more expensive because they reflect the computational effort of the model generating new content. Furthermore, costs vary significantly depending on the specific LLM used (more advanced models generally cost more per token). Some providers may offer volume discounts or tiered pricing based on usage levels, potentially allowing for lower rates if you commit to a certain minimum usage. Always check the specific provider’s terms.
Calculating Costs
The general formula to calculate the cost for a specific API call, considering potentially different rates, is:
Total Cost = (Input Tokens / 1000 * Cost per 1k Input Tokens) + (Output Tokens / 1000 * Cost per 1k Output Tokens)
(Note: Pricing is almost universally quoted per 1,000 tokens, sometimes abbreviated as “/1k tokens” or “/kT”).
IMPORTANT PRICING DISCLAIMER: The token costs mentioned in the following examples are purely illustrative and used only to demonstrate the calculation method. They do not reflect current market rates, or any specific provider’s pricing. LLM pricing changes frequently and varies widely between providers (OpenAI, Anthropic, Google, Microsoft Azure AI, etc.) and different model versions (e.g., basic vs. advanced capabilities). You MUST always consult the official, current pricing pages of the specific LLM provider and model version you intend to use for accurate cost estimation.
Practical Examples:
Let’s apply the formula with illustrative pricing:
Example 1: Simple Chatbot Conversation
- User Input: 15 words
- AI Response: 25 words
- Calculation Steps:
- Estimate Tokens (using 1 word ≈ 1.33 tokens):
- Input Tokens: 15 words × 1.33 ≈ 20 tokens
- Output Tokens: 25 words × 1.33 ≈ 33 tokens
- Calculate Cost (using a single hypothetical rate of $0.002 per 1k tokens for simplicity here):
- Total Tokens = 20 + 33 = 53 tokens
- Total Cost = (53 / 1000) * $0.002 = $0.000106
- Result: Cost per conversation ≈ $0.0001 (less than a cent).
- Estimate Tokens (using 1 word ≈ 1.33 tokens):
Example 2: Document Summarization
- Input Document: 2,500 words (approx. 5 pages)
- Desired Summary: 500 words (approx. 1 page)
- Calculation Steps:
- Estimate Tokens:
- Input Tokens: 2,500 words × 1.33 ≈ 3,325 tokens
- Output Tokens: 500 words × 1.33 ≈ 665 tokens
- Calculate Cost (using different hypothetical rates: Input @ $0.03/1k tokens, Output @ $0.06/1k tokens):
- Input Cost = (3,325 / 1000) * $0.03 = $0.09975
- Output Cost = (665 / 1000) * $0.06 = $0.0399
- Total Cost = $0.09975 + $0.0399 = $0.13965
- Result: Cost to summarize ≈ $0.14.
- Estimate Tokens:
Example 3: Content Generation
- Task: Generate a 2,000-word article
- Input Prompt: 100 words
- Calculation Steps:
- Estimate Tokens:
- Input Tokens: 100 words × 1.33 ≈ 133 tokens
- Output Tokens: 2,000 words × 1.33 ≈ 2,660 tokens
- Calculate Cost (using same hypothetical rates as Ex 2: Input @ $0.03/1k, Output @ $0.06/1k):
- Input Cost = (133 / 1000) * $0.03 = $0.00399
- Output Cost = (2,660 / 1000) * $0.06 = $0.1596
- Total Cost = $0.00399 + $0.1596 = $0.16359
- Result: Cost to generate article ≈ $0.16.
- Estimate Tokens:
Example 4: Translation of a Document
- Document Length: 5,000 words (approx. 10 pages)
- Assumption: Output word count is roughly similar to input.
- Calculation Steps:
- Estimate Tokens:
- Input Tokens: 5,000 words × 1.33 ≈ 6,650 tokens
- Output Tokens: Assume approx. 6,650 tokens
- Calculate Cost (using a single hypothetical rate of $0.002 per 1k tokens for simplicity):
- Total Tokens = 6,650 + 6,650 = 13,300 tokens
- Total Cost = (13,300 / 1000) * $0.002 = $0.0266
- Result: Cost to translate ≈ $0.027.
- Estimate Tokens:
Beyond Per-Token Costs
While token-based pricing for API calls is common, keep in mind other potential costs when budgeting for AI integration:
- Fine-tuning: If you need to adapt a pre-trained model to your specific data or task, there are typically costs associated with the training process itself, and potentially higher ongoing costs for hosting the customized model.
- Dedicated Capacity / Provisioned Throughput: For applications requiring guaranteed performance levels or very high volume, some providers offer dedicated instances of models, often billed hourly or monthly with committed usage, rather than strictly per token.
- Platform Fees: If accessing LLMs through a third-party platform or software that bundles AI features, there might be subscription fees separate from or in addition to the underlying token usage costs.
- Integration & Development: Don’t forget the internal or external costs associated with the initial development work required to integrate the LLM API into your existing software or workflows.
Understanding these cost components allows for more realistic budgeting and ROI calculation when considering LLM adoption for your SME.