You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/source/introduction/overview.rst
+40-1
Original file line number
Diff line number
Diff line change
@@ -22,6 +22,45 @@ This flexibility ensures that scrapers remain functional even when website layou
22
22
We support many LLMs including **GPT, Gemini, Groq, Azure, Hugging Face** etc.
23
23
as well as local models which can run on your machine using **Ollama**.
24
24
25
+
AI Models and Token Limits
26
+
==========================
27
+
28
+
ScrapGraphAI supports a wide range of AI models from various providers. Each model has a specific token limit, which is important to consider when designing your scraping pipelines. Here's an overview of the supported models and their token limits:
29
+
30
+
OpenAI Models
31
+
-------------
32
+
- GPT-3.5 Turbo (16,385 tokens)
33
+
- GPT-4 (8,192 tokens)
34
+
- GPT-4 Turbo Preview (128,000 tokens)
35
+
36
+
Azure OpenAI Models
37
+
-------------------
38
+
- GPT-3.5 Turbo (16,385 tokens)
39
+
- GPT-4 (8,192 tokens)
40
+
- GPT-4 Turbo Preview (128,000 tokens)
41
+
42
+
Google AI Models
43
+
----------------
44
+
- Gemini Pro (128,000 tokens)
45
+
- Gemini 1.5 Pro (128,000 tokens)
46
+
47
+
Anthropic Models
48
+
----------------
49
+
- Claude Instant (100,000 tokens)
50
+
- Claude 2 (200,000 tokens)
51
+
- Claude 3 (200,000 tokens)
52
+
53
+
Mistral AI Models
54
+
-----------------
55
+
- Mistral Large (128,000 tokens)
56
+
- Open Mistral 7B (32,000 tokens)
57
+
- Open Mixtral 8x7B (32,000 tokens)
58
+
59
+
For a complete list of supported models and their token limits, please refer to the API documentation.
60
+
61
+
Understanding token limits is crucial for optimizing your scraping tasks. Larger token limits allow for processing more text in a single API call, which can be beneficial for scraping lengthy web pages or documents.
This module contains a comprehensive dictionary of AI models and their corresponding token limits. The `models_tokens` dictionary is organized by provider (e.g., OpenAI, Azure OpenAI, Google AI, etc.) and includes various models with their maximum token counts.
10
+
11
+
Example usage:
12
+
13
+
.. code-block:: python
14
+
15
+
from scrapegraphai.helpers.models_tokens import models_tokens
0 commit comments