|
1 | 1 | --- |
2 | 2 | title: Hugging Face |
3 | | -description: Access thousands of open-source AI models including Qwen, Llama, and CodeLlama through Hugging Face's cost-effective inference API. |
| 3 | +description: Access thousands of open-source AI models through Hugging Face's inference API. |
4 | 4 | --- |
5 | 5 |
|
6 | | -## Overview |
| 6 | +**Website:** [https://huggingface.co/](https://huggingface.co/) |
7 | 7 |
|
8 | | -Hugging Face democratizes AI by providing easy access to cutting-edge models. Their platform hosts models for text generation, code completion, analysis, and more, all accessible through a simple API interface. |
| 8 | +## Getting an API Key |
9 | 9 |
|
10 | | -<CardGroup cols={3}> |
11 | | - <Card title="Open Source Models" icon="unlock"> |
12 | | - Access to thousands of community models |
13 | | - </Card> |
14 | | - <Card title="Research Models" icon="microscope"> |
15 | | - Latest models from top AI research labs |
16 | | - </Card> |
17 | | - <Card title="Cost Effective" icon="dollar-sign"> |
18 | | - Affordable pricing for open-source models |
19 | | - </Card> |
20 | | -</CardGroup> |
| 10 | +1. Go to [Hugging Face](https://huggingface.co/) and sign in |
| 11 | +2. Navigate to [Settings > Access Tokens](https://huggingface.co/settings/tokens) |
| 12 | +3. Create a new token with "Read" permissions |
| 13 | +4. Copy the token immediately |
21 | 14 |
|
22 | | -## Available Models |
| 15 | +## Configuration |
23 | 16 |
|
24 | | -<AccordionGroup> |
25 | | - <Accordion title="Qwen Models" icon="code"> |
26 | | - ### Qwen Series |
27 | | - High-quality models from Alibaba Cloud, excellent for coding and general tasks. |
| 17 | +1. Click the settings icon (⚙️) in CodinIT |
| 18 | +2. Select "Hugging Face" as the API Provider |
| 19 | +3. Paste your token |
| 20 | +4. Choose your model |
28 | 21 |
|
29 | | - - **Qwen2.5-Coder-32B**: Specialized for code generation |
30 | | - - **Qwen2.5-72B**: General purpose large model |
31 | | - - **Best for**: Code completion, technical writing, analysis |
| 22 | +## Supported Models |
32 | 23 |
|
33 | | - </Accordion> |
| 24 | +- **Qwen series:** Qwen2.5-Coder-32B, Qwen2.5-72B |
| 25 | +- **Llama series:** Llama-3.1-70B, Llama-3.1-405B |
| 26 | +- **CodeLlama:** CodeLlama-34B |
| 27 | +- **Yi series:** Yi-1.5-34B |
| 28 | +- **Hermes series:** Hermes-3-Llama-3.1-8B |
34 | 29 |
|
35 | | - <Accordion title="Llama Models" icon="brain"> |
36 | | - ### Meta Llama Series |
37 | | - Industry-leading open-source models from Meta. |
| 30 | +## Features |
38 | 31 |
|
39 | | - - **Llama-3.1-70B**: Powerful general-purpose model |
40 | | - - **Llama-3.1-405B**: Massive model for complex tasks |
41 | | - - **Best for**: Advanced reasoning, creative tasks |
| 32 | +- **Open source:** All models openly available |
| 33 | +- **Community driven:** Constantly updated |
| 34 | +- **Research access:** Latest models from top labs |
| 35 | +- **Cost effective:** Affordable pricing |
| 36 | +- **Free tier:** Available for testing |
42 | 37 |
|
43 | | - </Accordion> |
| 38 | +## Notes |
44 | 39 |
|
45 | | - <Accordion title="CodeLlama" icon="terminal"> |
46 | | - ### CodeLlama Series |
47 | | - Specialized models for programming and code-related tasks. |
48 | | - |
49 | | - - **CodeLlama-34B**: Large coding model |
50 | | - - **Best for**: Code generation, debugging, technical analysis |
51 | | - |
52 | | - </Accordion> |
53 | | - |
54 | | - <Accordion title="Yi Models" icon="sparkle"> |
55 | | - ### Yi Series |
56 | | - High-performance models from 01.AI. |
57 | | - |
58 | | - - **Yi-1.5-34B**: Balanced performance and capability |
59 | | - - **Best for**: General AI tasks, analysis, writing |
60 | | - |
61 | | - </Accordion> |
62 | | - |
63 | | - <Accordion title="Hermes Models" icon="wand"> |
64 | | - ### Hermes Series |
65 | | - Fine-tuned models optimized for helpfulness and reasoning. |
66 | | - |
67 | | - - **Hermes-3-Llama-3.1-8B**: Efficient and capable |
68 | | - - **Best for**: Conversational AI, helpful responses |
69 | | - |
70 | | - </Accordion> |
71 | | -</AccordionGroup> |
72 | | - |
73 | | -## Setup Instructions |
74 | | - |
75 | | -<Steps> |
76 | | - <Step title="Create Hugging Face Account"> |
77 | | - Visit [Hugging Face](https://huggingface.co/) and create a free account |
78 | | - </Step> |
79 | | - <Step title="Generate API Token"> |
80 | | - Go to [Settings > Access Tokens](https://huggingface.co/settings/tokens) and create a new token |
81 | | - </Step> |
82 | | - <Step title="Configure Token Permissions">Ensure your token has "Read" permissions for model inference</Step> |
83 | | - <Step title="Add to Codinit">Enter your Hugging Face token in the provider settings</Step> |
84 | | - <Step title="Test Models">Try different models to find the best fit for your needs</Step> |
85 | | -</Steps> |
86 | | - |
87 | | -## Key Features |
88 | | - |
89 | | -<BadgeGroup> |
90 | | - <Badge variant="secondary">Open Source</Badge> |
91 | | - <Badge variant="secondary">Community Driven</Badge> |
92 | | - <Badge variant="secondary">Research Models</Badge> |
93 | | - <Badge variant="secondary">Cost Effective</Badge> |
94 | | - <Badge variant="secondary">Diverse Models</Badge> |
95 | | -</BadgeGroup> |
96 | | - |
97 | | -### Platform Advantages |
98 | | - |
99 | | -- **Open Source**: All models are openly available and auditable |
100 | | -- **Community Driven**: Constantly updated by global AI community |
101 | | -- **Research Access**: Latest models from top research institutions |
102 | | -- **Flexible Pricing**: Pay only for what you use |
103 | | -- **Wide Selection**: Models for every use case and skill level |
104 | | - |
105 | | -## Use Cases |
106 | | - |
107 | | -<AccordionGroup> |
108 | | - <Accordion title="Code Development" icon="code"> |
109 | | - ### Programming Tasks |
110 | | - Specialized models for software development and coding assistance. |
111 | | - |
112 | | - - Code generation and completion |
113 | | - - Code review and analysis |
114 | | - - Debugging assistance |
115 | | - - Technical documentation |
116 | | - |
117 | | - </Accordion> |
118 | | - |
119 | | - <Accordion title="Research & Analysis" icon="search"> |
120 | | - ### Academic Research |
121 | | - Powerful models for research, analysis, and academic work. |
122 | | - |
123 | | - - Scientific paper analysis |
124 | | - - Research summarization |
125 | | - - Data interpretation |
126 | | - - Academic writing assistance |
127 | | - |
128 | | - </Accordion> |
129 | | - |
130 | | - <Accordion title="Content Creation" icon="edit"> |
131 | | - ### Creative Writing |
132 | | - Models for content creation and creative tasks. |
133 | | - |
134 | | - - Creative writing |
135 | | - - Content generation |
136 | | - - Language translation |
137 | | - - Educational content |
138 | | - |
139 | | - </Accordion> |
140 | | - |
141 | | - <Accordion title="Business Applications" icon="building"> |
142 | | - ### Enterprise Use |
143 | | - Suitable for business and productivity applications. |
144 | | - |
145 | | - - Business analysis |
146 | | - - Report generation |
147 | | - - Customer communication |
148 | | - - Process automation |
149 | | - |
150 | | - </Accordion> |
151 | | -</AccordionGroup> |
152 | | - |
153 | | -## Pricing Information |
154 | | - |
155 | | -Hugging Face offers flexible pricing based on model size and usage: |
156 | | - |
157 | | -<TokenPricing> |
158 | | - <TokenPricingItem model="Small Models (<7B)" input="$0.0001" output="$0.0002" /> |
159 | | - <TokenPricingItem model="Medium Models (7-20B)" input="$0.0002" output="$0.0004" /> |
160 | | - <TokenPricingItem model="Large Models (20-40B)" input="$0.0004" output="$0.0008" /> |
161 | | - <TokenPricingItem model="XL Models (40B+)" input="$0.0008" output="$0.0016" /> |
162 | | -</TokenPricing> |
163 | | - |
164 | | -<Callout type="info">**Free Tier**: Hugging Face offers a generous free tier for testing and light usage.</Callout> |
165 | | - |
166 | | -<Callout type="tip"> |
167 | | - **Model Selection**: Start with smaller models for testing, then scale up to larger models for production use. |
168 | | -</Callout> |
169 | | - |
170 | | -<Callout type="warning"> |
171 | | - **Rate Limits**: Free tier has usage limits. Paid plans offer higher rate limits and priority access. |
172 | | -</Callout> |
173 | | - |
174 | | -## Model Performance Notes |
175 | | - |
176 | | -<AccordionGroup> |
177 | | - <Accordion title="Inference Speed" icon="zap"> |
178 | | - ### Speed Considerations |
179 | | - Model size affects response time and resource usage. |
180 | | - |
181 | | - - **Small models**: Fast responses, lower resource usage |
182 | | - - **Large models**: Slower responses, higher resource usage |
183 | | - - **Consider trade-offs**: Speed vs. quality based on your needs |
184 | | - |
185 | | - </Accordion> |
186 | | - |
187 | | - <Accordion title="Context Limits" icon="file-text"> |
188 | | - ### Token Limits |
189 | | - Different models have varying context window sizes. |
190 | | - |
191 | | - - **Most models**: 4K-8K token context windows |
192 | | - - **Specialized models**: May have different limits |
193 | | - - **Check documentation**: Verify limits for your chosen model |
194 | | - |
195 | | - </Accordion> |
196 | | - |
197 | | - <Accordion title="Model Updates" icon="refresh"> |
198 | | - ### Staying Current |
199 | | - Hugging Face models are frequently updated by the community. |
200 | | - |
201 | | - - **Regular updates**: New model versions released frequently |
202 | | - - **Version pinning**: Specify exact model versions for consistency |
203 | | - - **Community contributions**: New models added regularly |
204 | | - |
205 | | - </Accordion> |
206 | | -</AccordionGroup> |
| 40 | +- **Pricing:** Based on model size and usage |
| 41 | +- **Rate limits:** Free tier has usage limits |
0 commit comments