Skip to content

Commit e201195

Browse files
committed
New models
1 parent 44da319 commit e201195

File tree

3 files changed

+148
-6
lines changed

3 files changed

+148
-6
lines changed

data/com.jeffser.Alpaca.metainfo.xml.in

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -97,6 +97,7 @@
9797
<ul>
9898
<li>Option to delete all chats</li>
9999
<li>Button to refresh sample prompts</li>
100+
<li>New models</li>
100101
</ul>
101102
<p>Fixes</p>
102103
<ul>

src/available_models.json

Lines changed: 141 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -33,9 +33,7 @@
3333
],
3434
"author": "Qwen Team",
3535
"categories": [
36-
"huge",
37-
"code",
38-
"math"
36+
"huge"
3937
],
4038
"languages": [
4139
"en"
@@ -2918,7 +2916,6 @@
29182916
"author": "Paraphrase Team",
29192917
"categories": [
29202918
"small",
2921-
"multilingual",
29222919
"embedding"
29232920
],
29242921
"languages": [
@@ -3202,8 +3199,8 @@
32023199
"categories": [
32033200
"small",
32043201
"medium",
3205-
"embedding",
32063202
"code",
3203+
"embedding",
32073204
"multilingual"
32083205
],
32093206
"author": "IBM Research",
@@ -3456,5 +3453,144 @@
34563453
"zh",
34573454
"ja"
34583455
]
3456+
},
3457+
"gemma3": {
3458+
"tags": [
3459+
[
3460+
"1b",
3461+
"815\u202fMB"
3462+
],
3463+
[
3464+
"4b",
3465+
"3.3\u202fGB"
3466+
],
3467+
[
3468+
"12b",
3469+
"8.1\u202fGB"
3470+
],
3471+
[
3472+
"27b",
3473+
"17\u202fGB"
3474+
]
3475+
],
3476+
"url": "https://ollama.com/library/gemma3",
3477+
"categories": [
3478+
"small",
3479+
"medium",
3480+
"big",
3481+
"huge"
3482+
],
3483+
"author": "Google DeepMind",
3484+
"languages": [
3485+
"en"
3486+
]
3487+
},
3488+
"phi4-mini": {
3489+
"tags": [
3490+
[
3491+
"3.8b",
3492+
"2.5\u202fGB"
3493+
]
3494+
],
3495+
"url": "https://ollama.com/library/phi4-mini",
3496+
"categories": [
3497+
"small",
3498+
"math",
3499+
"multilingual"
3500+
],
3501+
"author": "Microsoft",
3502+
"languages": [
3503+
"en",
3504+
"ar",
3505+
"zh",
3506+
"cs",
3507+
"da",
3508+
"nl",
3509+
"fi",
3510+
"fr",
3511+
"de",
3512+
"he",
3513+
"hu",
3514+
"it",
3515+
"ja",
3516+
"ko",
3517+
"no",
3518+
"pl",
3519+
"pt",
3520+
"ru",
3521+
"es",
3522+
"sv",
3523+
"th",
3524+
"tr",
3525+
"uk"
3526+
]
3527+
},
3528+
"granite3.2-vision": {
3529+
"tags": [
3530+
[
3531+
"2b",
3532+
"2.4\u202fGB"
3533+
]
3534+
],
3535+
"url": "https://ollama.com/library/granite3.2-vision",
3536+
"categories": [
3537+
"small",
3538+
"vision"
3539+
],
3540+
"author": "IBM for Code Intelligence",
3541+
"languages": [
3542+
"en"
3543+
]
3544+
},
3545+
"granite3.2": {
3546+
"tags": [
3547+
[
3548+
"2b",
3549+
"1.5\u202fGB"
3550+
],
3551+
[
3552+
"8b",
3553+
"4.9\u202fGB"
3554+
]
3555+
],
3556+
"url": "https://ollama.com/library/granite3.2",
3557+
"categories": [
3558+
"small",
3559+
"medium",
3560+
"multilingual"
3561+
],
3562+
"author": "IBM for Code Intelligence",
3563+
"languages": [
3564+
"en",
3565+
"de",
3566+
"es",
3567+
"fr",
3568+
"ja",
3569+
"pt",
3570+
"ar",
3571+
"cs",
3572+
"it",
3573+
"ko",
3574+
"nl",
3575+
"zh"
3576+
]
3577+
},
3578+
"command-r7b-arabic": {
3579+
"tags": [
3580+
[
3581+
"7b",
3582+
"5.1\u202fGB"
3583+
]
3584+
],
3585+
"url": "https://ollama.com/library/command-r7b-arabic",
3586+
"categories": [
3587+
"medium",
3588+
"multilingual"
3589+
],
3590+
"author": "Cohere",
3591+
"languages": [
3592+
"en",
3593+
"ar"
3594+
]
34593595
}
34603596
}

src/available_models_descriptions.py

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
descriptions = {
22
'llama3.3': _("New state of the art 70B model. Llama 3.3 70B offers similar performance compared to the Llama 3.1 405B model."),
3-
'qwq': _("QwQ is an experimental research model focused on advancing AI reasoning capabilities."),
3+
'qwq': _("QwQ is the reasoning model of the Qwen series."),
44
'llama3.2-vision': _("Llama 3.2 Vision is a collection of instruction-tuned image reasoning generative models in 11B and 90B sizes."),
55
'llama3.2': _("Meta's Llama 3.2 goes small with 1B and 3B models."),
66
'llama3.1': _("Llama 3.1 is a new state-of-the-art model from Meta available in 8B, 70B and 405B parameter sizes."),
@@ -154,4 +154,9 @@
154154
'openthinker': _("A fully open-source family of reasoning models built using a dataset derived by distilling DeepSeek-R1."),
155155
'deepscaler': _("A fine-tuned version of Deepseek-R1-Distilled-Qwen-1.5B that surpasses the performance of OpenAI’s o1-preview with just 1.5B parameters on popular math evaluations."),
156156
'r1-1776': _("A version of the DeepSeek-R1 model that has been post trained to provide unbiased, accurate, and factual information by Perplexity."),
157+
'gemma3': _("The current strongest model that fits on a single GPU."),
158+
'phi4-mini': _("Phi-4-mini brings significant enhancements in multilingual support, reasoning, and mathematics, and now, the long-awaited function calling feature is finally supported."),
159+
'granite3.2-vision': _("A compact and efficient vision-language model, specifically designed for visual document understanding, enabling automated content extraction from tables, charts, infographics, plots, diagrams, and more."),
160+
'granite3.2': _("Granite-3.2 is a family of long-context AI models from IBM Granite fine-tuned for thinking capabilities."),
161+
'command-r7b-arabic': _("A new state-of-the-art version of the lightweight Command R7B model that excels in advanced Arabic language capabilities for enterprises in the Middle East and Northern Africa."),
157162
}

0 commit comments

Comments
 (0)