AI Portal Gun
Open-Source-Models
Tuned Model

Tuned Models

Within the realm of LLMs, the "Tune model" assumes a position of paramount significance. Exacting fine-tuning on dedicated datasets empowers it to excel in various tasks, encompassing image classification, sentiment analysis, and language translation. This refined model serves as the foundational underpinning for innovative AI applications, notably within the domains of healthcare, autonomous vehicle technology, and augmented reality.



training model img

Models

ModelsDeveloped byRelease DateParameterDescription
Mistral (opens in a new tab)Ashvini Kumar JindalArithmo-Mistral-7B (opens in a new tab)Trained to reason and answer mathematical problems and is also capable of writing a Python program that upon execution prints answer to the question, used Mistral-7B as a base model and used QLoRA to fine-tune it on a single RTX 4090 GPU.
Falcon (opens in a new tab)TekniumOpenHermes-2-Mistral-7B (opens in a new tab)Trained on 900k entries of GPT-4 generated data from diverse open AI datasets, it underwent extensive filtering and format conversion to ShareGPT, followed by ChatML transformation.
LLaVA (opens in a new tab)Haotian LiuLLaVA-7B (opens in a new tab), LLaVA-13B (opens in a new tab)LLaVA-1.5 excels in 11 benchmarks with minimal changes from LLaVA, using only public data, quick training in ~1 day on a single 8-A100 node, and outperforming billion-scale data methods. It's a powerful multimodal model, rivaling GPT-4 in chat capabilities and setting a new Science QA accuracy record.
zephyr-7b-alpha (opens in a new tab)HuggingFaceH4 (opens in a new tab)zephyr-7b-alpha (opens in a new tab)Zephyr-7B-Alpha, from Hugging Face's Zephyr series, is a GPT-like language model with 7B parameters. While it aims to be a helpful assistant, it's not fine-tuned for human preferences and may produce problematic outputs. It's based on Mistral-7B-v0.1, mainly trained on English data and available under the CC BY-NC 4.0 license.
SSD-1B (opens in a new tab)Segmind (opens in a new tab)SSD-1B (opens in a new tab)Segmind's SSD-1B is a compact, faster variant of SDXL, maintaining top-notch text-to-image generation. Trained on diverse datasets, it uses knowledge distillation from expert models like SDXL (opens in a new tab), ZavyChromaXL (opens in a new tab), and JuggernautXL (opens in a new tab) to create impressive visuals from text prompts.
MistralLite (opens in a new tab)Amazon (opens in a new tab)MistralLite (opens in a new tab)MistralLite, a fine-tuned AWS model on Hugging Face, handles long context efficiently (up to 32K tokens) with adapted Rotary Embedding. It's ideal for tasks like topic retrieval, summarization, and Q&A. Deployable on AWS g5.2x instances with Sagemaker Huggingface Text Generation Inference. Supports Python usage and is Apache 2.0 licensed.
MetaMath (opens in a new tab)Meta-Math (opens in a new tab)MetaMath-7B (opens in a new tab)MetaMath is a project for enhancing mathematical questions for language models. It builds the MetaMathQA dataset and fine-tunes LLaMA-2 models, creating specialized mathematical reasoning models. Results show MetaMath's significant performance lead on GSM8K and MATH benchmarks, even surpassing models of the same size.
Alfred-40B (opens in a new tab)LightOn (opens in a new tab)Alfred-40B-1023 (opens in a new tab), Alfred-40B-0723 (opens in a new tab)Catering specifically to enterprise needs, the enhanced model surpasses its forerunner. Its standout features include minimized hallucinations, heightened self-awareness, advanced 'Chat with Docs,' and an expanded 8K token context. Derived from Falcon-40B, it primarily undergoes training in English, German, Spanish, and French, with limited exposure to other languages.