Skip to content
Thoughts on GitHub Models?

Phi-3.5-mini instruct (128k)

Refresh of Phi-3-mini model.
Context
131k input · 4k output
Training date
Aug 2024
Rate limit tier
Provider support
Try Phi-3.5-mini instruct (128k)
Get early access to our playground for modelsJoin our limited beta waiting list today and be among the first to try out an easy way to test models

Model navigation navigation

Microsoft

Phi-3.5-mini is a lightweight, state-of-the-art open model built upon datasets used for Phi-3 - synthetic data and filtered publicly available websites - with a focus on very high-quality, reasoning dense data. The model belongs to the Phi-3 model family and supports 128K token context length. The model underwent a rigorous enhancement process, incorporating both supervised fine-tuning, proximal policy optimization, and direct preference optimization to ensure precise instruction adherence and robust safety measures.

Resources

🏡 Phi-3 Portal

📰 Phi-3 Microsoft Blog

📖 Phi-3 Technical Report

👩‍🍳 Phi-3 Cookbook

Model Architecture

Phi-3.5-mini has 3.8B parameters and is a dense decoder-only Transformer model using the same tokenizer as Phi-3 Mini. It is a text-only model best suited for prompts using chat format.

Training Data

Phi-3.5-mini is a static model trained on an offline dataset with 3.4T tokens and a cutoff date October 2023 for publicly available data. Future versions of the tuned models may be released as we improve models.

Languages

 (23)
English, Arabic, Chinese, Czech, Danish, Dutch, Finnish, French, German, Hebrew, Hungarian, Italian

About

Refresh of Phi-3-mini model.
Context
131k input · 4k output
Training date
Aug 2024
Rate limit tier
Provider support

Languages

 (23)
English, Arabic, Chinese, Czech, Danish, Dutch, Finnish, French, German, Hebrew, Hungarian, Italian