You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This is a Phi Family of SLMs book for getting started with Phi Models. Phi a family of open sourced AI models developed by Microsoft. Phi models are the most capable and cost-effective small language models (SLMs) available, outperforming models of the same size and next size up across a variety of language, reasoning, coding, and math benchmarks
This repository provides everything you need to perform Supervised Fine-Tuning (SFT) of the Qwen2.5-Coder-1.5B-Instruct model—or any of its larger variants (7B, 14B, 32B)—on the Qwen Models, using the nvidia/OpenCodeReasoning dataset.
AtomMind is a lightweight Small Scientific Language Model (Sslm) for reasoning across Math, Physics, Chemistry, and Biology using domain experts, symbolic reasoning, and optimization modules. It supports optional memory and self-monitoring to improve problem-solving and accuracy.
Probing whether reasoning can be structurally crystallised into a small LLM via cyclic domain training + gap phases. Testing emergence at 300M parameters from scratch. SAE-based interpretability. Pruning crossover as depth measurement. Open collaboration.
A comprehensive Natural Language Processing project focused on clinical dialogue analysis, featuring medical text classification, summarization and generation using traditional machine learning, custom deep learning architectures, and transformer-based fine-tuning approaches.