Showing 67 open source projects for "scratch"

View related business solutions
  • The Most Powerful Software Platform for EHSQ and ESG Management Icon
    The Most Powerful Software Platform for EHSQ and ESG Management

    Addresses the needs of small businesses and large global organizations with thousands of users in multiple locations.

    Choose from a complete set of software solutions across EHSQ that address all aspects of top performing Environmental, Health and Safety, and Quality management programs.
    Learn More
  • AestheticsPro Medical Spa Software Icon
    AestheticsPro Medical Spa Software

    Our new software release will dramatically improve your medspa business performance while enhancing the customer experience

    AestheticsPro is the most complete Aesthetics Software on the market today. HIPAA Cloud Compliant with electronic charting, integrated POS, targeted marketing and results driven reporting; AestheticsPro delivers the tools you need to manage your medical spa business. It is our mission To Provide an All-in-One Cutting Edge Software to the Aesthetics Industry.
    Learn More
  • 1
    LLMs-from-scratch

    LLMs-from-scratch

    Implement a ChatGPT-like LLM in PyTorch from scratch, step by step

    LLMs-from-scratch is an educational codebase that walks through implementing modern large-language-model components step by step. It emphasizes building blocks—tokenization, embeddings, attention, feed-forward layers, normalization, and training loops—so learners understand not just how to use a model but how it works internally. The repository favors clear Python and NumPy or PyTorch implementations that can be run and modified without heavyweight frameworks obscuring the logic. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    llms-from-scratch-cn

    llms-from-scratch-cn

    Build a large language model from 0 only with Python foundation

    llms-from-scratch-cn is an educational open-source project designed to teach developers how to build large language models step by step using practical code and conceptual explanations. The repository provides a hands-on learning path that begins with the fundamentals of natural language processing and gradually progresses toward implementing full GPT-style architectures from the ground up.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    MiniMind

    MiniMind

    Train a 26M-parameter GPT from scratch in just 2h

    minimind is a framework that enables users to train a 26-million-parameter GPT (Generative Pre-trained Transformer) model from scratch in approximately two hours. It provides a streamlined process for data preparation, model training, and evaluation, making it accessible for individuals and organizations to develop their own language models without extensive computational resources.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 4
    GPT All Star

    GPT All Star

    AI-powered code generation tool for scratch development of web apps

    AI-powered code generation tool for scratch development of web applications with a team collaboration of autonomous AI agents. This is a research project, and its primary value is to explore the possibility of autonomous AI agents.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Skillfully - The future of skills based hiring Icon
    Skillfully - The future of skills based hiring

    Realistic Workplace Simulations that Show Applicant Skills in Action

    Skillfully transforms hiring through AI-powered skill simulations that show you how candidates actually perform before you hire them. Our platform helps companies cut through AI-generated resumes and rehearsed interviews by validating real capabilities in action. Through dynamic job specific simulations and skill-based assessments, companies like Bloomberg and McKinsey have cut screening time by 50% while dramatically improving hire quality.
    Learn More
  • 5
    DeepSeek Coder

    DeepSeek Coder

    DeepSeek Coder: Let the Code Write Itself

    DeepSeek-Coder is a series of code-specialized language models designed to generate, complete, and infill code (and mixed code + natural language) with high fluency in both English and Chinese. The models are trained from scratch on a massive corpus (~2 trillion tokens), of which about 87% is code and 13% is natural language. This dataset covers project-level code structure (not just line-by-line snippets), using a large context window (e.g. 16K) and a secondary fill-in-the-blank objective to encourage better contextual completions and infilling. Multiple sizes of the model are offered (e.g. 1B, 5.7B, 6.7B, 33B) so users can trade off inference cost vs capability. ...
    Downloads: 10 This Week
    Last Update:
    See Project
  • 6
    Hermes Agent

    Hermes Agent

    The agent that grows with you

    ...Rather than functioning as a stateless chatbot, it maintains long-term memory across sessions and can generate searchable “Skill Documents” that capture how it solved complex tasks so it doesn’t start from scratch each time. The agent interfaces with messaging platforms like Telegram, Discord, Slack, and WhatsApp through a single gateway process, and also offers an interactive terminal user interface with history, autocomplete, and streamable tool output. It supports scheduled automation in natural language, allowing users to set up recurring tasks such as daily briefings or system audits that it runs unattended.
    Downloads: 80 This Week
    Last Update:
    See Project
  • 7
    SimpleLLM

    SimpleLLM

    950 line, minimal, extensible LLM inference engine built from scratch

    SimpleLLM is a minimal, extensible large language model inference engine implemented in roughly 950 lines of code, built from scratch to serve both as a learning tool and a research platform for novel inference techniques. It provides the core components of an LLM runtime—such as tokenization, batching, and asynchronous execution—without the abstraction overhead of more complex engines, making it easier for developers and researchers to understand and modify.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Happy-LLM

    Happy-LLM

    Large Language Model Principles and Practice Tutorial from Scratch

    Happy-LLM is an open-source educational project created by the Datawhale AI community that provides a structured and comprehensive tutorial for understanding and building large language models from scratch. The project guides learners through the entire conceptual and practical pipeline of modern LLM development, starting with foundational natural language processing concepts and gradually progressing to advanced architectures and training techniques. It explains the Transformer architecture, pre-training paradigms, and model scaling strategies while also providing hands-on coding examples so readers can implement and experiment with their own models. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Machine learning algorithms

    Machine learning algorithms

    Minimal and clean examples of machine learning algorithms

    Machine learning algorithms is an open-source repository that provides minimal and clean implementations of machine learning algorithms written primarily in Python. The project focuses on demonstrating how fundamental machine learning methods work internally by implementing them from scratch rather than relying on high-level libraries. This approach allows learners to study the mathematical and algorithmic details behind widely used models in a transparent and readable way. The repository includes implementations of both supervised and unsupervised learning techniques, along with dimensionality reduction and clustering methods. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Award-Winning Medical Office Software Designed for Your Specialty Icon
    Award-Winning Medical Office Software Designed for Your Specialty

    Succeed and scale your practice with cloud-based, data-backed, AI-powered healthcare software.

    RXNT is an ambulatory healthcare technology pioneer that empowers medical practices and healthcare organizations to succeed and scale through innovative, data-backed, AI-powered software.
    Learn More
  • 10
    LTX-2

    LTX-2

    Python inference and LoRA trainer package for the LTX-2 audio–video

    ...It is architected to give developers low-level control over rendering pipelines, GPU resource management, shader orchestration, and cross-platform abstractions so they can craft visually compelling experiences without starting from scratch. Beyond basic rendering scaffolding, LTX-2 includes optimized math libraries, resource loaders, utilities for texture and buffer handling, and integration points for native event loops and input systems. The framework targets both interactive graphical applications and media-rich experiences, making it a solid foundation for games, creative tools, or visualization systems that demand both performance and flexibility. ...
    Downloads: 40 This Week
    Last Update:
    See Project
  • 11
    TurboQuant PyTorch

    TurboQuant PyTorch

    From-scratch PyTorch implementation of Google's TurboQuant

    TurboQuant PyTorch is a specialized deep learning optimization framework designed to accelerate neural network inference and training through advanced quantization techniques within the PyTorch ecosystem. The project focuses on reducing the computational and memory footprint of models by converting floating-point representations into lower-precision formats while preserving performance. It provides tools for experimenting with different quantization strategies, enabling developers to balance...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 12
    Norfair

    Norfair

    Lightweight Python library for adding real-time multi-object tracking

    ...It can easily be inserted into complex video processing pipelines to add tracking to existing projects. At the same time, it is possible to build a video inference loop from scratch using just Norfair and a detector. Supports moving camera, re-identification with appearance embeddings, and n-dimensional object tracking. Norfair provides several predefined distance functions to compare tracked objects and detections. The distance functions can also be defined by the user, enabling the implementation of different tracking strategies.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Transformers

    Transformers

    State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX

    Hugging Face Transformers provides APIs and tools to easily download and train state-of-the-art pre-trained models. Using pre-trained models can reduce your compute costs, carbon footprint, and save you the time and resources required to train a model from scratch. These models support common tasks in different modalities. Text, for tasks like text classification, information extraction, question answering, summarization, translation, text generation, in over 100 languages. Images, for tasks like image classification, object detection, and segmentation. Audio, for tasks like speech recognition and audio classification. ...
    Downloads: 23 This Week
    Last Update:
    See Project
  • 14
    Vision Transformer Pytorch

    Vision Transformer Pytorch

    Implementation of Vision Transformer, a simple way to achieve SOTA

    This repository provides a from-scratch, minimalist implementation of the Vision Transformer (ViT) in PyTorch, focusing on the core architectural pieces needed for image classification. It breaks down the model into patch embedding, positional encoding, multi-head self-attention, feed-forward blocks, and a classification head so you can understand each component in isolation.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 15
    HolyClaude

    HolyClaude

    AI coding workstation: Claude Code + web UI + 5 AI CLIs + headless

    ...HolyClaude emphasizes productivity by reducing friction in iterative development cycles, allowing users to refine outputs quickly without repeatedly crafting instructions from scratch. The toolkit is modular in nature, enabling developers to adapt it to different coding scenarios or integrate it into their existing workflows. It also reflects broader trends in AI-assisted development, where prompt design becomes a critical factor in output quality.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 16
    Argilla

    Argilla

    The open-source data curation platform for LLMs

    Argilla is a production-ready framework for building and improving datasets for NLP projects. Deploy your own Argilla Server on Spaces with a few clicks. Use embeddings to find the most similar records with the UI. This feature uses vector search combined with traditional search (keyword and filter based). Argilla is free, open-source, and 100% compatible with major NLP libraries (Hugging Face transformers, spaCy, Stanford Stanza, Flair, etc.). In fact, you can use and combine your preferred...
    Downloads: 9 This Week
    Last Update:
    See Project
  • 17
    MiniMind-V

    MiniMind-V

    "Big Model" trains a visual multimodal VLM with 26M parameters

    MiniMind-V is an experimental open-source project that aims to train a very small multimodal vision–language model (VLM) from scratch with extremely low compute and cost, making research and experimentation accessible to more people. The repository showcases training workflows and code designed to produce a 26-million parameter model—including both image and text capabilities—using minimal resources in very little time, reflecting a trend toward democratizing AI research.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 18
    AudioCraft

    AudioCraft

    Audiocraft is a library for audio processing and generation

    ...The repo provides inference scripts, checkpoints, and simple Python APIs so you can generate clips from prompts or incorporate the models into applications. It also contains training code and recipes, so researchers can fine-tune on custom data or explore new objectives without building infrastructure from scratch. Example notebooks, CLI tools, and audio utilities help with prompt design, conditioning on reference audio, and post-processing to produce ready-to-share outputs.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 19
    xFormers

    xFormers

    Hackable and optimized Transformers building blocks

    ...The library includes memory-efficient operator implementations in both Python and optimized C++/CUDA, ensuring that performance isn’t sacrificed for modularity. It also integrates with PyTorch seamlessly so you can drop in its blocks to existing models, replace default attention layers, or build new architectures from scratch. xformers includes training, deployment, and memory profiling tools.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 20
    EmoLLM

    EmoLLM

    Pre & Post-training & Dataset & Evaluation & Depoly & RAG

    ...The project is designed to help users through mental health conversations and has been fine-tuned from existing instruction-following LLMs rather than built as a base model from scratch. Its repository includes multiple model variants and training configurations spanning several underlying model families, including InternLM, Qwen, DeepSeek, Mixtral, LLaMA, and others, which shows that the initiative is structured as a broad ecosystem rather than a single release. The project also covers more than just model weights, with material for datasets, fine-tuning, evaluation, deployment, demos, RAG, and related subprojects such as its psychological digital assistant work.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 21
    TorchCode

    TorchCode

    Practice implementing softmax, attention, GPT-2 and more

    TorchCode is an interactive learning and practice platform designed to help developers master PyTorch by implementing core machine learning operations and architectures from scratch. It is structured similarly to competitive programming platforms like LeetCode but focuses specifically on tensor operations and deep learning concepts. The platform provides a collection of curated problems that cover fundamental topics such as activation functions, normalization layers, attention mechanisms, and full transformer architectures. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    LLM Foundry

    LLM Foundry

    LLM training code for MosaicML foundation models

    Introducing MPT-7B, the first entry in our MosaicML Foundation Series. MPT-7B is a transformer trained from scratch on 1T tokens of text and code. It is open source, available for commercial use, and matches the quality of LLaMA-7B. MPT-7B was trained on the MosaicML platform in 9.5 days with zero human intervention at a cost of ~$200k. Large language models (LLMs) are changing the world, but for those outside well-resourced industry labs, it can be extremely difficult to train and deploy these models. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 23
    PySpur

    PySpur

    Visual tool for building, testing, and deploying AI agent workflows

    ...By offering a visual representation of workflows, PySpur makes it easier to debug interactions between components and identify failures in complex pipelines. It supports iterative experimentation, allowing developers to rapidly improve agents without rebuilding systems from scratch. PySpur also enables deployment of finalized workflows after testing, making it suitable for both development and production use. Overall, it acts as an integrated environment for designing, evaluating, and managing AI-driven processes.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 24
    OpenFold

    OpenFold

    Trainable, memory-efficient, and GPU-friendly PyTorch reproduction

    ...In cases where the Nature paper differs from the source, we always defer to the latter. OpenFold is trainable in full precision, half precision, or bfloat16 with or without DeepSpeed, and we've trained it from scratch, matching the performance of the original. We've publicly released model weights and our training data — some 400,000 MSAs and PDB70 template hit files — under a permissive license. Model weights are available via scripts in this repository while the MSAs are hosted by the Registry of Open Data on AWS (RODA).
    Downloads: 2 This Week
    Last Update:
    See Project
  • 25
    AI Agent Deep Dive

    AI Agent Deep Dive

    AI Agent Source Code Deep Research Report

    AI Agent Deep Dive is a comprehensive educational repository designed to provide a deep and structured understanding of how modern AI agents work, focusing on architecture, workflows, and real-world implementation patterns. It breaks down complex concepts such as planning, tool usage, memory management, and multi-step reasoning into digestible explanations and practical examples. The project is organized as a learning resource rather than a standalone framework, making it particularly useful...
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • Next
MongoDB Logo MongoDB