Search Results for "learning vector quantization"

65 projects for "learning vector quantization" with 1 filter applied:

  • Complete Data Management for Nonprofits Icon
    Complete Data Management for Nonprofits

    Designed to fit with multi-level non-profit organization, across any sector

    NewOrg is a robust platform built with enhanced features to help non-profit organizations that capture and integrate the information from all of their operational areas to better manage volunteers, clients, programs, outcome reporting, activity sign-ups & scheduling, communications, surveys, fundraising activities and Development campaigns. NewOrg can truly deliver an intuitive product that will help manage your Committees, Donors, Events, and Memberships so that the organization runs efficiently.
    Learn More
  • The top-rated AI recruiting platform for faster, smarter hiring. Icon
    The top-rated AI recruiting platform for faster, smarter hiring.

    Humanly is an AI recruiting platform that automates candidate conversations, screening, and scheduling.

    Humanly is an AI-first recruiting platform that helps talent teams hire in days, not months—without adding headcount. Our intuitive CRM pairs with powerful agentic AI to engage and screen every candidate instantly, surfacing top talent fast. Built on insights from over 4 million candidate interactions, Humanly delivers speed, structure, and consistency at scale—engaging 100% of interested candidates and driving pipeline growth through targeted outreach and smart re-engagement. We integrate seamlessly with all major ATSs to reduce manual work, improve data flow, and enhance recruiter efficiency and candidate experience. Independent audits ensure our AI remains fair and bias-free, so you can hire confidently.
    Learn More
  • 1
    TurboQuant+

    TurboQuant+

    Implementation of TurboQuant (ICLR 2026)

    ...It is designed to be used in conjunction with modern machine learning workflows, particularly those involving large models that require optimization for deployment. TurboQuant Plus focuses on experimentation and performance tuning, allowing developers to test different configurations and evaluate trade-offs. Its architecture supports extensibility, enabling further development of quantization methods and integration with existing ML pipelines.
    Downloads: 24 This Week
    Last Update:
    See Project
  • 2
    TurboQuant PyTorch

    TurboQuant PyTorch

    From-scratch PyTorch implementation of Google's TurboQuant

    TurboQuant PyTorch is a specialized deep learning optimization framework designed to accelerate neural network inference and training through advanced quantization techniques within the PyTorch ecosystem. The project focuses on reducing the computational and memory footprint of models by converting floating-point representations into lower-precision formats while preserving performance.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 3
    Machine learning basics

    Machine learning basics

    Plain python implementations of basic machine learning algorithms

    Machine learning basics repository is an educational project that provides plain Python implementations of fundamental machine learning algorithms designed to help learners understand how these methods work internally. Instead of relying on external machine learning libraries, the algorithms are implemented from scratch so that users can explore the mathematical logic and computational structure behind each technique.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    Machine learning algorithms

    Machine learning algorithms

    Minimal and clean examples of machine learning algorithms

    Machine learning algorithms is an open-source repository that provides minimal and clean implementations of machine learning algorithms written primarily in Python. The project focuses on demonstrating how fundamental machine learning methods work internally by implementing them from scratch rather than relying on high-level libraries. This approach allows learners to study the mathematical and algorithmic details behind widely used models in a transparent and readable way. The repository...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Self-hosted n8n: No-code AI workflows Icon
    Self-hosted n8n: No-code AI workflows

    Connect workflows. Integrate data

    A free-to-use workflow automation tool, n8n lets you connect all your apps and data in one customizable, no-code platform. Design workflows and process data from a simple, unified dashboard.
    Learn More
  • 5
    DLRM

    DLRM

    An implementation of a deep learning recommendation model (DLRM)

    ...The implementation is optimized for performance at scale, supporting multi-GPU and multi-node execution, quantization, embedding partitioning, and pipelined I/O to feed huge embeddings efficiently. It includes data loaders for standard benchmarks (like Criteo), training scripts, evaluation tools, and capabilities like mixed precision, gradient compression, and memory fusion to maximize throughput.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    ggml

    ggml

    Tensor library for machine learning

    ggml is an open-source tensor library designed for efficient machine learning computation with a focus on running models locally and with minimal dependencies. Written primarily in C and C++, the library provides low-level tensor operations and automatic differentiation that allow developers to implement machine learning algorithms and neural networks efficiently. The project emphasizes portability and performance, enabling machine learning inference across a wide range of hardware...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    MyScaleDB

    MyScaleDB

    A @ClickHouse fork that supports high-performance vector search

    MyScaleDB is an open-source SQL vector database designed for building large-scale AI and machine learning applications that require both analytical queries and semantic vector search. The system is built on top of the ClickHouse database engine and extends it with specialized indexing and search capabilities optimized for vector embeddings. This design allows developers to store structured data, unstructured text, and high-dimensional vector embeddings within a single database platform. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    SWIFT LLM

    SWIFT LLM

    Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs

    SWIFT LLM is a comprehensive framework developed within the ModelScope ecosystem for training, fine-tuning, evaluating, and deploying large language models and multimodal models. The platform provides a full machine learning pipeline that supports tasks ranging from model pre-training to reinforcement learning alignment techniques. It integrates with popular inference engines such as vLLM and LMDeploy to accelerate deployment and runtime performance. The framework also includes support for...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 9
    Tencent-Hunyuan-Large

    Tencent-Hunyuan-Large

    Open-source large language model family from Tencent Hunyuan

    Tencent-Hunyuan-Large is the flagship open-source large language model family from Tencent Hunyuan, offering both pre-trained and instruct (fine-tuned) variants. It is designed with long-context capabilities, quantization support, and high performance on benchmarks across general reasoning, mathematics, language understanding, and Chinese / multilingual tasks. It aims to provide competitive capability with efficient deployment and inference. FP8 quantization support to reduce memory usage...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Manage your hosting business with our vacation rental software Icon
    Manage your hosting business with our vacation rental software

    Empowering your short-term rental business to succeed

    Whether you’re a new or established business, you can rely on Lodgify’s vacation rental property management software for support through every step of your journey.
    Learn More
  • 10
    RuVector

    RuVector

    Self-Learning, Vector Graph Neural Network, and Database built in Rust

    RuVector is part of the broader rUv ecosystem of AI engineering tools and focuses on enabling advanced vector-based processing and intelligent system development within agentic and AI-driven pipelines. The project fits into a larger vision of modular, composable AI infrastructure designed to support autonomous agents, data retrieval, and intelligent automation workflows. It emphasizes extensibility and interoperability with modern AI stacks, allowing developers to integrate vector operations...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    hls4ml

    hls4ml

    Machine learning on FPGAs using HLS

    hls4ml is an open-source framework that enables machine learning models to be implemented directly on hardware such as FPGAs and ASICs using high-level synthesis techniques. The system converts trained neural network models from common machine learning frameworks into hardware description code suitable for ultra-low-latency inference. This approach allows machine learning algorithms to run directly on specialized hardware, making them suitable for applications that require extremely fast...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    whisper.cpp

    whisper.cpp

    Port of OpenAI's Whisper model in C/C++

    whisper.cpp is a lightweight, C/C++ reimplementation of OpenAI’s Whisper automatic speech recognition (ASR) model—designed for efficient, standalone transcription without external dependencies. The entire high-level implementation of the model is contained in whisper.h and whisper.cpp. The rest of the code is part of the ggml machine learning library. The command downloads the base.en model converted to custom ggml format and runs the inference on all .wav samples in the folder samples. whisper.cpp supports integer quantization of the Whisper ggml models. Quantized models require less memory and disk space and depending on the hardware can be processed more efficiently.
    Downloads: 358 This Week
    Last Update:
    See Project
  • 13
    MatMul-Free LM

    MatMul-Free LM

    Implementation for MatMul-free LM

    ...Since matrix multiplication is one of the most computationally expensive components of modern language models, the project explores alternative computational strategies that reduce hardware requirements while maintaining comparable performance. The architecture relies on quantization-aware training and lightweight operations to replace conventional dense matrix multiplications with more efficient alternatives. These optimizations can significantly reduce memory consumption and potentially improve computational efficiency during both training and inference. The repository provides implementations of models at several parameter scales and includes tools for experimenting with the architecture using modern machine learning frameworks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    BitNet

    BitNet

    BitNet: Scaling 1-bit Transformers for Large Language Models

    BitNet is a machine learning research implementation that explores extremely low-precision neural network architectures designed to dramatically reduce the computational cost of large language models. The project implements the BitNet architecture described in research on scaling transformer models using extremely low-bit quantization techniques. In this approach, neural network weights are quantized to approximately one bit per parameter, allowing models to operate with far lower memory usage than traditional 16-bit or 32-bit neural networks. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    Book4_Power-of-Matrix

    Book4_Power-of-Matrix

    Book_4_Matrix Power | The Iris Book: From Addition, Subtraction

    Book4_Power-of-Matrix is an open educational repository that forms part of the Visualize-ML book series, focusing on explaining matrix mathematics and linear algebra concepts through visual and intuitive methods. The project is designed to help readers progress from basic arithmetic toward machine learning fundamentals by building a strong conceptual understanding of vectors, matrices, and their operations. It combines explanatory text, diagrams, and Python examples to bridge theory and...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    LLM Course

    LLM Course

    Course to get into Large Language Models (LLMs)

    ...The materials also cover inference optimization and quantization to make serving LLMs feasible on commodity GPUs or even CPUs, which is crucial for side projects and startups. Evaluation is treated as a first-class topic, with examples of automatic and human-in-the-loop methods to catch regressions and verify quality beyond simple loss values. By the end, students have a mental model and a practical toolkit for iterating on datasets, training configs, etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    AgentGuide

    AgentGuide

    AI Agent Development Guide, LangGraph in Action, Advanced RAG

    AgentGuide is an open-source learning resource designed to provide a structured pathway for understanding and building AI agents. The project aggregates tutorials, research papers, frameworks, and practical resources related to agent development with large language models. Instead of presenting scattered resources, the repository organizes them into a systematic learning roadmap that guides learners from foundational concepts to advanced AI agent systems. The guide covers topics such as...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    spacy-transformers

    spacy-transformers

    Use pretrained transformers like BERT, XLNet and GPT-2 in spaCy

    spaCy supports a number of transfer and multi-task learning workflows that can often help improve your pipeline’s efficiency or accuracy. Transfer learning refers to techniques such as word vector tables and language model pretraining. These techniques can be used to import knowledge from raw text into your pipeline, so that your models are able to generalize better from your annotated examples.
    Downloads: 16 This Week
    Last Update:
    See Project
  • 19
    Earth Engine API

    Earth Engine API

    Python and JavaScript bindings for calling the Earth Engine API

    The Earth Engine API provides Python and JavaScript client libraries for Google Earth Engine, a planetary-scale geospatial analysis platform. With it, users compose lazy, server-side computations over massive catalogs of satellite imagery and vector datasets without handling raw files locally. The API exposes functional operators for map algebra, reducers, joins, and machine learning that scale transparently on Earth Engine’s backend. Developers authenticate once, work interactively in notebooks or the Code Editor, and export results to Cloud Storage, Drive, or asset collections. ...
    Downloads: 12 This Week
    Last Update:
    See Project
  • 20
    Korvus

    Korvus

    Korvus is a search SDK that unifies the entire RAG pipeline

    ...The project consolidates the typical steps of a RAG pipeline—including embedding generation, document retrieval, reranking, and text generation—into a single query executed within the Postgres ecosystem. By leveraging PostgresML and vector extensions such as pgvector, Korvus eliminates the need for external microservices typically used for AI search architectures, reducing both system complexity and latency. The architecture enables machine learning operations to occur directly in the database, minimizing data transfer between services and improving overall performance for large datasets.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    RAG from Scratch

    RAG from Scratch

    Demystify RAG by building it from scratch

    ...Instead of relying on complex frameworks or cloud services, the repository demonstrates the entire RAG pipeline using transparent and minimal implementations. The project walks through key concepts such as generating embeddings, building vector databases, retrieving relevant documents, and integrating the retrieved context into language model prompts. Each example is written with detailed explanations so that developers can understand the internal mechanics of semantic search and context-aware language generation. The repository emphasizes learning through direct implementation, allowing users to see how each component of the RAG architecture functions independently.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Supermemory

    Supermemory

    Memory engine and app that is extremely fast, scalable

    Supermemory is an ambitious and extensible AI-powered personal knowledge management system that aims to help users capture, organize, retrieve, and reason over information in a manner that mimics human memory structures. The platform allows individuals to ingest text, documents, and other content forms, then uses advanced retrieval and embedding techniques to index and relate information intelligently so that users can recall relevant knowledge in context rather than just by keyword match....
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    FastDeploy

    FastDeploy

    High-performance Inference and Deployment Toolkit for LLMs and VLMs

    FastDeploy is an open-source inference and deployment toolkit designed to simplify the process of running and serving deep learning models across a wide range of hardware platforms. Developed within the PaddlePaddle ecosystem, the toolkit focuses on providing high-performance deployment capabilities for modern AI models including large language models and vision-language systems. The platform enables developers to deploy trained models quickly using optimized inference pipelines that support...
    Downloads: 3 This Week
    Last Update:
    See Project
  • 24
    WavTokenizer

    WavTokenizer

    SOTA discrete acoustic codec models with 40/75 tokens per second

    ...The model uses a single-quantizer design together with temporal compression to achieve extreme compression without sacrificing reconstruction fidelity. Its architecture incorporates a broader vector-quantization space, extended contextual windows, and improved attention networks, combined with multi-scale discriminators and inverse Fourier transform blocks to enhance waveform reconstruction. Extensive experiments show that WavTokenizer matches or surpasses previous neural codecs across speech, music, and general audio on both objective metrics and subjective listening tests.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Hindsight

    Hindsight

    Hindsight: Agent Memory That Learns

    Hindsight is an advanced, open-source memory system for AI agents designed to enable long-term learning, reasoning, and consistency across interactions by treating memory as a first-class component of intelligence rather than a simple retrieval layer. It addresses one of the core limitations of modern AI agents, which is their inability to retain and meaningfully use past experiences over time, by introducing a structured, biomimetic memory architecture inspired by how human memory works. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • Next
MongoDB Logo MongoDB