Showing 115 open source projects for "transformers"

View related business solutions
  • Peer to Peer Recognition Brings Teams Together Icon
    Peer to Peer Recognition Brings Teams Together

    The modern employee engagement platform for the modern workforce

    Create a positive and energetic workplace environment with Motivosity, an innovative employee recognition and engagement platform. With Motivosity, employees can give each other small monetary bonuses for doing great things, promoting trust, collaboration, and appreciation in the workplace. The software solution comes with features such as an open-currency open-reward system, insights and analytics, dynamic organization chart, award programs, milestones, and more.
    Learn More
  • Paladin Point of Sale is a powerful retail management system designed to simplify daily store operations for independent retailers. Icon
    Paladin Point of Sale is a powerful retail management system designed to simplify daily store operations for independent retailers.

    It enables businesses to sell from anywhere using mobile point-of-sale tools while also providing a ready-built online store for seamless omnichannel

    Paladin Point of Sale is ideal for independent retailers in hardware, lumber, pharmacy, and specialty retail industries seeking an easy-to-use, industry-specific POS system with strong support and flexibility.
    Learn More
  • 1
    KoGPT

    KoGPT

    KakaoBrain KoGPT (Korean Generative Pre-trained Transformer)

    KoGPT is a Korean language model based on OpenAI’s GPT architecture, designed for various natural language processing (NLP) tasks such as text generation, summarization, and dialogue systems.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 2
    FARM

    FARM

    Fast & easy transfer learning for NLP

    ...AMP optimizers (~35% faster) and parallel preprocessing (16 CPU cores => ~16x faster). Modular design of language models and prediction heads. Switch between heads or combine them for multitask learning. Full Compatibility with HuggingFace Transformers' models and model hub. Smooth upgrading to newer language models. Integration of custom datasets via Processor class. Powerful experiment tracking & execution.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    jiant

    jiant

    jiant is an nlp toolkit

    Jiant is a multitask NLP framework for fine-tuning transformer-based models on multiple natural language understanding (NLU) tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    SimCSE

    SimCSE

    SimCSE: Simple Contrastive Learning of Sentence Embeddings

    SimCSE (Simple Contrastive Learning of Sentence Embeddings) is a machine learning framework for training sentence embeddings using contrastive learning. It improves representation learning for NLP tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • White Labeled Fintech Software Solutions | Centrex Icon
    White Labeled Fintech Software Solutions | Centrex

    Centrex is a full suite of white labeled fintech solutions built and designed for brokers, lenders, banks, investors, fintechs

    The Centrex products include: CRM, loan origination, loan and advance servicing software, syndication management, white labeled mobile app, money manager, underwriting, Esign, and website smart app builder. The Centrex services include: fintech software consulting, admin retainer services, and managed data cloud.
    Learn More
  • 5
    Tabnine

    Tabnine

    Vim client for TabNine

    Tabnine is an AI-powered code completion extension trusted by millions of developers around the world. Whether you’re just getting started as a developer or if you’ve been doing it for decades, Tabnine will help you code twice as fast with half the keystrokes – all in your favorite IDE. Whether you call it IntelliSense, intelliCode, autocomplete, AI-assisted code completion, AI-powered code completion, AI copilot, AI code snippets, code suggestion, code prediction, code hinting, or...
    Downloads: 21 This Week
    Last Update:
    See Project
  • 6
    TextBrewer

    TextBrewer

    A PyTorch-based knowledge distillation toolkit

    TextBrewer is a PyTorch-based model distillation toolkit for natural language processing. It includes various distillation techniques from both NLP and CV field and provides an easy-to-use distillation framework, which allows users to quickly experiment with the state-of-the-art distillation methods to compress the model with a relatively small sacrifice in the performance, increasing the inference speed and reducing the memory usage.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Image GPT

    Image GPT

    Large-scale autoregressive pixel model for image generation by OpenAI

    ...Researchers can use the code to sample new images, evaluate generative loss on datasets like ImageNet or CIFAR-10, and explore the impact of scaling on performance. While the repository is archived and provided as-is, it remains a valuable starting point for experimenting with autoregressive transformers applied directly to raw pixel data. By demonstrating GPT’s flexibility across modalities, Image-GPT influenced subsequent multimodal generative research.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 8
    Sparse Attention

    Sparse Attention

    "Generating Long Sequences with Sparse Transformers" examples

    Sparse Attention is OpenAI’s code release for the Sparse Transformer model, introduced in the paper Generating Long Sequences with Sparse Transformers. It explores how modifying the self-attention mechanism with sparse patterns can reduce the quadratic scaling of standard transformers, making it possible to model much longer sequences efficiently. The repository provides implementations of sparse attention layers, training code, and evaluation scripts for benchmark datasets. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    TFKit

    TFKit

    Handling multiple nlp task in one pipeline

    TFKit is a tool kit mainly for language generation. It leverages the use of transformers on many tasks with different models in this all-in-one framework. All you need is a little change of config. You can use tfkit for model training and evaluation with tfkit-train and tfkit-eval. The key to combine different task together is to make different task with same data format. All data will be in csv format - tfkit will use csv for all task, normally it will have two columns, first columns is the input of models, the second column is the output of models. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Project Planning and Management Software | Planview Icon
    Project Planning and Management Software | Planview

    For Enterprise PMOs

    Planview® ProjectAdvantage (formerly Sciforma) is an enterprise-centric project and portfolio management (PPM) software designed to enable change, drive innovation, and lead in a company's digital transformation. With ProjectAdvantage, teams can strategically track and monitor project data in order to make relevant decisions. It offers multiple features focused on strategic management, functional management, and execution management. A highly scalable and cost-effective solution, ProjectAdvantage is available in various deployment models.
    Learn More
  • 10
    DETR

    DETR

    End-to-end object detection with transformers

    PyTorch training code and pretrained models for DETR (DEtection TRansformer). We replace the full complex hand-crafted object detection pipeline with a Transformer, and match Faster R-CNN with a ResNet-50, obtaining 42 AP on COCO using half the computation power (FLOPs) and the same number of parameters. Inference in 50 lines of PyTorch. What it is. Unlike traditional computer vision techniques, DETR approaches object detection as a direct set prediction problem. It consists of a set-based...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    PixelCNN

    PixelCNN

    Code for the paper "PixelCNN++: A PixelCNN Implementation..."

    ...The project serves as both a research reference and a practical framework for experimenting with autoregressive generative models. Although archived, PixelCNN has influenced a wide range of later work in generative modeling, including advancements in image transformers and diffusion models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 12
    BlockSparse

    BlockSparse

    Efficient GPU kernels for block-sparse matrix multiplication

    ...The idea is to exploit block-level sparsity — i.e. treat matrices or weight tensors as composed of blocks, many of which may be zero or unused — to save compute and memory when sparsity patterns are structured. This is particularly useful in models like Sparse Transformers, where attention matrices or intermediate layers may adopt block-sparse patterns to scale better. The repo implements both blocksparse and blockwise convolution/transpose-convolution primitives, with support for preparing, executing, and verifying those ops on NVIDIA GPUs. In addition to low-level kernels, it includes wrapper code for integrating with TensorFlow, example scripts (e.g. a transformer on the enwik8 dataset), transformer logic that uses blocksparse operations, and debugging helpers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    XML and Python-based webserving framework, featuring XSLT, embedded Python evaluation, pipeline processing model and caching. Works with several XSL transformers and webserver interfaces.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    ZopeXMLMethods provides methods to apply to Zope objects for XML/XSLT processing. XSLTMethod associates XSLT transformers with XML documents. ZopeXMLMethods succeeds XMLTransform. It features file-system caching and works with many XML/XSLT libraries.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    gpt-oss-20b

    gpt-oss-20b

    OpenAI’s compact 20B open model for fast, agentic, and local use

    ...Like its larger sibling (gpt-oss-120b), it offers adjustable reasoning depth and full chain-of-thought visibility for better interpretability. It’s released under a permissive Apache 2.0 license, allowing unrestricted commercial and research use. GPT-OSS-20B is compatible with Transformers, vLLM, Ollama, PyTorch, and other tools. It is ideal for developers building lightweight AI agents or experimenting with fine-tuning on consumer-grade hardware.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB