Showing 3 open source projects for "parallel language"

View related business solutions
  • Peer to Peer Recognition Brings Teams Together Icon
    Peer to Peer Recognition Brings Teams Together

    The modern employee engagement platform for the modern workforce

    Create a positive and energetic workplace environment with Motivosity, an innovative employee recognition and engagement platform. With Motivosity, employees can give each other small monetary bonuses for doing great things, promoting trust, collaboration, and appreciation in the workplace. The software solution comes with features such as an open-currency open-reward system, insights and analytics, dynamic organization chart, award programs, milestones, and more.
    Learn More
  • The most user-friendly sales commission software for revenue-focused teams Icon
    The most user-friendly sales commission software for revenue-focused teams

    Everstage is a trusted ICM for public companies and enterprises worldwide-across industries

    Rated as #1 sales compensation management software, Everstage helps businesses streamline commission administration, boost sales performance and improve ROI with actionable insights. Top features: No-code plan designer, detailed commission statements, advanced commission forecasting, quota management, queries & approval workflows, deferred commissions (ASC606), BI-powered reporting, and more.
    Learn More
  • 1
    Megatron

    Megatron

    Ongoing research training transformer models at scale

    Megatron is a large, powerful transformer developed by the Applied Deep Learning Research team at NVIDIA. This repository is for ongoing research on training large transformer language models at scale. We developed efficient, model-parallel (tensor, sequence, and pipeline), and multi-node pre-training of transformer based models such as GPT, BERT, and T5 using mixed precision. Megatron is also used in NeMo Megatron, a framework to help enterprises overcome the challenges of building and training sophisticated natural language processing models with billions and trillions of parameters. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Colossal-AI

    Colossal-AI

    Making large AI models cheaper, faster and more accessible

    ...However, distributed training, especially model parallelism, often requires domain expertise in computer systems and architecture. It remains a challenge for AI researchers to implement complex distributed training solutions for their models. Colossal-AI provides a collection of parallel components for you. We aim to support you to write your distributed deep learning models just like how you write your model on your laptop.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    GPT-NeoX

    GPT-NeoX

    Implementation of model parallel autoregressive transformers on GPUs

    This repository records EleutherAI's library for training large-scale language models on GPUs. Our current framework is based on NVIDIA's Megatron Language Model and has been augmented with techniques from DeepSpeed as well as some novel optimizations. We aim to make this repo a centralized and accessible place to gather techniques for training large-scale autoregressive language models, and accelerate research into large-scale training. For those looking for a TPU-centric codebase, we...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB