Showing 51 open source projects for "parallel language"

View related business solutions
  • We help you deliver Virtual and Hybrid Events using our Award Winning end-to-end Event Management Platform Icon
    We help you deliver Virtual and Hybrid Events using our Award Winning end-to-end Event Management Platform

    Designed by event planners for event planners, the EventsAIR platform gives you the ability to manage your event, conference, meeting or function with

    EventsAIR have been anticipating and responding to the ever-changing event industry needs for over 30 years, providing innovative solutions that empower event organizers to create successful events around the globe.
    Learn More
  • Network Performance Monitoring | Statseeker Icon
    Network Performance Monitoring | Statseeker

    Statseeker is a powerful network performance monitoring solution for businesses

    Using just a single server or virtual machine, Statseeker can be up and running within minutes, and discovering your entire network in less than an hour, without any significant effect on your bandwidth availability.
    Learn More
  • 1
    DeepEval
    DeepEval is a simple-to-use, open-source LLM evaluation framework, for evaluating and testing large-language model systems. It is similar to Pytest but specialized for unit testing LLM outputs. DeepEval incorporates the latest research to evaluate LLM outputs based on metrics such as G-Eval, hallucination, answer relevancy, RAGAS, etc., which uses LLMs and various other NLP models that run locally on your machine for evaluation. Whether your application is implemented via RAG or fine-tuning,...
    Downloads: 4 This Week
    Last Update:
    See Project
  • 2
    fairseq2

    fairseq2

    FAIR Sequence Modeling Toolkit 2

    ...It supports multi-GPU and multi-node distributed training using DDP, FSDP, and tensor parallelism, capable of scaling up to 70B+ parameter models. The framework integrates seamlessly with PyTorch 2.x features such as torch.compile, Fully Sharded Data Parallel (FSDP), and modern configuration management.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 3
    BeeAI Framework

    BeeAI Framework

    Build production-ready AI agents in both Python and Typescript

    ...The framework supports both Python and TypeScript with full feature parity, making it accessible to a wide range of developers and teams. It includes a unified backend layer that connects seamlessly to multiple large language model providers, allowing flexible deployment across different AI infrastructures without vendor lock-in. BeeAI also provides orchestration tools for designing dynamic workflows, enabling multiple agents to coordinate tasks through structured execution flows, retries, and parallel processing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    higgsfield

    higgsfield

    Fault-tolerant, highly scalable GPU orchestration

    Higgsfield is an open-source, fault-tolerant, highly scalable GPU orchestration, and a machine learning framework designed for training models with billions to trillions of parameters, such as Large Language Models (LLMs).
    Downloads: 5 This Week
    Last Update:
    See Project
  • Planview is the leading end-to-end platform for Strategic Portfolio Management (SPM) and Digital Product Development (DPD) Icon
    Planview is the leading end-to-end platform for Strategic Portfolio Management (SPM) and Digital Product Development (DPD)

    Manage project and product portfolios enterprise-wide

    Planview AdaptiveWork (formerly Clarizen) with embedded AI helps you proactively plan and deliver any type and size of portfolio, project, and work. Gain AI-enhanced visibility and insights, drive collaboration, and achieve better business outcomes across your organization.
    Learn More
  • 5
    Functionary

    Functionary

    Chat language model that can use tools and interpret the results

    Functionary is an open-source large language model specifically designed for interpreting and executing structured functions or external tools within conversational AI systems. The model extends traditional chat-based language models by enabling them to determine when external functions should be called and how to extract the necessary parameters from natural language input.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    YiVal

    YiVal

    Your Automatic Prompt Engineering Assistant for GenAI Applications

    YiVal is an open-source framework designed to automate prompt engineering and evaluation workflows for generative AI applications, enabling developers to systematically improve the performance of large language models. It focuses on experimentation and optimization by allowing users to test multiple prompt variations, configurations, and model parameters in parallel, then evaluate their outputs using structured metrics and scoring systems. The platform is particularly useful in production environments where prompt quality directly impacts user experience, as it provides a repeatable and data-driven approach to refining prompts rather than relying on manual trial and error. ...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    Graph of Thoughts

    Graph of Thoughts

    Official Implementation of "Graph of Thoughts

    ...The framework executes these operations using a large language model as the reasoning engine while evaluating intermediate results to guide the search process. This approach enables models to explore multiple reasoning strategies in parallel and choose the most promising solutions during problem solving.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    Petals

    Petals

    Run 100B+ language models at home, BitTorrent-style

    Parallel inference reaches hundreds of tokens/sec. Beyond classic language model APIs — you can employ any fine-tuning and sampling methods, execute custom paths through the model, or see its hidden states. You get the comforts of an API with the flexibility of PyTorch. You can also host BLOOMZ, a version of BLOOM fine-tuned to follow human instructions in the zero-shot regime — just replace bloom-petals with bloomz-petals.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 9
    Medusa

    Medusa

    Framework for Accelerating LLM Generation with Multiple Decoding Heads

    Medusa is a framework aimed at accelerating the generation capabilities of Large Language Models (LLMs) by employing multiple decoding heads. This approach allows for parallel processing during text generation, significantly enhancing throughput and reducing response times. Medusa is designed to be simple to implement and integrates with existing LLM infrastructures, making it a practical solution for scaling LLM applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Intelligent Appointment Reminders Icon
    Intelligent Appointment Reminders

    For doctors, clinics and hospitals

    DoctorConnect provides industry leading patient engagement. In business for over 25 years, we provide highly customizable services to thousands of doctors, clinics and hospitals. Appointment Reminders, After Care Surveys, Automated No-Show and Recall Messaging, and more. We can directly interface with hundreds of EMR and PM systems. We'd love to hear from you and show you how we increase your revenue and your patient satisfaction.
    Get a Demo
  • 10
    GPT-NeoX

    GPT-NeoX

    Implementation of model parallel autoregressive transformers on GPUs

    This repository records EleutherAI's library for training large-scale language models on GPUs. Our current framework is based on NVIDIA's Megatron Language Model and has been augmented with techniques from DeepSpeed as well as some novel optimizations. We aim to make this repo a centralized and accessible place to gather techniques for training large-scale autoregressive language models, and accelerate research into large-scale training. For those looking for a TPU-centric codebase, we...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    TextBox

    TextBox

    A text generation library with pre-trained language models github.com

    TextBox 2.0 is an up-to-date text generation library based on Python and PyTorch focusing on building a unified and standardized pipeline for applying pre-trained language models to text generation. From a task perspective, we consider 13 common text generation tasks such as translation, story generation, and style transfer, and their corresponding 83 widely-used datasets. From a model perspective, we incorporate 47 pre-trained language models/modules covering the categories of general, translation, Chinese, dialogue, controllable, distilled, prompting, and lightweight models (modules). ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Fairseq

    Fairseq

    Facebook AI Research Sequence-to-Sequence Toolkit written in Python

    Fairseq(-py) is a sequence modeling toolkit that allows researchers and developers to train custom models for translation, summarization, language modeling and other text generation tasks. We provide reference implementations of various sequence modeling papers. Recent work by Microsoft and Google has shown that data parallel training can be made significantly more efficient by sharding the model parameters and optimizer state across data parallel workers. These ideas are encapsulated in the new FullyShardedDataParallel (FSDP) wrapper provided by fairscale. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    GPT Neo

    GPT Neo

    An implementation of model parallel GPT-2 and GPT-3-style models

    An implementation of model & data parallel GPT3-like models using the mesh-tensorflow library. If you're just here to play with our pre-trained models, we strongly recommend you try out the HuggingFace Transformer integration. Training and inference is officially supported on TPU and should work on GPU as well. This repository will be (mostly) archived as we move focus to our GPU-specific repo, GPT-NeoX. NB, while neo can technically run a training step at 200B+ parameters, it is very...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 14
    FARM

    FARM

    Fast & easy transfer learning for NLP

    ...With FARM you can build fast proofs-of-concept for tasks like text classification, NER or question answering and transfer them easily into production. Easy fine-tuning of language models to your task and domain language. AMP optimizers (~35% faster) and parallel preprocessing (16 CPU cores => ~16x faster). Modular design of language models and prediction heads. Switch between heads or combine them for multitask learning. Full Compatibility with HuggingFace Transformers' models and model hub. Smooth upgrading to newer language models. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    XLM (Cross-lingual Language Model)

    XLM (Cross-lingual Language Model)

    PyTorch original implementation of Cross-lingual Language Model

    XLM (Cross-lingual Language Model) is a family of multilingual pretraining methods that align representations across languages to enable strong zero-shot transfer. It popularized objectives like Masked Language Modeling (MLM) across many languages and Translation Language Modeling (TLM) that jointly trains on parallel sentence pairs to tighten cross-lingual alignment.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Weld

    Weld

    High-performance runtime for data analytics applications

    Weld is a programming language and runtime designed to improve the performance of data-intensive applications by optimizing computations across multiple libraries. Instead of optimizing individual functions independently, Weld introduces an intermediate representation that allows different frameworks to share optimization opportunities. This approach reduces data movement between libraries and enables the system to generate highly optimized machine code for parallel execution. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    OpenSeq2Seq

    OpenSeq2Seq

    Toolkit for efficient experimentation with Speech Recognition

    ...Its core goal is to give researchers a flexible, modular framework for building and training encoder–decoder architectures while fully leveraging distributed and mixed-precision training. The toolkit includes ready-made models for neural machine translation, automatic speech recognition, speech synthesis, language modeling, and additional NLP tasks such as sentiment analysis. It supports multi-GPU and multi-node data-parallel training, and integrates with Horovod to scale out across large GPU clusters. Mixed-precision support (float16) is optimized for NVIDIA Volta and Turing GPUs, allowing significant speedups and memory savings without sacrificing model quality. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18

    popt4jlib

    Parallel Optimization Library for Java

    popt4jlib is an open-source parallel optimization library for the Java programming language supporting both shared memory and distributed message passing models. Implements a number of meta-heuristic algorithms for Non-Linear Programming, including Genetic Algorithms, Differential Evolution, Evolutionary Algorithms, Simulated Annealing, Particle Swarm Optimization, Firefly Algorithm, Monte-Carlo Search, Local Search algorithms, Gradient-Descent-based algorithms, as well as some well-known network flow and other graph algorithms. ...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19

    PADIC

    A multilingual Parallel Arabic DIalectal Corpus

    PADIC (Parallel Arabic DIalectal Corpus) is a multi-dialectal corpus built in the framework of the National Research Project "TORJMAN", led by Scientific and Technical Research Center for the Development of Arabic Language and funded by the Algerian Ministry of Higher Education and Scientific Research. PADIC is composed of 6 dialects: two Algerian dialects (Algiers and Annaba cities), Palestinian, Syrian, Tunisian, Moroccan) and MSA.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 20
    Osman Arabic Text Readability

    Osman Arabic Text Readability

    Open Source tool for Arabic text readability

    We present OSMAN (Open Source Metric for Measuring Arabic Narratives) - a novel open source Arabic readability metric and tool. The open source Java tool allows users to calculate readability for Arabic text (with and without diacritics). The tool provides methods to split the text into words and sentence, count syllables, Faseeh letters, hard and complex words in addition to adding diacritics (vocalise text). This makes the tool useful for researchers and educators working with Arabic text....
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    HLearn

    HLearn

    Homomorphic machine learning

    HLearn is a Haskell-based machine learning library focused on composability, algebraic structure, and performance. It provides a functional approach to building machine learning algorithms by leveraging algebraic properties such as monoids and groups. This allows for parallel, incremental, and distributed computation in a mathematically consistent way. HLearn aims to provide implementations of common algorithms like k-means, naive Bayes, and others while maintaining the expressiveness and safety of the Haskell language.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22

    English-Khmer S. Machine Translation

    English-Khmer Automatic Statistic Machine Translation (SMT)

    Automatic Machine Translation from English to Khmer project is the first effort in Natural Language Processing field for translating English to Khmer (Cambodian) language. This project uses Domy CE, an open source SMT toolkit, for training parallel corpus and web technologies such as Python, Apache2, HTML, XML, and XSLT for developing web-based application. This project is developed by Ms. Kim Sokphyrum (DU) and Ms. Suos Samak (Jamia), under Supervision of Mr. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    CRFSharp

    CRFSharp

    CRFSharp is a .NET(C#) implementation of Conditional Random Field

    CRFSharp(aka CRF#) is a .NET(C#) implementation of Conditional Random Fields, an machine learning algorithm for learning from labeled sequences of examples. It is widely used in Natural Language Process (NLP) tasks, for example: word breaker, postagging, named entity recognized, query chunking and so on. CRF#'s mainly algorithm is the same as CRF++ written by Taku Kudo. It encodes model parameters by L-BFGS. Moreover, it has many significant improvement than CRF++, such as totally parallel encoding, optimizing memory usage and so on. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Simdist lets you harness the power of cluster computing without any knowledge of parallel libraries such as MPI, and with no restrictions on programming language. Primarily targeted at evolutionary computing and similar master-slave configurations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Sanchay
    Sanchay is a collection of tools and APIs for language researchers. It has some implementations of NLP algorithms, some flexible APIs, several user friendly annotation interfaces and Sanchay Query Language for language resources.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB