Showing 139 open source projects for "parallel language"

View related business solutions
  • Professional Streaming and Video Hosting - GDPR Compliant - 3Q Icon
    Professional Streaming and Video Hosting - GDPR Compliant - 3Q

    Secure hosting, scalable streaming, and easy integration for internal and external communications

    3Q offers a multifunctional video platform for hosting, managing and distributing video and audio content on all channels. Live and on-demand.
    Learn More
  • HR Outsourcing Built for Small and Midsize Businesses Icon
    HR Outsourcing Built for Small and Midsize Businesses

    Payroll. Benefits. Compliance. Technology. All in one place.

    TriNet is a leading provider of HR outsourcing solutions built for small and midsize businesses. Its platform combines payroll, benefits, risk management, compliance, and HR technology in one integrated system. Through its PEO (Professional Employer Organization) and HR Plus (ASO) offerings, TriNet helps companies streamline HR administration, stay compliant, and access enterprise-level benefits. Businesses can run payroll efficiently, manage compliance with complex state and federal regulations, and offer competitive employee benefits with ease. The company’s intuitive HR platform also automates time tracking, leave requests, and onboarding. With TriNet, organizations can focus on growth while ensuring their people and processes are supported by expert HR guidance.
    Learn More
  • 1
    Large Language Models (LLMs)

    Large Language Models (LLMs)

    Connect MATLAB to LLM APIs, including OpenAI® Chat Completions

    This repository enables MATLAB to connect with large language models (LLMs) such as OpenAI's ChatGPT, DALL-E, Azure OpenAI, and Ollama, integrating their natural language processing and image generation capabilities directly within MATLAB environments. It facilitates creating chatbots, summarizing text, and image generation, among other tasks.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 2
    The Julia Programming Language

    The Julia Programming Language

    High-level, high-performance dynamic language for technical computing

    Julia is a fast, open source high-performance dynamic language for technical computing. It can be used for data visualization and plotting, deep learning, machine learning, scientific computing, parallel computing and so much more. Having a high level syntax, Julia is easy to use for programmers of every level and background. Julia has more than 2,800 community-registered packages including various mathematical libraries, data manipulation tools, and packages for general purpose computing. ...
    Downloads: 19 This Week
    Last Update:
    See Project
  • 3
    multi-agent-shogun

    multi-agent-shogun

    Samurai-inspired multi-agent system for Claude Code

    multi-agent-shogun is a multi-agent orchestration system designed to coordinate multiple AI coding agents working in parallel. Inspired by the hierarchy of a feudal Japanese military structure, the system organizes agents into roles such as Shogun, Karo, and Ashigaru, which correspond to strategist, coordinator, and worker agents. A user interacts primarily with the Shogun agent by issuing natural language instructions that describe the desired tasks.
    Downloads: 10 This Week
    Last Update:
    See Project
  • 4
    Apache Beam

    Apache Beam

    Unified programming model for Batch and Streaming

    Apache Beam is an open source, unified programming model to define both batch and streaming data-parallel processing pipelines, as well as certain language-specific SDKs for constructing pipelines and Runners. These pipelines are executed on one of Beam’s supported distributed processing back-ends, which include Apache Apex, Apache Flink, Apache Spark, and Google Cloud Dataflow. Beam is especially useful for Embarrassingly Parallel data processing tasks, and caters to the different needs and backgrounds of end users, SDK writers and runner writers.
    Downloads: 1 This Week
    Last Update:
    See Project
  • All-in-One Mental Health EHR Icon
    All-in-One Mental Health EHR

    Simplify your systems. Strengthen your cash flow. Start fresh with Ensora Health

    Ensora Health’s Mental Health EHR is designed for mental health professionals, therapists, and practice managers looking for a secure, user-friendly solution to streamline administrative tasks and improve efficiency in their practice management
    Learn More
  • 5
    MiroFish

    MiroFish

    A Simple and Universal Swarm Intelligence Engine

    MiroFish is a next-generation artificial intelligence prediction engine that leverages multi-agent technology and swarm-intelligence simulation to model, simulate, and forecast complex real-world scenarios. The system extracts “seed” information from sources such as breaking news, policy documents, and market signals to construct a high-fidelity digital parallel world populated by thousands of virtual agents with independent memory and behavior rules. Users can inject variables or conditions...
    Downloads: 1,464 This Week
    Last Update:
    See Project
  • 6
    LLMCompiler

    LLMCompiler

    An LLM Compiler for Parallel Function Calling

    LLMCompiler is an open-source framework designed to optimize how large language models orchestrate multiple external tool or function calls during complex reasoning tasks. Traditional LLM agent systems typically execute tool calls sequentially, which can create latency, higher costs, and reduced reliability when solving multi-step problems. LLMCompiler addresses this limitation by applying principles from classical compilers to analyze a task and construct an execution plan that allows multiple functions to run in parallel whenever possible. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Bend

    Bend

    A massively parallel, high-level programming language

    Bend is an interactive programming environment (REPL) built on top of the Kotlin language, designed to allow users to explore, experiment, and learn Kotlin in a live, feedback-driven manner. The tool lets you define variables, functions, or values at the prompt and iteratively refine them—immediately seeing output and types—while preserving state across commands. It emphasizes discoverability and experimentation: users can inspect functions, call them on sample inputs, and evolve logic...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    vLLM

    vLLM

    A high-throughput and memory-efficient inference and serving engine

    vLLM is a fast and easy-to-use library for LLM inference and serving. High-throughput serving with various decoding algorithms, including parallel sampling, beam search, and more.
    Downloads: 42 This Week
    Last Update:
    See Project
  • 9
    nndeploy

    nndeploy

    An Easy-to-Use and High-Performance AI Deployment Framework

    nndeploy is an open-source framework designed to simplify the deployment of artificial intelligence models across multiple hardware platforms and devices. The framework focuses on making it easier to transform trained AI models into production-ready applications that can run efficiently on desktops, mobile devices, servers, and edge computing hardware. Developers can use visual workflows to design and configure AI processing pipelines by connecting modular nodes that represent different...
    Downloads: 14 This Week
    Last Update:
    See Project
  • The AI coach for teams, built on validated assessments. Icon
    The AI coach for teams, built on validated assessments.

    Cloverleaf is an assessment-backed AI Coach that fully understands your people and the context of their workday.

    Give managers and teams proactive, contextual coaching to lead effectively, communicate clearly, and navigate real work situations as they happen.
    Learn More
  • 10
    Bazel

    Bazel

    Fast, scalable, multi-language and extensible build system

    Build and test software of any size, quickly and reliably. Bazel only rebuilds what is necessary. With advanced local and distributed caching, optimized dependency analysis and parallel execution, you get fast and incremental builds. Build and test Java, C++, Android, iOS, Go and a wide variety of other language platforms. Bazel runs on Windows, macOS, and Linux. Bazel helps you scale your organization, codebase and Continuous Integration system. It handles codebases of any size, in multiple repositories or a huge monorepo. ...
    Downloads: 76 This Week
    Last Update:
    See Project
  • 11

    Halide

    A language for fast, portable data-parallel computation

    Halide is a programming language for fast, portable data-parallel computation. It was designed to make writing high-performance image and array processing code much easier on modern machines. It works on all major operating systems and with several CPU architectures (X86, ARM, MIPS, Hexagon, PowerPC) and GPU Compute APIs (CUDA, OpenCL, OpenGL, among others). It isn't a standalone programming language however; rather it is embedded in C++ which means that you write C++ code, building an in-memory representation of a Halide pipeline using Halide's C++ API. ...
    Downloads: 11 This Week
    Last Update:
    See Project
  • 12
    ChainForge

    ChainForge

    An open-source visual programming environment

    ChainForge is an open-source visual programming environment designed to help developers systematically test, compare, and evaluate prompts and outputs across multiple large language models in a structured and scalable way. Instead of relying on isolated prompt experimentation, it introduces a dataflow-based interface that allows users to create complex prompt pipelines and evaluate them across different models, parameters, and datasets simultaneously. The platform enables rapid experimentation by generating permutations of prompts and inputs, making it possible to test hundreds of variations in parallel and analyze performance trends more effectively. ...
    Downloads: 7 This Week
    Last Update:
    See Project
  • 13
    Kimi K2.5

    Kimi K2.5

    Moonshot's most powerful AI model

    Kimi K2.5 is Moonshot AI’s open-source, native multimodal agentic model built through continual pretraining on approximately 15 trillion mixed vision and text tokens. Based on a 1T-parameter Mixture-of-Experts (MoE) architecture with 32B activated parameters, it integrates advanced language reasoning with strong visual understanding. K2.5 supports both “Thinking” and “Instant” modes, enabling either deep step-by-step reasoning or low-latency responses depending on the task. Designed for agentic workflows, it features an Agent Swarm mechanism that decomposes complex problems into coordinated sub-agents executing in parallel. ...
    Downloads: 39 This Week
    Last Update:
    See Project
  • 14
    Superset LLM

    Superset LLM

    Run an army of Claude Code, Codex, etc. on your machine

    Superset is a development environment and terminal-based platform designed to orchestrate multiple AI coding agents simultaneously within a single workspace. The tool enables developers to run many autonomous coding agents in parallel without the typical overhead of manually managing multiple terminals, repositories, or branches. Each agent task is isolated in its own Git worktree, ensuring that code changes from different agents do not interfere with each other while allowing developers to...
    Downloads: 19 This Week
    Last Update:
    See Project
  • 15
    Starlark in Go

    Starlark in Go

    Starlark in Go: the Starlark configuration language, implemented in Go

    starlark-go is the official Go implementation of the Starlark programming language, a Python-like, dynamically typed configuration and scripting language originally developed for the Bazel build system. Starlark combines Python’s expressiveness and readability with deterministic execution, making it ideal for defining build configurations, scripting automation, and embedding domain-specific logic in applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    CogVLM

    CogVLM

    A state-of-the-art open visual language model

    CogVLM is an open-source visual–language model suite—and its GUI-oriented sibling CogAgent—aimed at image understanding, grounding, and multi-turn dialogue, with optional agent actions on real UI screenshots. The flagship CogVLM-17B combines ~10B visual parameters with ~7B language parameters and supports 490×490 inputs; CogAgent-18B extends this to 1120×1120 and adds plan/next-action outputs plus grounded operation coordinates for GUI tasks.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Megatron

    Megatron

    Ongoing research training transformer models at scale

    Megatron is a large, powerful transformer developed by the Applied Deep Learning Research team at NVIDIA. This repository is for ongoing research on training large transformer language models at scale. We developed efficient, model-parallel (tensor, sequence, and pipeline), and multi-node pre-training of transformer based models such as GPT, BERT, and T5 using mixed precision. Megatron is also used in NeMo Megatron, a framework to help enterprises overcome the challenges of building and training sophisticated natural language processing models with billions and trillions of parameters. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Servo

    Servo

    Embed web technologies in applications

    Servo is an experimental, highly parallel, and embeddable browser rendering engine written in Rust. It leverages Rust’s memory-safety and concurrency strengths, supports modern GPU-powered rendering (WebGL/WebGPU), and serves as a research-forward alternative to traditional browser engines. Servo is a prototype web browser engine written in the Rust language. It is currently developed on 64-bit macOS, 64-bit Linux, 64-bit Windows, 64-bit OpenHarmony, and Android.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 19
    CubeCL

    CubeCL

    Multi-platform high-performance compute language extension for Rust

    CubeCL is a low-level compute language and compiler framework designed to simplify and optimize GPU programming for high-performance workloads, particularly in machine learning and numerical computing. It provides an abstraction layer that allows developers to write portable, hardware-efficient compute kernels without directly dealing with complex GPU APIs such as CUDA or OpenCL. CubeCL focuses on delivering predictable performance and composability by exposing explicit control over memory...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 20
    Numaflow

    Numaflow

    Kubernetes-native platform to run massively parallel data/streaming

    Numaflow is a Kubernetes-native tool for running massively parallel stream processing. A Numaflow Pipeline is implemented as a Kubernetes custom resource and consists of one or more source, data processing, and sink vertices. Numaflow installs in a few minutes and is easier and cheaper to use for simple data processing applications than a full-featured stream processing platform.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 21
    Bevy

    Bevy

    A refreshingly simple data-driven game engine built in Rust

    A refreshingly simple data-driven game engine built in Rust. All engine and game logic uses Bevy ECS, a custom Entity Component System. Massively Parallel and Cache-Friendly. The fastest ECS according to some benchmarks. Components are Rust structs, Systems are Rust functions. Queries, Global Resources, Local Resources, Change Detection, Lock-Free Parallel Scheduler. Bevy is still in the very early stages of development. APIs can and will change (now is the time to make suggestions!). ...
    Downloads: 14 This Week
    Last Update:
    See Project
  • 22
    wllama

    wllama

    WebAssembly binding for llama.cpp - Enabling on-browser LLM inference

    wllama is a WebAssembly-based library that enables large language model inference directly inside a web browser. Built as a binding for the llama.cpp inference engine, the project allows developers to run LLM models locally without requiring a server backend or dedicated GPU hardware. The library leverages WebAssembly SIMD capabilities to achieve efficient execution within modern browsers while maintaining compatibility across platforms. By running models locally on the user’s device, wllama...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 23
    ispc

    ispc

    Intel SPMD Program Compiler

    ispc is a compiler for a variant of the C programming language, with extensions for single programs, and multiple data programming. Under the SPMD model, the programmer writes a program that generally appears to be a regular serial program, though the execution model is actually that a number of program instances execute in parallel on the hardware. ispc compiles a C-based SPMD programming language to run on the SIMD units of CPUs and GPUs; it frequently provides a 3x or more speedup on architectures with 4-wide vector SSE units and 5x-6x on architectures with 8-wide AVX vector units, without any of the difficulty of writing intrinsics code. ...
    Downloads: 22 This Week
    Last Update:
    See Project
  • 24
    CGraph

    CGraph

    A general, three-party dependency-free, cross-platform

    CGraph is a high-performance, cross-platform Directed Acyclic Graph (DAG) framework implemented in pure C++ with no third-party dependencies, designed for building complex task pipelines and parallel execution workflows. It allows developers to model computational processes as graph structures, where nodes represent tasks and edges define dependencies, enabling efficient scheduling and execution. The framework includes a pipeline system that supports sequential and parallel execution,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Parallax

    Parallax

    Parallax is a distributed model serving framework

    Parallax is a decentralized inference framework designed to run large language models across distributed computing resources. Instead of relying on centralized GPU clusters in data centers, the system allows multiple heterogeneous machines to collaborate in serving AI inference workloads. Parallax divides model layers across different nodes and dynamically coordinates them to form a complete inference pipeline. A two-stage scheduling architecture determines how model layers are allocated to...
    Downloads: 5 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB