Open Source TypeScript Large Language Models (LLM)

TypeScript Large Language Models (LLM)

View 362 business solutions

Browse free open source TypeScript Large Language Models (LLM) and projects below. Use the toggles on the left to filter open source TypeScript Large Language Models (LLM) by OS, license, language, programming language, and project status.

  • SoftCo: Enterprise Invoice and P2P Automation Software Icon
    SoftCo: Enterprise Invoice and P2P Automation Software

    For companies that process over 20,000 invoices per year

    SoftCo Accounts Payable Automation processes all PO and non-PO supplier invoices electronically from capture and matching through to invoice approval and query management. SoftCoAP delivers unparalleled touchless automation by embedding AI across matching, coding, routing, and exception handling to minimize the number of supplier invoices requiring manual intervention. The result is 89% processing savings, supported by a context-aware AI Assistant that helps users understand exceptions, answer questions, and take the right action faster.
    Learn More
  • Data management solutions for confident marketing Icon
    Data management solutions for confident marketing

    For companies wanting a complete Data Management solution that is native to Salesforce

    Verify, deduplicate, manipulate, and assign records automatically to keep your CRM data accurate, complete, and ready for business.
    Learn More
  • 1
    Clippy

    Clippy

    Clippy, now with some AI

    Clippy is an open-source desktop assistant that allows users to run modern large language models locally while presenting them through a nostalgic interface inspired by Microsoft’s classic Clippy assistant from the 1990s. The project serves as both a playful homage to the early days of personal computing and a practical demonstration of local AI inference. Clippy integrates with the llama.cpp runtime to run models directly on a user’s computer without requiring cloud-based AI services. It supports models in the GGUF format, which allows it to run many publicly available open-source LLMs efficiently on consumer hardware. Users interact with the system through a simple animated assistant interface that can answer questions, generate text, and perform conversational tasks. The application includes one-click installation support for several popular models such as Meta’s Llama, Google’s Gemma, and other open models.
    Downloads: 44 This Week
    Last Update:
    See Project
  • 2
    OpenClaude

    OpenClaude

    Claude Code opened to any LLM

    OpenClaude is an open-source alternative or extension inspired by Claude-style agent systems, designed to provide similar capabilities in a customizable and self-hosted environment. The project focuses on enabling users to run their own AI agents with full control over data, workflows, and integrations, reducing reliance on proprietary platforms. It likely includes support for executing tasks, managing context, and interacting with external tools, allowing agents to perform real-world actions beyond simple text generation. The architecture emphasizes flexibility, enabling developers to adapt the system to different use cases and integrate it with various models or APIs. It may also include modular components for extending functionality, such as plugins or skills. The project reflects a broader trend toward open and decentralized AI systems that prioritize transparency and control.
    Downloads: 38 This Week
    Last Update:
    See Project
  • 3
    Eidos

    Eidos

    An extensible framework for Personal Data Management

    Eidos is an extensible personal data management platform designed to help users organize and interact with their information using a local-first architecture. The system transforms SQLite into a flexible personal database that can store structured and unstructured information such as notes, documents, datasets, and knowledge resources. Its interface is inspired by tools like Notion, allowing users to create documents, databases, and custom views to organize personal information. Unlike cloud-based knowledge tools, Eidos runs entirely on the user’s machine, ensuring privacy and high performance through local storage. The platform integrates large language models to enable AI-assisted features such as summarizing documents, translating content, and interacting with stored data conversationally. It also includes an extension system that allows developers to create custom tools, scripts, and workflows using programming languages such as TypeScript or Python.
    Downloads: 33 This Week
    Last Update:
    See Project
  • 4
    AI as Workspace

    AI as Workspace

    An elegant AI chat client. Full-featured, lightweight

    AI as Workspace, short for AI as Workspace, is an open-source AI client application that provides a unified interface for interacting with multiple large language models and AI tools within a single workspace environment. The platform is designed as a lightweight yet powerful desktop or web application that organizes AI interactions through structured workspaces. Instead of managing individual chat sessions separately, users can group conversations, artifacts, and tasks within customizable workspaces that support different projects or contexts. AIaW supports multiple AI providers and models through a flexible interface compatible with common API formats used by services such as OpenAI-style endpoints. The application also includes a plugin system that allows developers to extend the platform with additional capabilities such as automation tools, integrations, or custom AI utilities.
    Downloads: 25 This Week
    Last Update:
    See Project
  • Skillfully - The future of skills based hiring Icon
    Skillfully - The future of skills based hiring

    Realistic Workplace Simulations that Show Applicant Skills in Action

    Skillfully transforms hiring through AI-powered skill simulations that show you how candidates actually perform before you hire them. Our platform helps companies cut through AI-generated resumes and rehearsed interviews by validating real capabilities in action. Through dynamic job specific simulations and skill-based assessments, companies like Bloomberg and McKinsey have cut screening time by 50% while dramatically improving hire quality.
    Learn More
  • 5
    AingDesk

    AingDesk

    AI assistant that supports knowledge bases, model APIs

    AingDesk is an open-source desktop and server-based AI assistant platform designed to provide a user-friendly environment for interacting with language models and building AI-powered tools. The software enables users to run local AI models or connect to external model APIs through a unified interface. One of its primary goals is to simplify the process of building knowledge-based assistants by allowing users to create local knowledge bases that the AI can search and analyze. The system supports additional features such as web search, intelligent agent workflows, and multi-model conversations within a single session. AingDesk can be deployed locally on personal machines or installed as a server using containerized environments. Its design emphasizes accessibility, making it suitable for both beginners and experienced developers who want to experiment with AI tools.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 6
    Langflow

    Langflow

    Low-code app builder for RAG and multi-agent AI applications

    Langflow is a low-code app builder for RAG and multi-agent AI applications. It’s Python-based and agnostic to any model, API, or database.
    Downloads: 19 This Week
    Last Update:
    See Project
  • 7
    Superset LLM

    Superset LLM

    Run an army of Claude Code, Codex, etc. on your machine

    Superset is a development environment and terminal-based platform designed to orchestrate multiple AI coding agents simultaneously within a single workspace. The tool enables developers to run many autonomous coding agents in parallel without the typical overhead of manually managing multiple terminals, repositories, or branches. Each agent task is isolated in its own Git worktree, ensuring that code changes from different agents do not interfere with each other while allowing developers to track their progress independently. The platform includes built-in monitoring capabilities so users can observe the activity of each agent, receive notifications when tasks are completed, and quickly review changes produced by automated coding workflows. Superset also integrates tools for reviewing code differences, editing generated outputs, and managing the development environment directly from the interface.
    Downloads: 18 This Week
    Last Update:
    See Project
  • 8
    FreedomGPT

    FreedomGPT

    React and Electron-based app that executes the FreedomGPT LLM locally

    FreedomGPT is a locally executed large language model (LLM) application built using React and Electron, allowing users to interact with AI models privately on their Mac or Windows devices. The app enables offline operation, ensuring privacy and security while providing a chat-based interface for seamless communication with the AI. It supports integration with models like Liberty Edge and offers an open-source solution for those seeking more control over their AI interactions. The app's setup is simple, and it includes clear installation guides for both macOS and Windows platforms, as well as detailed instructions for building necessary libraries like llama.cpp.
    Downloads: 17 This Week
    Last Update:
    See Project
  • 9
    MCP Router

    MCP Router

    A Unified MCP Server Management App (MCP Manager)

    MCP Router is an open-source management platform designed to simplify the deployment and coordination of Model Context Protocol (MCP) servers used by AI agents. MCP is an emerging standard that allows language models and AI assistants to connect to external tools, data sources, and services through a structured interface. The MCP Router project acts as a centralized manager that helps developers run, configure, and coordinate multiple MCP servers within a single environment. This enables AI applications to access multiple tools and knowledge sources through a unified interface rather than connecting to each service individually. The project provides infrastructure for routing requests between clients and MCP servers, enabling scalable multi-tool agent systems. Developers building AI agents can use the platform to manage tool endpoints, control service availability, and simplify agent integration workflows.
    Downloads: 17 This Week
    Last Update:
    See Project
  • Field Sales+ for MS Dynamics 365 and Salesforce Icon
    Field Sales+ for MS Dynamics 365 and Salesforce

    Maximize your sales performance on the go.

    Bring Dynamics 365 and Salesforce wherever you go with Resco’s solution. With powerful offline features and reliable data syncing, your team can access CRM data on mobile devices anytime, anywhere. This saves time, cuts errors, and speeds up customer visits.
    Learn More
  • 10
    node-llama-cpp

    node-llama-cpp

    Run AI models locally on your machine with node.js bindings for llama

    node-llama-cpp is a JavaScript and Node.js binding that allows developers to run large language models locally using the high-performance inference engine provided by llama.cpp. The library enables applications built with Node.js to interact directly with local LLM models without requiring a remote API or external service. By using native bindings and optimized model execution, the framework allows developers to integrate advanced language model capabilities into desktop applications, server software, and command-line tools. The system automatically detects the available hardware on a machine and selects the most appropriate compute backend, including CPU or GPU acceleration. Developers can use the library to perform tasks such as text generation, conversational chat, embedding generation, and structured output generation. Because it runs models locally, the platform is particularly useful for privacy-sensitive environments or offline AI deployments.
    Downloads: 17 This Week
    Last Update:
    See Project
  • 11
    Read Frog

    Read Frog

    Open Source Immersive Translate

    Read Frog is an open-source browser extension designed to transform everyday web reading into an immersive language learning experience powered by artificial intelligence. The tool integrates translation, contextual explanations, and content analysis directly into the browsing workflow so users can learn languages naturally while reading authentic online content. Instead of forcing learners to switch between translation tools and the original text, the extension displays translations alongside the source language, making comprehension immediate and continuous. The system automatically extracts the main content of an article using intelligent parsing techniques, allowing users to focus on the most relevant text without distractions. AI models are used to generate summaries, introductions, and explanations for words, phrases, and sentences based on the learner’s language level, making the experience personalized and adaptive.
    Downloads: 16 This Week
    Last Update:
    See Project
  • 12
    Deta Surf

    Deta Surf

    Personal AI Notebooks. Organize files & webpages and generate notes

    Surf is an open-source AI-driven development tool designed to simplify the process of building and experimenting with artificial intelligence applications. The platform provides a streamlined development environment where developers can test models, run experiments, and deploy small AI services with minimal infrastructure overhead. It focuses on simplicity and speed, allowing developers to prototype ideas quickly without managing complex cloud configurations. Surf integrates modern AI workflows such as prompt-based applications, lightweight APIs, and automated deployment pipelines. The platform is particularly useful for developers who want to experiment with AI models locally while maintaining the option to deploy them in production environments later. Its architecture is designed to minimize setup complexity while still supporting scalable application structures.
    Downloads: 14 This Week
    Last Update:
    See Project
  • 13
    Dify

    Dify

    One API for plugins and datasets, one interface for prompt engineering

    Dify is an easy-to-use LLMOps platform designed to empower more people to create sustainable, AI-native applications. With visual orchestration for various application types, Dify offers out-of-the-box, ready-to-use applications that can also serve as Backend-as-a-Service APIs. Unify your development process with one API for plugins and datasets integration, and streamline your operations using a single interface for prompt engineering, visual analytics, and continuous improvement. Out-of-the-box web sites supporting form mode and chat conversation mode A single API encompassing plugin capabilities, context enhancement, and more, saving you backend coding effort Visual data analysis, log review, and annotation for applications
    Downloads: 14 This Week
    Last Update:
    See Project
  • 14
    Flowise

    Flowise

    Drag & drop UI to build your customized LLM flow

    Open source UI visual tool to build your customized LLM flow using LangchainJS, written in Node Typescript/Javascript. Conversational agent for a chat model which utilizes chat-specific prompts and buffer memory. Open source is the core of Flowise, and it will always be free for commercial and personal usage. Flowise support different environment variables to configure your instance. You can specify the following variables in the .env file inside the packages/server folder.
    Downloads: 14 This Week
    Last Update:
    See Project
  • 15
    Hollama

    Hollama

    A minimal LLM chat app that runs entirely in your browser

    Hollama is a lightweight open-source chat application designed to run entirely within the browser while interacting with large language model servers. The project provides a minimal but powerful user interface for communicating with local or remote LLMs, including servers powered by Ollama or OpenAI-compatible APIs. Because the application runs as a static web interface, it does not require complex backend infrastructure and can be easily deployed or self-hosted. Hollama supports both text-based and multimodal interactions, allowing users to work with models that process images as well as text. The interface includes features for editing prompts, retrying responses, copying generated code snippets, and storing conversation history locally within the browser. Mathematical expressions can be rendered using KaTeX, and Markdown formatting allows code blocks and structured outputs to appear clearly within conversations.
    Downloads: 13 This Week
    Last Update:
    See Project
  • 16
    Pluely

    Pluely

    The Open Source Alternative to Cluely

    Pluely is an open-source AI automation framework designed to simplify the development and deployment of AI-driven workflows across applications and services. The system focuses on orchestrating tasks performed by large language models and other AI components, allowing developers to define structured workflows where models interact with tools, APIs, and external systems. By providing a modular architecture for building AI pipelines, the platform enables developers to connect multiple processing steps such as data retrieval, prompt execution, analysis, and response generation. The project emphasizes flexibility, allowing developers to extend the platform with custom integrations and automation logic. This makes the framework suitable for building intelligent assistants, automated business workflows, and data-processing pipelines that rely on generative AI capabilities.
    Downloads: 12 This Week
    Last Update:
    See Project
  • 17
    llama.vscode

    llama.vscode

    VS Code extension for LLM-assisted code/text completion

    llama.vscode is a Visual Studio Code extension that provides AI-assisted coding features powered primarily by locally running language models. The extension is designed to be lightweight and efficient, enabling developers to use AI tools even on consumer-grade hardware. It integrates with the llama.cpp runtime to run language models locally, eliminating the need to rely entirely on external APIs or cloud providers. The extension supports common AI development features such as code completion, conversational chat assistance, and AI-assisted code editing directly within the IDE. Developers can select and manage models through a configuration interface that automatically downloads and runs the required models locally. The extension also supports agent-style coding workflows, where AI tools can perform more complex tasks such as analyzing project context or editing multiple files.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 18
    opcode

    opcode

    A powerful GUI app and Toolkit for Claude Code

    opcode is an open source desktop application and toolkit designed to enhance the developer experience when working with Claude Code by providing a graphical interface and advanced workflow management tools. The project acts as a command center for AI-assisted programming, bridging the gap between command-line workflows and modern visual development environments. Built using the Tauri framework, Opcode enables developers to manage multiple Claude sessions, create custom agents, and track usage in a centralized interface. The platform is intended to make AI-assisted coding more intuitive by providing visual tools for monitoring agent activity, organizing projects, and reviewing development timelines. It includes features that help developers coordinate tasks between agents and human collaborators while maintaining transparency over the actions performed by AI systems.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 19
    NativeMind Extension

    NativeMind Extension

    Your fully private, open-source, on-device AI assistant

    NativeMindExtension is an open-source browser extension that provides a private, on-device AI assistant designed to run without cloud dependencies. The project is built around a privacy-first model in which conversations, document analysis, translations, and writing assistance stay on the user’s device rather than being sent to external servers. It integrates with local model back ends such as Ollama and also supports WebLLM for quick in-browser trials, giving users a choice between stronger local setups and lighter no-install demonstrations. The extension is aimed at everyday browser workflows, offering features like multi-tab context awareness, webpage summarization, document understanding, contextual toolbars, and AI-assisted rewriting directly inside the browsing experience. Because it runs locally after setup, it is also positioned as an always-available assistant that avoids API quotas, network latency, and service outages common in cloud-based AI tools.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 20
    TaxHacker

    TaxHacker

    Self-hosted AI accounting app. LLM analyzer for receipts

    TaxHacker is an open-source, self-hosted accounting application that uses artificial intelligence to automate financial record management for freelancers, independent developers, and small businesses. The system is designed to simplify bookkeeping by automatically processing financial documents such as receipts, invoices, and transaction records. It integrates large language models to analyze these documents, extract relevant financial information, and categorize expenses or income based on configurable rules. Users can deploy the application on their own infrastructure, ensuring that financial data remains private and under their control rather than being processed by external services. The software provides tools for tracking income streams, monitoring expenses, and organizing financial records in a structured format. Because the system supports customizable prompts and categories, users can adapt the AI analysis to match their accounting workflows or tax requirements.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 21
    Lunary

    Lunary

    The production toolkit for LLMs. Observability, prompt management

    Lunary helps developers of LLM Chatbots develop and improve them.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 22
    PasteGuard

    PasteGuard

    Masks sensitive data and secrets before they reach AI

    PasteGuard is an open-source privacy proxy that protects sensitive information like personal data and API secrets by detecting and masking them before they reach large language model APIs such as OpenAI or Anthropic Claude. It sits between an application and the LLM provider, automatically replacing names, emails, tokens, and other personally identifiable information (PII) with placeholders so that external services never see raw sensitive values, and then optionally unmasking them in the returned output. PasteGuard supports two primary modes: mask mode, which anonymizes data and still uses external APIs; and route mode, which forwards sensitive requests to a local LLM inference engine while sending the rest to the cloud. It can be self-hosted via Docker, works with a wide range of SDKs and tools, and includes a browser extension for automatic protection in everyday AI chats.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 23
    Pezzo

    Pezzo

    Open-source, developer-first LLMOps platform

    Pezzo enables you to build, test, monitor and instantly ship AI all in one platform, while constantly optimizing for cost and performance. Packed with powerful features to streamline your workflow, so you can focus on what matters. Pezzo is a fully cloud-native and open-source LLMOps platform. Seamlessly observe and monitor your AI operations, troubleshoot issues, save up to 90% on costs and latency, collaborate and manage your prompts in one place, and instantly deliver AI changes.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 24
    oh my PI

    oh my PI

    AI Coding agent for the terminal

    Oh-My-Pi is an open-source AI agent toolkit focused on creating intelligent coding assistants that operate directly from the terminal environment. The project provides a command-line coding agent capable of analyzing repositories, generating commits, editing code, and interacting with development tools through an integrated tool system. Instead of functioning as a simple prompt-based assistant, the system includes an agent architecture that can inspect Git repositories, analyze changes, and perform development actions with fine-grained control. The platform also supports tool-based workflows where the agent can run shell commands, read files, modify code, and stage changes during development tasks. It includes infrastructure for integrating different AI providers and models through a unified API layer, allowing developers to switch between models while keeping the same agent interface.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 25
    promptfoo

    promptfoo

    Evaluate and compare LLM outputs, catch regressions, improve prompts

    Ensure high-quality LLM outputs with automatic evals. Use a representative sample of user inputs to reduce subjectivity when tuning prompts. Use built-in metrics, LLM-graded evals, or define your own custom metrics. Compare prompts and model outputs side-by-side, or integrate the library into your existing test/CI workflow. Use OpenAI, Anthropic, and open-source models like Llama and Vicuna, or integrate custom API providers for any LLM API.
    Downloads: 7 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • 2
  • 3
  • 4
  • Next
MongoDB Logo MongoDB