Open Source Unix Shell Artificial Intelligence Software - Page 2

Unix Shell Artificial Intelligence Software

View 13574 business solutions

Browse free open source Unix Shell Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Unix Shell Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • SoftCo: Enterprise Invoice and P2P Automation Software Icon
    SoftCo: Enterprise Invoice and P2P Automation Software

    For companies that process over 20,000 invoices per year

    SoftCo Accounts Payable Automation processes all PO and non-PO supplier invoices electronically from capture and matching through to invoice approval and query management. SoftCoAP delivers unparalleled touchless automation by embedding AI across matching, coding, routing, and exception handling to minimize the number of supplier invoices requiring manual intervention. The result is 89% processing savings, supported by a context-aware AI Assistant that helps users understand exceptions, answer questions, and take the right action faster.
    Learn More
  • The Most Powerful Software Platform for EHSQ and ESG Management Icon
    The Most Powerful Software Platform for EHSQ and ESG Management

    Addresses the needs of small businesses and large global organizations with thousands of users in multiple locations.

    Choose from a complete set of software solutions across EHSQ that address all aspects of top performing Environmental, Health and Safety, and Quality management programs.
    Learn More
  • 1
    Xianyu Intelligent Monitor Bot

    Xianyu Intelligent Monitor Bot

    AI tool for real-time monitoring and analysis of Goofish listings

    ai-goofish-monitor is an open source automation tool designed to monitor listings on the Goofish second-hand marketplace and analyze them using artificial intelligence. It combines browser automation with AI-based analysis to automatically search, collect, and evaluate newly posted items that match a user’s purchase criteria. It uses Playwright to simulate real user interactions with the marketplace, allowing the system to retrieve product data and track updates in near real time. ai-goofish-monitor can run multiple monitoring tasks simultaneously, each configured with specific keywords, price ranges, and filtering conditions. A built-in web management interface allows users to create tasks, review results, and manage monitoring rules without relying solely on command line tools. AI models analyze product descriptions, images, and seller information to determine whether a listing meets defined requirements and should be recommended to the user.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 2
    YuE

    YuE

    Open source AI model for generating full songs from lyrics prompts

    YuE is an open source project that provides a foundation model designed for full-song music generation using artificial intelligence. It focuses on transforming text inputs such as lyrics and genre prompts into complete musical compositions that include both vocal and instrumental tracks. Unlike many shorter audio generators, the model is capable of producing songs that last several minutes while maintaining coherent musical structure and alignment with the provided lyrics. YuE introduces a family of models built on large language model architectures that process music generation as a sequence prediction task. YuE also incorporates techniques such as track-decoupled prediction and progressive conditioning to help manage complex audio signals and maintain consistency throughout long compositions. It includes inference scripts, prompt examples, evaluation tools, and training components that enable researchers and developers to experiment with AI-based music.
    Downloads: 8 This Week
    Last Update:
    See Project
  • 3
    Cactus

    Cactus

    Low-latency AI inference engine optimized for mobile devices

    Cactus is a low-latency, energy-efficient AI inference framework designed specifically for mobile devices and wearables, enabling advanced machine learning capabilities directly on-device. It provides a full-stack architecture composed of an inference engine, a computation graph system, and highly optimized hardware kernels tailored for ARM-based processors. Cactus emphasizes efficient memory usage through techniques such as zero-copy computation graphs and quantized model formats, allowing large models to run within the constraints of mobile hardware. It supports a wide range of AI tasks including text generation, speech-to-text, vision processing, and retrieval-augmented workflows through a unified API interface. A notable feature of Cactus is its hybrid execution model, which can dynamically route tasks between on-device processing and cloud services when additional compute is required.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 4
    ChatLab

    ChatLab

    Local-first AI chat analysis tool for insights from conversation data

    ChatLab is an open source desktop application designed to help users analyze and better understand their personal chat histories through structured data exploration and AI-assisted insights. It enables users to import chat exports from multiple messaging platforms and transform them into a unified data model for consistent analysis. By combining a flexible SQL engine with AI agents, the tool allows users to query, summarize, and explore conversation patterns in a more interactive and intelligent way. ChatLab emphasizes a local-first approach, meaning all chat data is processed and stored on the user’s device rather than being uploaded to external servers. It supports large-scale datasets through streaming parsing and multi-worker processing, allowing it to handle millions of messages efficiently. ChatLab also includes visualization features that present trends, activity patterns, and interaction metrics in a clear and accessible format.
    Downloads: 7 This Week
    Last Update:
    See Project
  • MicroStation by Bentley Systems is the trusted computer-aided design (CAD) software built specifically for infrastructure design. Icon
    MicroStation by Bentley Systems is the trusted computer-aided design (CAD) software built specifically for infrastructure design.

    Microstation enables architects, engineers, and designers to create precise 2D and 3D drawings that bring complex projects to life.

    MicroStation is the only computer-aided design software for infrastructure design, helping architects and engineers like you bring their vision to life, present their designs to their clients, and deliver their projects to the community.
    Learn More
  • 5
    StableSwarmUI

    StableSwarmUI

    Multi-user UI for managing and running Stable Diffusion workflows tool

    StableSwarmUI is a web-based interface designed to manage and coordinate Stable Diffusion image generation workflows in a multi-user environment. It focuses on enabling multiple users to interact with shared resources, making it suitable for collaborative or server-based deployments. It provides a centralized system where users can submit, monitor, and manage generation tasks through a browser interface. It abstracts much of the complexity involved in running diffusion models by offering a structured environment for handling prompts, outputs, and processing queues. StableSwarmUI is built to work alongside backend systems that execute the actual image generation, allowing separation between user interaction and compute workloads. It also emphasizes scalability, making it useful for setups where multiple jobs need to be processed efficiently. Overall, it serves as a coordination layer for Stable Diffusion usage rather than a standalone model implementation.
    Downloads: 7 This Week
    Last Update:
    See Project
  • 6
    Inbox Zero

    Inbox Zero

    AI assistant that automates email tasks to help achieve inbox zero

    Inbox Zero is an open source AI-powered email assistant designed to help users manage and process their inbox more efficiently. It aims to reduce the time spent handling email by automatically organizing, prioritizing, and responding to messages using customizable automation rules and artificial intelligence. Users can define prompts or rule-based actions that guide how the assistant processes incoming messages, enabling automated workflows for sorting, replying, or handling routine communication. Inbox Zero is structured as a modern web application built with a monorepo architecture that contains multiple applications and shared packages, allowing modular development and easier maintenance. It integrates with email services and can automate actions such as scheduling tasks, generating replies, and managing follow-ups. Inbox Zero is designed to allow users to retain precise control over automation rules while still benefiting from AI-driven suggestions and analysis.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 7
    Text Embeddings Inference

    Text Embeddings Inference

    High-performance inference server for text embeddings models API layer

    Text Embeddings Inference is a high-performance server designed to serve text embedding models efficiently in production environments. It focuses on delivering fast and scalable embedding generation by leveraging optimized inference techniques and modern hardware acceleration. It is built to support transformer-based embedding models, making it suitable for tasks such as semantic search, clustering, and retrieval-augmented systems. It provides an API interface that allows developers to integrate embedding capabilities into applications without managing model internals directly. Text Embeddings Inference is optimized for throughput and low latency, enabling it to handle large volumes of requests reliably. It also emphasizes ease of deployment, often using containerization and configurable runtime options to adapt to different infrastructure setups.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 8
    WeKnora

    WeKnora

    LLM framework for document understanding and semantic retrieval

    WeKnora is an open source framework developed for deep document understanding and semantic information retrieval using large language models. It focuses on analyzing complex and heterogeneous documents by combining multiple processing stages such as multimodal document parsing, vector indexing, and intelligent retrieval. It follows the Retrieval-Augmented Generation (RAG) paradigm, where relevant document segments are retrieved and used by language models to generate accurate, context-aware responses. This approach enables the system to provide more reliable answers by grounding model reasoning in the content of uploaded documents. WeKnora is designed with a modular architecture that separates components for document processing, search strategies, and model inference, allowing developers to customize or extend different parts of the pipeline. It supports knowledge base management and conversational question answering built on top of structured and unstructured documents.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 9
    CodeGeeX2

    CodeGeeX2

    CodeGeeX2: A More Powerful Multilingual Code Generation Model

    CodeGeeX2 is the second-generation multilingual code generation model from ZhipuAI, built upon the ChatGLM2-6B architecture and trained on 600B code tokens. Compared to the first generation, it delivers a significant boost in programming ability across multiple languages, outperforming even larger models like StarCoder-15B in some benchmarks despite having only 6B parameters. The model excels at code generation, translation, summarization, debugging, and comment generation, and it supports over 100 programming languages. With improved inference efficiency, quantization options, and multi-query/flash attention, CodeGeeX2 achieves faster generation speeds and lightweight deployment, requiring as little as 6GB GPU memory at INT4 precision. Its backend powers the CodeGeeX IDE plugins for VS Code, JetBrains, and other editors, offering developers interactive AI assistance with features like infilling and cross-file completion.
    Downloads: 5 This Week
    Last Update:
    See Project
  • Next-Gen Encryption for Post-Quantum Security | CLEAR by Quantum Knight Icon
    Next-Gen Encryption for Post-Quantum Security | CLEAR by Quantum Knight

    Lock Down Any Resource, Anywhere, Anytime

    CLEAR by Quantum Knight is a FIPS-140-3 validated encryption SDK engineered for enterprises requiring top-tier security. Offering robust post-quantum cryptography, CLEAR secures files, streaming media, databases, and networks with ease across over 30 modern platforms. Its compact design, smaller than a single smartphone image, ensures maximum efficiency and low energy consumption.
    Learn More
  • 10
    Kalavai

    Kalavai

    Turn everyday devices into your own AI cluster

    Kalavai is a self-hosted platform that turns everyday devices into your very own AI cluster. Do you have an old desktop or a gaming laptop gathering dust? Aggregate resources from multiple machines and say goodbye to CUDA out-of-memory errors. Deploy your favorite open-source LLM, fine-tune it with your own data, or simply run your distributed work, zero-DevOps. Simple. Private. Yours.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 11
    LitServe

    LitServe

    Minimal Python framework for scalable AI inference servers fast

    LitServe is a minimal Python framework designed for building custom AI inference servers with full control over how models are executed and served. It allows developers to define their own inference logic, making it suitable for complex systems such as multi-model pipelines, agents, and retrieval-augmented generation workflows. Unlike traditional serving tools that enforce rigid abstractions, LitServe focuses on flexibility by letting users control request handling, batching strategies, and output processing directly in Python. LitServe is built on top of FastAPI and extends it with AI-specific optimizations such as efficient multi-worker execution, which can significantly improve throughput. It includes built-in capabilities for batching, streaming responses, and automatic scaling across CPUs and GPUs, enabling high-performance deployments.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 12
    MCP Toolbox for Databases

    MCP Toolbox for Databases

    Open source MCP server that exposes database tools for AI agents

    GenAI Toolbox, also known as MCP Toolbox for Databases, is an open source server designed to simplify how generative AI applications interact with databases. It provides a central service that exposes database operations as reusable tools that can be consumed by AI agents and developer workflows. It handles common infrastructure concerns such as authentication, connection pooling, and performance optimization so developers do not have to implement them individually in each application. By defining tools and data sources through configuration files, developers can standardize how AI systems access and operate on database resources. GenAI Toolbox is designed to integrate with agent frameworks and development environments so that AI assistants can execute database-related tasks with proper context and security. It also supports observability through built-in metrics and tracing capabilities, allowing developers to monitor how tools are used and debug interactions.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 13
    Memobase

    Memobase

    Fast backend for long-term AI user memory via structured profiles

    Memobase is an open source backend system that enables long-term user memory functionality for AI applications by capturing and structuring information about users across interactions. Its design centers on creating user profiles and recording event timelines, allowing AI systems to remember, understand, and evolve in their behaviour toward individual users over time. Instead of relying purely on traditional embedding-based retrieval or RAG systems, Memobase uses profile and timeline structures to deliver memory that reflects user context efficiently and meaningfully. The system focuses on three principal performance metrics: high search performance, reduced large language model (LLM) costs through batch processing techniques, and low latency with minimal SQL operations. Memobase supports integration with existing LLM workflows via APIs and SDKs (including Python, Node, and Go), making it easy to adopt within diverse application stacks.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 14
    RuoYi AI

    RuoYi AI

    Enterprise AI platform for building, deploying, and managing apps

    RuoYi AI is a full-stack enterprise-oriented AI development platform designed to help developers rapidly build, deploy, and manage intelligent applications using modern large language models and AI ecosystems. It provides a unified framework for integrating multiple AI models from different providers, allowing teams to switch or combine models through a consistent interface without vendor lock-in. RuoYi AI includes built-in support for retrieval-augmented generation, enabling organizations to create secure, private knowledge bases with high-accuracy search and reasoning capabilities. It also offers visual workflow orchestration tools that allow users to design complex AI pipelines, automate tasks, and coordinate multi-agent systems for advanced decision-making scenarios. In addition to backend capabilities, RuoYi AI includes frontend components and administrative dashboards built with modern web technologies, making it a complete end-to-end solution.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 15
    ShaHaN SSH Panel

    ShaHaN SSH Panel

    SSH User Management With Add/Delete Users

    SSH user management with add/delete users, online users, and limit users.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 16
    TaskingAI

    TaskingAI

    Open platform for building, deploying, and managing LLM agents

    TaskingAI is an open source platform designed to simplify the development and deployment of applications powered by large language models. It follows a Backend as a Service approach, allowing developers to separate AI logic from frontend product development while maintaining a structured and scalable workflow. TaskingAI integrates hundreds of language models from multiple providers into a unified system, enabling developers to switch models or combine capabilities without major reconfiguration. It includes a modular architecture that supports components such as assistants, tools, retrieval systems, and conversation management, all accessible through a consistent interface. TaskingAI also provides a built-in user interface for managing projects, testing workflows, and configuring AI agents without needing to rely entirely on code. It supports advanced techniques like retrieval-augmented generation and plugin-based extensions, allowing developers to enhance agent capabilities.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 17
    fairseq2

    fairseq2

    FAIR Sequence Modeling Toolkit 2

    fairseq2 is a modern, modular sequence modeling framework developed by Meta AI Research as a complete redesign of the original fairseq library. Built from the ground up for scalability, composability, and research flexibility, fairseq2 supports a broad range of language, speech, and multimodal content generation tasks, including instruction fine-tuning, reinforcement learning from human feedback (RLHF), and large-scale multilingual modeling. Unlike the original fairseq—which evolved into a large, monolithic codebase—fairseq2 introduces a clean, plugin-oriented architecture designed for long-term maintainability and rapid experimentation. It supports multi-GPU and multi-node distributed training using DDP, FSDP, and tensor parallelism, capable of scaling up to 70B+ parameter models. The framework integrates seamlessly with PyTorch 2.x features such as torch.compile, Fully Sharded Data Parallel (FSDP), and modern configuration management.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 18
    ChatGLM3

    ChatGLM3

    ChatGLM3 series: Open Bilingual Chat LLMs | Open Source Bilingual Chat

    ChatGLM3 is ZhipuAI & Tsinghua KEG’s third-gen conversational model suite centered on the 6B-parameter ChatGLM3-6B. It keeps the series’ smooth dialog and low deployment cost while adding native tool use (function calling), a built-in code interpreter, and agent-style workflows. The family includes base and long-context variants (8K/32K/128K). The repo ships Python APIs, CLI and web demos (Gradio/Streamlit), an OpenAI-format API server, and a compact fine-tuning kit. Quantization (4/8-bit), CPU/MPS support, and accelerator backends (TensorRT-LLM, OpenVINO, chatglm.cpp) enable lightweight local or edge deployment.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 19
    Claude-Flow

    Claude-Flow

    The leading agent orchestration platform for Claude

    Claude-Flow v2 Alpha is an advanced AI orchestration and automation framework designed for enterprise-grade, large-scale AI-driven development. It enables developers to coordinate multiple specialized AI agents in real time through a hive-mind architecture, combining swarm intelligence, neural reasoning, and a powerful set of 87 Modular Control Protocol (MCP) tools. The platform supports both quick swarm tasks and persistent multi-agent sessions known as hives, facilitating distributed AI collaboration with persistent contextual memory. At its core, Claude-Flow integrates Dynamic Agent Architecture (DAA) for self-organizing agent management, neural pattern recognition accelerated by WebAssembly SIMD, and a SQLite-based memory system for context retention and knowledge persistence across tasks. It automates development workflows via pre- and post-operation hooks, providing seamless coordination, code formatting, validation, and performance optimization.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 20
    Generative AI for Beginners (Version 3)

    Generative AI for Beginners (Version 3)

    21 Lessons, Get Started Building with Generative AI

    Generative AI for Beginners is a 21-lesson course by Microsoft Cloud Advocates that teaches the fundamentals of building generative AI applications in a practical, project-oriented way. Lessons are split into “Learn” modules for core concepts and “Build” modules with hands-on code in Python and TypeScript, so you can jump in at any point that matches your goals. The course covers everything from model selection, prompt engineering, and chat/text/image app patterns to secure development practices and UX for AI. It also walks through modern application techniques such as function calling, RAG with vector databases, working with open source models, agents, fine-tuning, and using SLMs. Each lesson includes a short video, a written guide, runnable samples for Azure OpenAI, the GitHub Marketplace Model Catalog, and the OpenAI API, plus a “Keep Learning” section for deeper study.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 21
    Kaldi

    Kaldi

    kaldi-asr/kaldi is the official location of the Kaldi project

    Kaldi is an open source toolkit for speech recognition research. It provides a powerful framework for building state-of-the-art automatic speech recognition (ASR) systems, with support for deep neural networks, Gaussian mixture models, hidden Markov models, and other advanced techniques. The toolkit is widely used in both academia and industry due to its flexibility, extensibility, and strong community support. Kaldi is designed for researchers who need a highly customizable environment to experiment with new algorithms, as well as for practitioners who want robust, production-ready ASR pipelines. It includes extensive tools for data preparation, feature extraction, acoustic and language modeling, decoding, and evaluation. With its modular design, Kaldi allows users to adapt the system to a wide range of languages and domains. As one of the most influential projects in speech recognition, it has become a foundation for much of the modern work in ASR.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 22
    MobileLLM

    MobileLLM

    MobileLLM Optimizing Sub-billion Parameter Language Models

    MobileLLM is a lightweight large language model (LLM) framework developed by Facebook Research, optimized for on-device deployment where computational and memory efficiency are critical. Introduced in the ICML 2024 paper “MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases”, it focuses on delivering strong reasoning and generalization capabilities in models under one billion parameters. The framework integrates several architectural innovations—SwiGLU activation, deep and thin network design, embedding sharing, and grouped-query attention (GQA)—to achieve a superior trade-off between model size, inference speed, and accuracy. MobileLLM demonstrates remarkable performance, with the 125M and 350M variants outperforming previous state-of-the-art models of the same scale by up to 4.3% on zero-shot commonsense reasoning tasks.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 23
    NOFX

    NOFX

    Open source AI trading OS for autonomous multi-model trading systems

    NOFX is an open source AI-powered trading operating system designed to automate financial trading workflows using autonomous AI agents. It acts as an infrastructure layer that transforms market data into AI-driven trade decisions and execution. Instead of requiring users to manually configure machine learning models, data sources, and API integrations, the system allows AI components to perceive market conditions, select models, and perform trading actions automatically. It supports running multiple AI models simultaneously and allows them to compete or collaborate when making trading decisions. NOFX integrates trading infrastructure such as exchange connectivity, strategy management, and performance monitoring into a single environment. It also includes components for strategy development, backtesting, and real-time monitoring so traders and researchers can evaluate algorithmic trading approaches.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 24
    Open Semantic Search

    Open Semantic Search

    Open source semantic search and text analytics for large document sets

    Open Semantic Search is an open source research and analytics platform designed for searching, analyzing, and exploring large collections of documents using semantic search technologies. It provides an integrated search server combined with a document processing pipeline that supports crawling, text extraction, and automated analysis of content from many different sources. Open Semantic Search includes an ETL framework that can ingest documents, process them through analysis steps, and enrich the data with extracted information such as named entities and metadata. It also supports optical character recognition to extract text from images and scanned documents, including images embedded inside PDF files. It integrates text mining and analytics capabilities that allow users to examine relationships, topics, and structured data within document collections.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 25
    Preswald

    Preswald

    Python tool for browser-based interactive data apps in one file

    Preswald is an open source Python-based framework and static-site generator designed for building interactive data applications that run entirely in the browser. It packages application logic, data processing, and user interface components into a single self-contained output, enabling easy sharing and deployment without requiring local dependencies. Preswald leverages a WebAssembly runtime along with technologies like Pyodide and DuckDB to execute Python code directly in the browser environment. This approach allows developers to create dashboards, reports, notebooks, and data tools that are portable, fast, and capable of running offline. Preswald emphasizes a code-first workflow where users define applications entirely in Python while using built-in UI components such as tables, charts, and forms. It also includes a reactive execution model that only recomputes necessary parts of the app, improving performance and responsiveness.
    Downloads: 4 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB