Open Source Unix Shell Artificial Intelligence Software - Page 3

Unix Shell Artificial Intelligence Software

View 13574 business solutions

Browse free open source Unix Shell Artificial Intelligence Software and projects below. Use the toggles on the left to filter open source Unix Shell Artificial Intelligence Software by OS, license, language, programming language, and project status.

  • Data management solutions for confident marketing Icon
    Data management solutions for confident marketing

    For companies wanting a complete Data Management solution that is native to Salesforce

    Verify, deduplicate, manipulate, and assign records automatically to keep your CRM data accurate, complete, and ready for business.
    Learn More
  • The Most Powerful Software Platform for EHSQ and ESG Management Icon
    The Most Powerful Software Platform for EHSQ and ESG Management

    Addresses the needs of small businesses and large global organizations with thousands of users in multiple locations.

    Choose from a complete set of software solutions across EHSQ that address all aspects of top performing Environmental, Health and Safety, and Quality management programs.
    Learn More
  • 1
    Rhasspy

    Rhasspy

    Offline private voice assistant for many human languages

    Rhasspy (ˈɹæspi) is an open-source, fully offline set of voice assistant services for many human languages that works well with Hermes protocol-compatible services (Snips.AI), Home Assistant and Hass.io, Node-RED, Jeedom, OpenHAB. Rhasspy will produce JSON events that can trigger action in home automation software, such as a Node-RED flow. Rhasspy comes with a snazzy web interface that lets you configure, program, and test your voice assistant remotely from your web browser. All of the web UI's functionality is exposed in a comprehensive HTTP API. You can easily extend or replace functionality in Rhasspy by using the appropriate messages. Many of these messages can be also sent and received over the HTTP API and the WebSocket API. Rhasspy is intended for savvy amateurs or advanced users that want to have a private voice interface to their chosen home automation software.
    Downloads: 4 This Week
    Last Update:
    See Project
  • 2
    Desktop Commander MCP

    Desktop Commander MCP

    AI-powered MCP server for desktop file and terminal automation

    Desktop Commander MCP is an advanced Model Context Protocol server designed to extend AI assistants with direct control over a user’s local machine, including the file system and terminal. It integrates with clients like Claude Desktop to enable AI-driven workflows such as editing files, executing commands, and automating development tasks from a single conversational interface. Desktop Commander MCP builds on top of an MCP filesystem server and enhances it with powerful search, replace, and code editing capabilities tailored for real-world development environments. It allows users to run terminal commands with streaming output, manage long-running processes, and even execute code in memory without saving files. It also supports working with structured and document formats such as Excel, PDF, and DOCX, enabling AI to read, modify, and generate these files directly.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 3
    Everywhere

    Everywhere

    Context-aware desktop AI assistant that understands screen content

    Everywhere is a context-aware desktop AI assistant designed to interact directly with the content displayed on a user’s screen. It distinguishes itself from traditional AI tools by eliminating the need for manual input methods such as copying text or taking screenshots, instead allowing users to invoke assistance instantly through a shortcut. It can analyze on-screen information in real time and provide contextual responses, making it useful for tasks like troubleshooting errors, summarizing articles, translating text, and refining written content. It integrates with multiple large language model providers and supports various tools, enabling flexible and extensible AI-powered workflows. Everywhere features a modern design with interactive elements such as markdown rendering, keyboard shortcuts, and voice input capabilities. Additionally, the project emphasizes seamless workflow integration by operating alongside existing applications rather than requiring users to switch.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 4
    HumanLayer

    HumanLayer

    Open source IDE for orchestrating AI coding agents in large codebases

    HumanLayer is an open source development environment designed to help developers orchestrate and manage AI coding agents working within complex software projects. It provides a framework and tooling that allow AI agents to research, plan, and implement changes in large codebases while maintaining structured workflows. It focuses on enabling AI-assisted development through coordinated agent workflows rather than isolated code generation tasks. HumanLayer integrates with modern AI models and coding assistants to automate tasks such as code research, planning, and implementation while maintaining a structured development process. HumanLayer introduces advanced context management techniques that help AI agents understand large repositories and operate effectively across multiple tasks. It also supports collaborative workflows where developers and AI agents can work together with human oversight and control.
    Downloads: 3 This Week
    Last Update:
    See Project
  • Rezku Point of Sale Icon
    Rezku Point of Sale

    Designed for Real-World Restaurant Operations

    Rezku is an all-inclusive ordering platform and management solution for all types of restaurant and bar concepts. You can now get a fully custom branded downloadable smartphone ordering app for your restaurant exclusively from Rezku.
    Learn More
  • 5
    LoLLMs WEBUI

    LoLLMs WEBUI

    Local AI WebUI for running and managing large language models offlineA

    lollms-webui is a locally hosted web interface designed to run and manage large language models without relying on external services. It provides users with a centralized environment to interact with multiple AI models, making it suitable for experimentation, development, and personal use. lollms-webui emphasizes offline capability, allowing users to maintain privacy and control over their data while still accessing advanced AI features. It integrates model management tools that help users download, configure, and switch between different language models with ease. It is built to be user-friendly while still offering advanced customization options for power users who want deeper control over model behavior. Additionally, it supports extensibility through plugins or modular components, enabling users to expand functionality as needed. Overall, it serves as a flexible platform for running AI locally with a focus on usability and adaptability.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 6
    Ralph for Claude Code

    Ralph for Claude Code

    Autonomous development loop that iteratively improves projects

    Ralph for Claude Code is an autonomous AI development loop framework designed to continuously iterate on a software project until predefined goals are achieved. It implements a technique that enables Claude Code to repeatedly analyze, modify, and improve a codebase through structured development cycles. It automates the process of running AI-assisted development tasks, allowing the model to progressively refine a project without constant manual intervention. Ralph introduces mechanisms to detect completion signals and determine when the development loop should stop, preventing endless execution cycles. It also includes built-in safeguards such as rate limiting and circuit breaker protections to avoid excessive API usage or runaway processes. Ralph for Claude Code is designed to be installed globally so it can operate as a command-line utility available in any directory, enabling developers to apply autonomous development workflows across multiple projects.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 7
    VGGSfM

    VGGSfM

    VGGSfM: Visual Geometry Grounded Deep Structure From Motion

    VGGSfM is an advanced structure-from-motion (SfM) framework jointly developed by Meta AI Research (GenAI) and the University of Oxford’s Visual Geometry Group (VGG). It reconstructs 3D geometry, dense depth, and camera poses directly from unordered or sequential images and videos. The system combines learned feature matching and geometric optimization to generate high-quality camera calibrations, sparse/dense point clouds, and depth maps in standard COLMAP format. Version 2.0 adds support for dynamic scene handling, dense point cloud export, video-based reconstruction (1000+ frames), and integration with Gaussian Splatting pipelines. It leverages tools like PyCOLMAP, poselib, LightGlue, and PyTorch3D for feature matching, pose estimation, and visualization. With minimal configuration, users can process single scenes or full video sequences, apply motion masks to exclude moving objects, and train neural radiance or splatting models directly from reconstructed outputs.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    CogVLM

    CogVLM

    A state-of-the-art open visual language model

    CogVLM is an open-source visual–language model suite—and its GUI-oriented sibling CogAgent—aimed at image understanding, grounding, and multi-turn dialogue, with optional agent actions on real UI screenshots. The flagship CogVLM-17B combines ~10B visual parameters with ~7B language parameters and supports 490×490 inputs; CogAgent-18B extends this to 1120×1120 and adds plan/next-action outputs plus grounded operation coordinates for GUI tasks. The repo provides multiple ways to run models (CLI, web demo, and OpenAI-Vision–style APIs), along with quantization options that reduce VRAM needs (e.g., 4-bit). It includes checkpoints for chat, base, and grounding variants, plus recipes for model-parallel inference and LoRA fine-tuning. The documentation covers task prompts for general dialogue, visual grounding (box→caption, caption→box, caption+boxes), and GUI agent workflows that produce structured actions with bounding boxes.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 9
    Open SaaS

    Open SaaS

    Open source SaaS boilerplate for React, NodeJS apps with Wasp stack

    Open SaaS is a free and open source starter template designed to help developers quickly build and launch Software-as-a-Service applications. It is built on the Wasp full stack framework, which combines React, NodeJS, and Prisma to manage both client and server code within a unified architecture. Open SaaS provides a production-ready foundation that includes common SaaS functionality such as authentication, payments, analytics, and file uploads. Developers can use it as a boilerplate to avoid writing repetitive setup code and instead focus on building product features. It integrates several commonly used services and tools, including payment processing systems, email providers, analytics platforms, and AI integrations. It also includes an admin dashboard, testing setup, and deployment configuration to streamline development workflows. By bundling these components together, Open SaaS aims to reduce development time and make it easier to create scalable web applications.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Premier Construction Software Icon
    Premier Construction Software

    Premier is a global leader in financial construction ERP software.

    Rated #1 Construction Accounting Software by Forbes Advisor in 2022 & 2023. Our modern SAAS solution is designed to meet the needs of General Contractors, Developers/Owners, Homebuilders & Specialty Contractors.
    Learn More
  • 10
    PySpur

    PySpur

    Visual tool for building, testing, and deploying AI agent workflows

    PySpur is a visual development environment designed to help AI engineers build, test, and iterate on agent-based workflows more efficiently. It provides a structured playground where users can define test cases, construct agents either through Python code or a graphical interface, and continuously refine their behavior. It addresses common challenges in AI agent development such as prompt tuning difficulties and lack of visibility into workflow execution. By offering a visual representation of workflows, PySpur makes it easier to debug interactions between components and identify failures in complex pipelines. It supports iterative experimentation, allowing developers to rapidly improve agents without rebuilding systems from scratch. PySpur also enables deployment of finalized workflows after testing, making it suitable for both development and production use. Overall, it acts as an integrated environment for designing, evaluating, and managing AI-driven processes.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 11
    Super Magic

    Super Magic

    All-in-one AI productivity platform with agents, workflows, and IM

    Magic is an open source all-in-one AI productivity platform designed to help organizations build, deploy, and scale AI-driven applications efficiently. It is not a single tool but a complete product ecosystem composed of multiple integrated systems that work together to enhance productivity across different business scenarios. Magic centers around a general-purpose AI agent system called Super Magic, which can autonomously understand tasks, plan actions, execute workflows, and perform error correction. Alongside this, Magic includes a visual workflow engine that enables users to design complex AI processes using a drag-and-drop interface without requiring extensive coding knowledge. It also provides an enterprise-grade instant messaging system that integrates AI conversations with internal communication, allowing teams to collaborate while leveraging intelligent assistants. Its architecture is built using a microservices approach with containerized services.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 12
    VisualGLM-6B

    VisualGLM-6B

    Chinese and English multimodal conversational language model

    VisualGLM-6B is an open-source multimodal conversational language model developed by ZhipuAI that supports both images and text in Chinese and English. It builds on the ChatGLM-6B backbone, with 6.2 billion language parameters, and incorporates a BLIP2-Qformer visual module to connect vision and language. In total, the model has 7.8 billion parameters. Trained on a large bilingual dataset — including 30 million high-quality Chinese image-text pairs from CogView and 300 million English pairs — VisualGLM-6B is designed for image understanding, description, and question answering. Fine-tuning on long visual QA datasets further aligns the model’s responses with human preferences. The repository provides inference APIs, command-line demos, web demos, and efficient fine-tuning options like LoRA, QLoRA, and P-tuning. It also supports quantization down to INT4, enabling local deployment on consumer GPUs with as little as 6.3 GB VRAM.
    Downloads: 2 This Week
    Last Update:
    See Project
  • 13
    Acontext

    Acontext

    Context data platform for building observable, self-learning AI agents

    Acontext is a cloud-native context data platform designed to support the development and operation of advanced AI agents. It provides a unified system to store and manage contexts, multimodal messages, artifacts, and task workflows, enabling developers to engineer context effectively for their agent products. The platform observes agent tasks and user feedback in real time, offering robust observability into workflows and helping teams understand how agents perform over time. Acontext also supports agent self-learning by distilling structured skills and experiences from previously completed tasks, which can later be reused or searched to improve future performance. It includes tools to interact with session data, background agents that monitor progress, and a dashboard that visualizes success rates, artifacts, and learned skills. By combining persistent storage, observability, and learning capabilities, Acontext aims to make AI agents more scalable, reliable, and capable.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    Bear Stone Smart Home

    Bear Stone Smart Home

    Custom Home Assistant configuration with automations and scripts setup

    Bear Stone Smart Home contains a personalized configuration setup for Home Assistant, an open source home automation platform. It defines how various smart home devices, services, and integrations are organized and controlled within a single environment. It includes configuration files that manage entities such as lights, sensors, switches, and media devices, enabling centralized automation and monitoring. It demonstrates how to structure Home Assistant YAML files for scalability and maintainability in a real-world deployment. Bear Stone Smart Home also showcases custom automations and scripts designed to improve convenience, energy efficiency, and overall smart home behavior. Additionally, it may include examples of dashboards and user interface customization to enhance usability and visualization of home data. Overall, it serves as a practical reference for building and refining a tailored Home Assistant setup.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    ChatGLM2-6B

    ChatGLM2-6B

    ChatGLM2-6B: An Open Bilingual Chat LLM

    ChatGLM2-6B is the second-gen Chinese-English conversational LLM from ZhipuAI/Tsinghua. It upgrades the base model with GLM’s hybrid pretraining objective, 1.4 TB bilingual data, and preference alignment—delivering big gains on MMLU, CEval, GSM8K, and BBH. The context window extends up to 32K (FlashAttention), and Multi-Query Attention improves speed and memory use. The repo includes Python APIs, CLI & web demos, OpenAI-style/FASTAPI servers, and quantized checkpoints for lightweight local deployment on GPUs or CPU/MPS.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Coconut

    Coconut

    Training Large Language Model to Reason in a Continuous Latent Space

    Coconut is the official PyTorch implementation of the research paper “Training Large Language Models to Reason in a Continuous Latent Space.” The framework introduces a novel method for enhancing large language models (LLMs) with continuous latent reasoning steps, enabling them to generate and refine reasoning chains within a learned latent space rather than relying solely on discrete symbolic reasoning. It supports training across multiple reasoning paradigms—including standard Chain-of-Thought (CoT), no-thought, and hybrid configurations—using configurable training stages and latent representations. The repository is built with Hugging Face Transformers, PyTorch Distributed, and Weights & Biases (wandb) for logging, supporting large-scale experiments on mathematical and logical reasoning datasets such as GSM8K, ProntoQA, and ProsQA.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    DeepAnalyze

    DeepAnalyze

    Autonomous LLM agent for end-to-end data science workflows

    DeepAnalyze is an open source project that introduces an agentic large language model designed to perform autonomous data science tasks from start to finish. It is built to handle the entire data science pipeline, including data preparation, analysis, modeling, visualization, and report generation without requiring continuous human guidance. DeepAnalyze is capable of conducting open-ended data research across multiple data formats such as structured tables, semi-structured files, and unstructured text, enabling flexible and comprehensive analysis workflows. It integrates execution-based reasoning by generating and running code as part of its analysis process, allowing it to iteratively refine results and produce more accurate outputs. DeepAnalyze provides multiple interaction interfaces, including a web-based UI, a command-line interface, and a Jupyter-style notebook environment for interactive workflows.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 18
    GitDiagram

    GitDiagram

    AI tool that converts GitHub repositories into interactive diagrams

    GitDiagram is an open source web application designed to help developers quickly understand the structure and architecture of GitHub repositories by automatically generating interactive diagrams. It analyzes repository metadata such as the file tree and project documentation to build a visual representation of how different components of a project relate to one another. It uses an AI-powered pipeline to interpret repository structure and transform that information into system design diagrams rendered with Mermaid visualization. These diagrams provide a high-level overview of a codebase, making it easier for developers to explore unfamiliar projects or understand large and complex repositories. Users can interact with the generated diagrams by clicking components to navigate directly to related files or directories within the repository. GitDiagram combines a modern web frontend with a backend service that processes repository data and generates diagrams dynamically.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 19
    ImageReward

    ImageReward

    [NeurIPS 2023] ImageReward: Learning and Evaluating Human Preferences

    ImageReward is the first general-purpose human preference reward model (RM) designed for evaluating text-to-image generation, introduced alongside the NeurIPS 2023 paper ImageReward: Learning and Evaluating Human Preferences for Text-to-Image Generation. Trained on 137k expert-annotated image pairs, ImageReward significantly outperforms existing scoring methods like CLIP, Aesthetic, and BLIP in capturing human visual preferences. It is provided as a Python package (image-reward) that enables quick scoring of generated images against textual prompts, with APIs for ranking, scoring, and filtering outputs. Beyond evaluation, ImageReward supports Reward Feedback Learning (ReFL), a method for directly fine-tuning diffusion models such as Stable Diffusion using human-preference feedback, leading to demonstrable improvements in image quality.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 20
    OpenAI Harmony

    OpenAI Harmony

    Renderer for the harmony response format to be used with gpt-oss

    Harmony is a response format developed by OpenAI for use with the gpt-oss model series. It defines a structured way for language models to produce outputs, including regular text, reasoning traces, tool calls, and structured data. By mimicking the OpenAI Responses API, Harmony provides developers with a familiar interface while enabling more advanced capabilities such as multiple output channels, instruction hierarchies, and tool namespaces. The format is essential for ensuring gpt-oss models operate correctly, as they are trained to rely on this structure for generating and organizing their responses. For users accessing gpt-oss through third-party providers like HuggingFace, Ollama, or vLLM, Harmony formatting is handled automatically, but developers building custom inference setups must implement it directly. With its flexible design, Harmony serves as the foundation for creating more interpretable, controlled, and extensible interactions with open-weight language models.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    Rig

    Rig

    Rust framework for building modular and scalable LLM-powered apps

    Rig is an open source Rust framework designed to help developers build modular and scalable applications powered by large language models. It provides a unified set of abstractions that allow applications to interact with many AI model providers and vector databases through a single interface. Its architecture emphasizes modularity, enabling developers to integrate only the components and integrations they need for a specific application. Rig includes built-in support for agent workflows, allowing systems to perform multi-turn reasoning, tool calling, and retrieval-based tasks within structured pipelines. It also supports capabilities such as text generation, embeddings, transcription, image generation, and audio generation depending on the provider used. Developers can integrate language models into their software with minimal boilerplate while maintaining flexibility for complex AI workflows.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 22
    SenseVoice

    SenseVoice

    Multilingual speech recognition and audio understanding model

    SenseVoice is a speech foundation model designed to perform multiple voice understanding tasks from audio input. It provides capabilities such as automatic speech recognition, spoken language identification, speech emotion recognition, and audio event detection within a single system. SenseVoice is trained on more than 400,000 hours of speech data and supports over 50 languages for multilingual recognition tasks. It is built to achieve high transcription accuracy while maintaining efficient inference performance. It includes different model variants optimized for either speed or accuracy, allowing developers to choose a configuration suitable for their use case. In addition to speech transcription, SenseVoice can detect emotional cues in speech and identify common sound events such as applause, laughter, or coughing. It also provides tools for running inference, exporting models to formats like ONNX or LibTorch, and deploying the system through APIs.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 23
    Stanford Machine Learning Course

    Stanford Machine Learning Course

    machine learning course programming exercise

    The Stanford Machine Learning Course Exercises repository contains programming assignments from the well-known Stanford Machine Learning online course. It includes implementations of a variety of fundamental algorithms using Python and MATLAB/Octave. The repository covers a broad set of topics such as linear regression, logistic regression, neural networks, clustering, support vector machines, and recommender systems. Each folder corresponds to a specific algorithm or concept, making it easy for learners to navigate and practice. The exercises serve as practical, hands-on reinforcement of theoretical concepts taught in the course. This collection is valuable for students and practitioners who want to strengthen their skills in machine learning through coding exercises.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 24
    Tracking Any Point (TAP)

    Tracking Any Point (TAP)

    DeepMind model for tracking arbitrary points across videos & robotics

    TAPNet is the official Google DeepMind repository for Tracking Any Point (TAP), bundling datasets, models, benchmarks, and demos for precise point tracking in videos. The project includes the TAP-Vid and TAPVid-3D benchmarks, which evaluate long-range tracking of arbitrary points in 2D and 3D across diverse real and synthetic videos. Its flagship models—TAPIR, BootsTAPIR, and the latest TAPNext—use matching plus temporal refinement or next-token style propagation to achieve state-of-the-art accuracy and speed on TAP-Vid. RoboTAP demonstrates how TAPIR-style tracks can drive real-world robot manipulation via efficient imitation, and ships with a dataset of annotated robotics videos. The repo provides JAX and PyTorch checkpoints, Colab demos, and a real-time live demo that runs on a GPU to let you select and track points interactively.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 25
    muse

    muse

    AI agent memory system—pure Markdown, zero dependencies, fully local

    MUSE gives AI coding agents persistent cross-session memory and multi-role governance through plain Markdown files. Supports Claude Code, OpenClaw, Cursor, Windsurf, Gemini CLI, and Codex via one-command install. Built-in MCP Server for programmatic access. 56 skills, auto memory capture, semantic compression, role-based governance, multi-project management. Pure Markdown, no database, no cloud. MIT open source.
    Downloads: 17 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB