Open Source Windows Object Detection Models - Page 2

Object Detection Models for Windows

View 1843 business solutions
  • Award-Winning Medical Office Software Designed for Your Specialty Icon
    Award-Winning Medical Office Software Designed for Your Specialty

    Succeed and scale your practice with cloud-based, data-backed, AI-powered healthcare software.

    RXNT is an ambulatory healthcare technology pioneer that empowers medical practices and healthcare organizations to succeed and scale through innovative, data-backed, AI-powered software.
    Learn More
  • Outbound sales software Icon
    Outbound sales software

    Unified cloud-based platform for dialing, emailing, appointment scheduling, lead management and much more.

    Adversus is an outbound dialing solution that helps you streamline your call strategies, automate manual processes, and provide valuable insights to improve your outbound workflows and efficiency.
    Learn More
  • 1
    Unofficial Action Cam Yi Desktop

    Unofficial Action Cam Yi Desktop

    Unofficial Action Camera Yi Desktop App

    Action Camera Yi Desktop App 2nd Version is an application that can be used to control Action Camera Yi through Desktop / Computer. This application offers faster data processing. In the Features section, You can see the features offered by this app. This app has the following advantages : -Low Latency Live Streaming.You can use this application for streaming to Streaming Provider (i.e. Youtube). -Adjustable Live Streaming Video’s FPS (Frame Per Second). -Selectable Image Processing for Live Streaming Video using Faster Image Processing. -Download Photo and Video from Album Feature, directly from Computer. -Configurable Action Camera Yi’s Setting from Settings Feature. -Object Detection -Simple Face Detection More Details, New Updates: http://bit.ly/2QHnRHS
    Downloads: 2 This Week
    Last Update:
    See Project
  • 2
    FastQR

    FastQR

    A Fast QR code detector for arbitrarily acquired images

    Applications of Quick Response (QR) codes enable rich context interaction through creation of links between physical objects and internet resources. In spite of the widespread use of this kind of barcode, applications for visually impaired people and robots are not common because existing decoders assume that the symbol is properly framed during image acquisition. This project implements a two-stage component-based approach to perform accurate detection of QR code symbols in arbitrarily acquired images. In the first stage a cascade classifier to detect parts of the symbol is trained using the rapid object detection framework proposed by Viola-Jones. In the second stage, detected patterns are aggregated in order to evaluate if they are spatially arranged in a way that is geometrically consistent with the components of a QR code symbol. OpenCV 2.2+ is required.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 3
    playqt

    playqt

    GUI version of ffplay for Windows

    playqt is a Windows GUI version of the well known ffplay program and has been enhanced with Object Detection capabilities. It can process multiple types of media including real time streams. An integrated camera control feature allows control over the camera parameters as well as automatic network configuration and connection. See the README under the files tab for configuration info. Real time object counting is implemented using YOLO detection algorithm. The program can be used with standard or customized models. A reduced version of the COCO dataset for most commonly observed types is available here. The program is based on ffplay and will respond to the familiar options if launched from the command line. This allows the program to be used with other command line tools such as youtube-dl. The source code is open and available here. It may be compiled using the contrib library provided along with Qt6, MSVC 2019 and NVIDIA cuda development library.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 4
    COCO Annotator

    COCO Annotator

    Web-based image segmentation tool for object detection & localization

    COCO Annotator is a web-based image annotation tool designed for versatility and efficiently label images to create training data for image localization and object detection. It provides many distinct features including the ability to label an image segment (or part of a segment), track object instances, label objects with disconnected visible parts, and efficiently store and export annotations in the well-known COCO format. The annotation process is delivered through an intuitive and customizable interface and provides many tools for creating accurate datasets. Several annotation tools are currently available, with most applications as a desktop installation. Once installed, users can manually define regions in an image and creating a textual description. Generally, objects can be marked by a bounding box, either directly, through a masking tool, or by marking points to define the containing area. COCO Annotator allows users to annotate images using free-form curves.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Iris Powered By Generali - Iris puts your customer in control of their identity. Icon
    Iris Powered By Generali - Iris puts your customer in control of their identity.

    Increase customer and employee retention by offering Onwatch identity protection today.

    Iris Identity Protection API sends identity monitoring and alerts data into your existing digital environment – an ideal solution for businesses that are looking to offer their customers identity protection services without having to build a new product or app from scratch.
    Learn More
  • 5
    ChainerCV

    ChainerCV

    ChainerCV: a Library for Deep Learning in Computer Vision

    ChainerCV is a collection of tools to train and run neural networks for computer vision tasks using Chainer. In ChainerCV, we define the object detection task as a problem of, given an image, bounding box-based localization and categorization of objects. Bounding boxes in an image are represented as a two-dimensional array of shape (R,4), where R is the number of bounding boxes and the second axis corresponds to the coordinates of bounding boxes. ChainerCV supports dataset loaders, which can be used to easily index examples with list-like interfaces. Dataset classes whose names end with BboxDataset contain annotations of where objects locate in an image and which categories they are assigned to. These datasets can be indexed to return a tuple of an image, bounding boxes and labels. ChainerCV provides several network implementations that carry out object detection.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Command Line Parser GetPot

    Command Line Parser GetPot

    Tool to parse the command line and configuration files.

    Powerful command line and configuration file parsing for C++, Python, Ruby and Java (others to come). This tool provides many features, such as separate treatment for options, variables, and flags, unrecognized object detection, prefixes and much more.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Computer Vision Pretrained Models

    Computer Vision Pretrained Models

    A collection of computer vision pre-trained models

    A pre-trained model is a model created by someone else to solve a similar problem. Instead of building a model from scratch to solve a similar problem, we can use the model trained on other problem as a starting point. A pre-trained model may not be 100% accurate in your application. For example, if you want to build a self-learning car. You can spend years building a decent image recognition algorithm from scratch or you can take the inception model (a pre-trained model) from Google which was built on ImageNet data to identify images in those pictures. The model generates bounding boxes and segmentation masks for each instance of an object in the image. It's based on Feature Pyramid Network (FPN) and a ResNet101 backbone. TensorFlow implementation of 'YOLO: Real-Time Object Detection', with training and an actual support for real-time running on mobile devices. MobileNets trade off between latency, size and accuracy while comparing favorably with popular models from the literature.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    CutLER

    CutLER

    Code release for Cut and Learn for Unsupervised Object Detection

    CutLER is an approach for unsupervised object detection and instance segmentation that trains detectors without human-annotated labels, and the repo also includes VideoCutLER for unsupervised video instance segmentation. The method follows a “Cut-and-LEaRn” recipe: bootstrap object proposals, refine them iteratively, and train detection/segmentation heads to discover objects across diverse datasets. The codebase provides training and inference scripts, model configs, and references to benchmarking results that report large gains over prior unsupervised baselines. It’s intended for researchers exploring self-supervised and unsupervised recognition, offering a practical path to scale beyond costly labeled corpora. The README links papers and gives a high-level overview of components and expected outputs, with pointers to demos and assets. The repository is actively starred and structured as a typical research release with license, contribution guidelines, and security policy.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    DETR

    DETR

    End-to-end object detection with transformers

    PyTorch training code and pretrained models for DETR (DEtection TRansformer). We replace the full complex hand-crafted object detection pipeline with a Transformer, and match Faster R-CNN with a ResNet-50, obtaining 42 AP on COCO using half the computation power (FLOPs) and the same number of parameters. Inference in 50 lines of PyTorch. What it is. Unlike traditional computer vision techniques, DETR approaches object detection as a direct set prediction problem. It consists of a set-based global loss, which forces unique predictions via bipartite matching, and a Transformer encoder-decoder architecture. Given a fixed small set of learned object queries, DETR reasons about the relations of the objects and the global image context to directly output the final set of predictions in parallel. Due to this parallel nature, DETR is very fast and efficient.
    Downloads: 0 This Week
    Last Update:
    See Project
  • The AI workplace management platform Icon
    The AI workplace management platform

    Plan smart spaces, connect teams, manage assets, and get insights with the leading AI-powered operating system for the built world.

    By combining AI workflows, predictive intelligence, and automated insights, OfficeSpace gives leaders a complete view of how their spaces are used and how people work. Facilities, IT, HR, and Real Estate teams use OfficeSpace to optimize space utilization, enhance employee experience, and reduce portfolio costs with precision.
    Learn More
  • 10
    DIGITS

    DIGITS

    Deep Learning GPU training system

    The NVIDIA Deep Learning GPU Training System (DIGITS) puts the power of deep learning into the hands of engineers and data scientists. DIGITS can be used to rapidly train the highly accurate deep neural network (DNNs) for image classification, segmentation and object detection tasks. DIGITS simplifies common deep learning tasks such as managing data, designing and training neural networks on multi-GPU systems, monitoring performance in real-time with advanced visualizations, and selecting the best performing model from the results browser for deployment. DIGITS is completely interactive so that data scientists can focus on designing and training networks rather than programming and debugging. DIGITS is available as a free download to the members of the NVIDIA Developer Program. DIGITS is available on NVIDIA GPU Cloud (NGC) as an optimized container for on-demand usage. Sign-up for an NGC account and get started with DIGITS in minutes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Deep Learning course

    Deep Learning course

    Slides and Jupyter notebooks for the Deep Learning lectures

    Slides and Jupyter notebooks for the Deep Learning lectures at Master Year 2 Data Science from Institut Polytechnique de Paris. This course is being taught at as part of Master Year 2 Data Science IP-Paris. Note: press "P" to display the presenter's notes that include some comments and additional references. This lecture is built and maintained by Olivier Grisel and Charles Ollion.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    DeepDetect

    DeepDetect

    Deep Learning API and Server in C++14 support for Caffe, PyTorch

    The core idea is to remove the error sources and difficulties of Deep Learning applications by providing a safe haven of commoditized practices, all available as a single core. While the Open Source Deep Learning Server is the core element, with REST API, and multi-platform support that allows training & inference everywhere, the Deep Learning Platform allows higher level management for training neural network models and using them as if they were simple code snippets. Ready for applications of image tagging, object detection, segmentation, OCR, Audio, Video, Text classification, CSV for tabular data and time series. Neural network templates for the most effective architectures for GPU, CPU, and Embedded devices. Training in a few hours and with small data thanks to 25+ pre-trained models. Full Open Source, with an ecosystem of tools (API clients, video, annotation, ...) Fast Server written in pure C++, a single codebase for Cloud, Desktop & Embedded.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    DetectAndTrack

    DetectAndTrack

    The implementation of an algorithm presented in the CVPR18 paper

    DetectAndTrack is the reference implementation for the CVPR 2018 paper “Detect-and-Track: Efficient Pose Estimation in Videos,” focusing on human keypoint detection and tracking across video frames. The system combines per-frame pose detection with a tracking mechanism to maintain identities over time, enabling efficient multi-person pose estimation in video. Code and instructions are organized to replicate paper results and to serve as a starting point for researchers working on pose in video. Although the repo has been archived and is now read-only, its issue tracker and artifacts remain useful for understanding implementation details and experimental settings. The project sits alongside other Facebook Research vision efforts, offering historical context for the evolution of video pose and tracking techniques. Researchers can still study the algorithms, adapt the pipeline, or port ideas into modern frameworks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Detic

    Detic

    Code release for "Detecting Twenty-thousand Classes

    Detic (“Detecting Twenty-thousand Classes using Image-level Supervision”) is a large-vocabulary object detector that scales beyond fully annotated datasets by leveraging image-level labels. It decouples localization from classification, training a strong box localizer on standard detection data while learning classifiers from weak supervision and large image-tag corpora. A shared region proposal backbone feeds a flexible classification head that can expand to tens of thousands of categories without exhaustive box annotations. The system supports zero- or few-shot extension to novel categories via semantic embeddings and class name supervision, making “open-world” detection practical. Built on Detectron2, the repo includes configs, pretrained weights, and conversion tools to mix fully and weakly supervised sources. Detic is especially useful for applications where label space is vast and long-tailed, but dense bounding-box annotation is infeasible.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Gluon CV Toolkit

    Gluon CV Toolkit

    Gluon CV Toolkit

    GluonCV provides implementations of state-of-the-art (SOTA) deep learning algorithms in computer vision. It aims to help engineers, researchers, and students quickly prototype products, validate new ideas and learn computer vision. It features training scripts that reproduce SOTA results reported in latest papers, a large set of pre-trained models, carefully designed APIs and easy-to-understand implementations and community support. From fundamental image classification, object detection, semantic segmentation and pose estimation, to instance segmentation and video action recognition. The model zoo is the one-stop shopping center for many models you are expecting. GluonCV embraces a flexible development pattern while is super easy to optimize and deploy without retaining a heavyweight deep learning framework.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16

    LibPaBOD

    LibPaBOD: a LIBrary for PArt-Based Object Detection in C++

    LibPaBOD is a library written in C++ that allows to perform object detection on still images. It implements the object detection method proposed by Felzenszwalb et al. in PAMI, 2010.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    LifeAI is an artificial intelligence system that can be applied to robotics, games, or business. It simulates key processes of our minds, such as organizing data into concepts and categories, planning actions based on their predicted outcome, and communication. LifeAI was designed to be simple, but powerful and flexible enough to have many applications.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18

    LiveVideo

    Real-time Video Analysis Software

    This is an open-source real-time object detection and tracking software for H.264/AVC videos. It applies probabilistic spatiotemporal macroblock filtering (PSMF) and partial decoding processes to effectively detect and track multiple objects with fast computation in H.264|AVC bitstreams with stationary background. The codes were written in Visual C++. For more details, please visit https://www.wonsangyou.com/research/aivision. If you need technical help, please send an email to wyou(at)kaist.ac.kr. When you use this software for your publications, please cite as follows. 1. Wonsang You, M.S. Houari Sabirin, and Munchurl Kim, "Moving object tracking in H.264/AVC bitstream," Lecture Notes in Computer Science, vol. 4577, 2007, pp. 483-492. 2. Wonsang You, M.S. Houari Sabirin, and Munchurl Kim, "Real-time detection and tracking of multiple objects with partial decoding in H.264/AVC bitstream domain," Proceedings of SPIE, vol. 7244, 72440D (February 2009).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    ML.NET

    ML.NET

    Open source and cross-platform machine learning framework for .NET

    With ML.NET, you can create custom ML models using C# or F# without having to leave the .NET ecosystem. ML.NET lets you re-use all the knowledge, skills, code, and libraries you already have as a .NET developer so that you can easily integrate machine learning into your web, mobile, desktop, games, and IoT apps. ML.NET offers Model Builder (a simple UI tool) and ML.NET CLI to make it super easy to build custom ML Models. These tools use Automated ML (AutoML), a cutting edge technology that automates the process of building best performing models for your Machine Learning scenario. All you have to do is load your data, and AutoML takes care of the rest of the model building process. ML.NET has been designed as an extensible platform so that you can consume other popular ML frameworks (TensorFlow, ONNX, Infer.NET, and more) and have access to even more machine learning scenarios, like image classification, object detection, and more.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    MMTracking

    MMTracking

    OpenMMLab Video Perception Toolbox

    MMTracking is an open-source video perception toolbox by PyTorch. It is a part of OpenMMLab project. We are the first open-source toolbox that unifies versatile video perception tasks include video object detection, multiple object tracking, single object tracking and video instance segmentation. We decompose the video perception framework into different components and one can easily construct a customized method by combining different modules. MMTracking interacts with other OpenMMLab projects. It is built upon MMDetection that we can capitalize any detector only through modifying the configs. All operations run on GPUs. The training and inference speeds are faster than or comparable to other implementations. We reproduce state-of-the-art models and some of them even outperform the official implementations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    Monk Computer Vision

    Monk Computer Vision

    A low code unified framework for computer vision and deep learning

    Monk is an open source low code programming environment to reduce the cognitive load faced by entry level programmers while catering to the needs of Expert Deep Learning engineers. There are three libraries in this opensource set. - Monk Classiciation- https://monkai.org. A Unified wrapper over major deep learning frameworks. Our core focus area is at the intersection of Computer Vision and Deep Learning algorithms. - Monk Object Detection - https://github.com/Tessellate-Imaging/Monk_Object_Detection. Monk object detection is our take on assembling state of the art object detection, image segmentation, pose estimation algorithms at one place, making them low code and easily configurable on any machine. - Monk GUI - https://github.com/Tessellate-Imaging/Monk_Gui. An interface over these low code tools for non coders.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
     Object detection is usually a software-based monitoring algorithm that will signal, for example in the surveillance camera to begin capturing the event when it detects motion. In object tracking, the object is located and the moving object is followed. One of the fundamental steps in many computer based vision systems for object tracking and motion detection is real-time segmentation of moving regions in the image sequences. Segmentation is done in order to detect the object accurately. Usually cameras are used as input sensors, for recording.  Front end- MATLAB
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    Norfair

    Norfair

    Lightweight Python library for adding real-time multi-object tracking

    Norfair is a customizable lightweight Python library for real-time multi-object tracking. Using Norfair, you can add tracking capabilities to any detector with just a few lines of code. Any detector expressing its detections as a series of (x, y) coordinates can be used with Norfair. This includes detectors performing tasks such as object or keypoint detection. It can easily be inserted into complex video processing pipelines to add tracking to existing projects. At the same time, it is possible to build a video inference loop from scratch using just Norfair and a detector. Supports moving camera, re-identification with appearance embeddings, and n-dimensional object tracking. Norfair provides several predefined distance functions to compare tracked objects and detections. The distance functions can also be defined by the user, enabling the implementation of different tracking strategies.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24

    ObjectDetector

    Car Detection,Face Detectiom,Object Detection

    Machine learning: This project is used for training new object like Car,Motor Cycle and so on and we use this model(xml file) for detecting in images.In this project we use viola jones algorithm.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Objectron

    Objectron

    A dataset of short, object-centric video clips

    The Objectron dataset is a collection of short, object-centric video clips, which are accompanied by AR session metadata that includes camera poses, sparse point-clouds and characterization of the planar surfaces in the surrounding environment. In each video, the camera moves around the object, capturing it from different angles. The data also contain manually annotated 3D bounding boxes for each object, which describe the object’s position, orientation, and dimensions. The dataset consists of 15K annotated video clips supplemented with over 4M annotated images in the following categories: bikes, books, bottles, cameras, cereal boxes, chairs, cups, laptops, and shoes. In addition, to ensure geo-diversity, our dataset is collected from 10 countries across five continents. Along with the dataset, we are also sharing a 3D object detection solution for four categories of objects — shoes, chairs, mugs, and cameras.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB