Open Source Mac Deep Learning Frameworks - Page 5

Deep Learning Frameworks for Mac

View 17 business solutions
  • AestheticsPro Medical Spa Software Icon
    AestheticsPro Medical Spa Software

    Our new software release will dramatically improve your medspa business performance while enhancing the customer experience

    AestheticsPro is the most complete Aesthetics Software on the market today. HIPAA Cloud Compliant with electronic charting, integrated POS, targeted marketing and results driven reporting; AestheticsPro delivers the tools you need to manage your medical spa business. It is our mission To Provide an All-in-One Cutting Edge Software to the Aesthetics Industry.
    Learn More
  • Rezku Point of Sale Icon
    Rezku Point of Sale

    Designed for Real-World Restaurant Operations

    Rezku is an all-inclusive ordering platform and management solution for all types of restaurant and bar concepts. You can now get a fully custom branded downloadable smartphone ordering app for your restaurant exclusively from Rezku.
    Learn More
  • 1
    ChainerRL

    ChainerRL

    ChainerRL is a deep reinforcement learning library

    ChainerRL (this repository) is a deep reinforcement learning library that implements various state-of-the-art deep reinforcement algorithms in Python using Chainer, a flexible deep learning framework. PFRL is the PyTorch analog of ChainerRL. ChainerRL has a set of accompanying visualization tools in order to aid developers' ability to understand and debug their RL agents. With this visualization tool, the behavior of ChainerRL agents can be easily inspected from a browser UI. Environments that support the subset of OpenAI Gym's interface (reset and step methods) can be used.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    ConvNetJS

    ConvNetJS

    Deep learning in Javascript to train convolutional neural networks

    ConvNetJS is a Javascript library for training Deep Learning models (Neural Networks) entirely in your browser. Open a tab and you're training. No software requirements, no compilers, no installations, no GPUs, no sweat. ConvNetJS is an implementation of Neural networks, together with nice browser-based demos. It currently supports common Neural Network modules (fully connected layers, non-linearities), classification (SVM/Softmax) and Regression (L2) cost functions, ability to specify and train Convolutional Networks that process images, and experimental Reinforcement Learning modules, based on Deep Q Learning. The library allows you to formulate and solve Neural Networks in Javascript. If you would like to add features to the library, you will have to change the code in src/ and then compile the library into the build/ directory. The compilation script simply concatenates files in src/ and then minifies the result.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Convolution arithmetic

    Convolution arithmetic

    A technical report on convolution arithmetic in deep learning

    A technical report on convolution arithmetic in the context of deep learning. The code and the images of this tutorial are free to use as regulated by the licence and subject to proper attribution. The animations will be output to the gif directory. Individual animation steps will be output in PDF format to the pdf directory and in PNG format to the png directory. We introduce a guide to help deep learning practitioners understand and manipulate convolutional neural network architectures. The guide clarifies the relationship between various properties (input shape, kernel shape, zero padding, strides and output shape) of convolutional, pooling and transposed convolutional layers, as well as the relationship between convolutional and transposed convolutional layers. Relationships are derived for various cases, and are illustrated in order to make them intuitive.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    DGL

    DGL

    Python package built to ease deep learning on graph

    Build your models with PyTorch, TensorFlow or Apache MXNet. Fast and memory-efficient message passing primitives for training Graph Neural Networks. Scale to giant graphs via multi-GPU acceleration and distributed training infrastructure. DGL empowers a variety of domain-specific projects including DGL-KE for learning large-scale knowledge graph embeddings, DGL-LifeSci for bioinformatics and cheminformatics, and many others. We are keen to bringing graphs closer to deep learning researchers. We want to make it easy to implement graph neural networks model family. We also want to make the combination of graph based modules and tensor based modules (PyTorch or MXNet) as smooth as possible. DGL provides a powerful graph object that can reside on either CPU or GPU. It bundles structural data as well as features for a better control. We provide a variety of functions for computing with graph objects including efficient and customizable message passing primitives for Graph Neural Networks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • The Most Powerful Software Platform for EHSQ and ESG Management Icon
    The Most Powerful Software Platform for EHSQ and ESG Management

    Addresses the needs of small businesses and large global organizations with thousands of users in multiple locations.

    Choose from a complete set of software solutions across EHSQ that address all aspects of top performing Environmental, Health and Safety, and Quality management programs.
    Learn More
  • 5
    DIG

    DIG

    A library for graph deep learning research

    The key difference with current graph deep learning libraries, such as PyTorch Geometric (PyG) and Deep Graph Library (DGL), is that, while PyG and DGL support basic graph deep learning operations, DIG provides a unified testbed for higher level, research-oriented graph deep learning tasks, such as graph generation, self-supervised learning, explainability, 3D graphs, and graph out-of-distribution. If you are working or plan to work on research in graph deep learning, DIG enables you to develop your own methods within our extensible framework, and compare with current baseline methods using common datasets and evaluation metrics without extra efforts. It includes unified implementations of data interfaces, common algorithms, and evaluation metrics for several advanced tasks. Our goal is to enable researchers to easily implement and benchmark algorithms.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    DLRM

    DLRM

    An implementation of a deep learning recommendation model (DLRM)

    DLRM (Deep Learning Recommendation Model) is Meta’s open-source reference implementation for large-scale recommendation systems built to handle extremely high-dimensional sparse features and embedding tables. The architecture combines dense (MLP) and sparse (embedding) branches, then interacts features via dot product or feature interactions before passing through further dense layers to predict click-through, ranking scores, or conversion probabilities. The implementation is optimized for performance at scale, supporting multi-GPU and multi-node execution, quantization, embedding partitioning, and pipelined I/O to feed huge embeddings efficiently. It includes data loaders for standard benchmarks (like Criteo), training scripts, evaluation tools, and capabilities like mixed precision, gradient compression, and memory fusion to maximize throughput.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Darts

    Darts

    A python library for easy manipulation and forecasting of time series

    darts is a Python library for easy manipulation and forecasting of time series. It contains a variety of models, from classics such as ARIMA to deep neural networks. The models can all be used in the same way, using fit() and predict() functions, similar to scikit-learn. The library also makes it easy to backtest models, combine the predictions of several models, and take external data into account. Darts supports both univariate and multivariate time series and models. The ML-based models can be trained on potentially large datasets containing multiple time series, and some of the models offer a rich support for probabilistic forecasting. We recommend to first setup a clean Python environment for your project with at least Python 3.7 using your favorite tool (conda, venv, virtualenv with or without virtualenvwrapper).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    Datasets

    Datasets

    Hub of ready-to-use datasets for ML models

    Datasets is a library for easily accessing and sharing datasets, and evaluation metrics for Natural Language Processing (NLP), computer vision, and audio tasks. Load a dataset in a single line of code, and use our powerful data processing methods to quickly get your dataset ready for training in a deep learning model. Backed by the Apache Arrow format, process large datasets with zero-copy reads without any memory constraints for optimal speed and efficiency. We also feature a deep integration with the Hugging Face Hub, allowing you to easily load and share a dataset with the wider NLP community. There are currently over 2658 datasets, and more than 34 metrics available. Datasets naturally frees the user from RAM memory limitation, all datasets are memory-mapped using an efficient zero-serialization cost backend (Apache Arrow). Smart caching: never wait for your data to process several times.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    DeeProtGO

    DeeProtGO

    DeeProtGO is a deep learning model for predicting GO terms of proteins

    This project contains the source code of DeeProtGO as well as an example of its use when predicting GO terms of the biological process sub-ontology for eukaryotic proteins.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Transforming NetOps Through No-Code Network Automation - NetBrain Icon
    Transforming NetOps Through No-Code Network Automation - NetBrain

    For anyone searching for a complete no-code automation platform for hybrid network observability and AIOps

    NetBrain, founded in 2004, provides a powerful no-code automation platform for hybrid network observability, allowing organizations to enhance their operational efficiency through automated workflows. The platform applies automation across three key workflows: troubleshooting, change management, and assessment.
    Learn More
  • 10
    Deep Learning

    Deep Learning

    Deep Learning Book Chinese Translation

    With the help and proofreading of many netizens, the Chinese version was finally published. Although there are still many problems, at least 90% of the content is readable and accurate. We have preserved the meaning of the original book Deep Learning as much as possible and retained the original language of the book. However, our level is limited, and we cannot eliminate the variance of many readers. We still need everyone's advice and help to reduce translation bias together. All you have to do is read, then aggregate your suggestions and raise issues (preferably not one by one). If you are sure that your suggestion does not need to be discussed, you can directly initiate a PR. Please download the PDF directly to read. We do not intend to provide formats such as EPUB. Please modify it yourself if necessary.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Deep Learning Drizzle

    Deep Learning Drizzle

    Drench yourself in Deep Learning, Reinforcement Learning

    Drench yourself in Deep Learning, Reinforcement Learning, Machine Learning, Computer Vision, and NLP by learning from these exciting lectures! Optimization courses which form the foundation for ML, DL, RL. Computer Vision courses which are DL & ML heavy. Speech recognition courses which are DL heavy. Structured Courses on Geometric, Graph Neural Networks. Section on Autonomous Vehicles. Section on Computer Graphics with ML/DL focus.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Deep Learning Is Nothing

    Deep Learning Is Nothing

    Deep learning concepts in an approachable style

    Deep-Learning-Is-Nothing presents deep learning concepts in an approachable, from-scratch style that demystifies the stack behind modern models. It typically begins with linear algebra, calculus, and optimization refreshers before moving to perceptrons, multilayer networks, and gradient-based training. Implementations favor small, readable examples—often NumPy first—to show how forward and backward passes work without depending solely on high-level frameworks. Once the fundamentals are clear, the material extends to CNNs, RNNs, and attention mechanisms, explaining why each architecture suits particular tasks. Practical sections cover data pipelines, regularization, and evaluation, emphasizing reproducibility and debugging techniques. The goal is to replace buzzwords with intuition so learners can reason about architectures and training dynamics with confidence.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Deep Learning Models

    Deep Learning Models

    A collection of various deep learning architectures, models, and tips

    This repository collects clear, well-documented implementations of deep learning models and training utilities written by Sebastian Raschka. The code favors readability and pedagogy: components are organized so you can trace data flow through layers, losses, optimizers, and evaluation. Examples span fundamental architectures—MLPs, CNNs, RNN/Transformers—and practical tasks like image classification or text modeling. Reproducible training scripts and configuration files make it straightforward to rerun experiments or adapt them to your own datasets. The repo often pairs implementations with notes on design choices and trade-offs, turning it into both a toolbox and a learning resource. It’s suitable for students, researchers prototyping ideas, and practitioners who want clean baselines before adding complexity.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Deep Learning Papers Reading Roadmap

    Deep Learning Papers Reading Roadmap

    Deep Learning papers reading roadmap for anyone who are eager to learn

    Deep Learning Papers Reading Roadmap is a widely known curated reading plan for deep learning that helps newcomers and practitioners navigate the vast literature in a structured and intentional way. It is built around several guiding principles: moving from outline to detail, from older foundational papers to state-of-the-art work, and from generic to more specialized areas while keeping a focus on impactful contributions. The roadmap organizes papers into categories such as fundamentals, convolutional networks, sequence models, unsupervised learning, generative models, optimization, and application areas like computer vision or NLP. For each section, it suggests an order that lets readers gradually build intuition and then dive deeper into more advanced or recent topics. It is particularly useful for students and engineers who want to systematically improve their understanding rather than randomly picking papers.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Deep Learning course

    Deep Learning course

    Slides and Jupyter notebooks for the Deep Learning lectures

    Slides and Jupyter notebooks for the Deep Learning lectures at Master Year 2 Data Science from Institut Polytechnique de Paris. This course is being taught at as part of Master Year 2 Data Science IP-Paris. Note: press "P" to display the presenter's notes that include some comments and additional references. This lecture is built and maintained by Olivier Grisel and Charles Ollion.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Deep Learning with Keras and Tensorflow

    Deep Learning with Keras and Tensorflow

    Introduction to Deep Neural Networks with Keras and Tensorflow

    Introduction to Deep Neural Networks with Keras and Tensorflow. To date tensorflow comes in two different packages, namely tensorflow and tensorflow-gpu, whether you want to install the framework with CPU-only or GPU support, respectively. NVIDIA Drivers and CuDNN must be installed and configured before hand. Please refer to the official Tensorflow documentation for further details. Since version 0.9 Theano introduced the libgpuarray in the stable release (it was previously only available in the development version). The goal of libgpuarray is (from the documentation) make a common GPU ndarray (n dimensions array) that can be reused by all projects that is as future proof as possible, while keeping it easy to use for simple need/quick test. The easiest way to get (most) these is to use an all-in-one installer such as Anaconda from Continuum. These are available for multiple architectures.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    Deep Learning with PyTorch

    Deep Learning with PyTorch

    Latest techniques in deep learning and representation learning

    This course concerns the latest techniques in deep learning and representation learning, focusing on supervised and unsupervised deep learning, embedding methods, metric learning, convolutional and recurrent nets, with applications to computer vision, natural language understanding, and speech recognition. The prerequisites include DS-GA 1001 Intro to Data Science or a graduate-level machine learning course. To be able to follow the exercises, you are going to need a laptop with Miniconda (a minimal version of Anaconda) and several Python packages installed. The following instruction would work as is for Mac or Ubuntu Linux users, Windows users would need to install and work in the Git BASH terminal. JupyterLab has a built-in selectable dark theme, so you only need to install something if you want to use the classic notebook interface.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Deep Reinforcement Learning TensorFlow

    Deep Reinforcement Learning TensorFlow

    TensorFlow implementation of Deep Reinforcement Learning papers

    Deep Reinforcement Learning TensorFlow is a comprehensive TensorFlow codebase that implements several foundational deep reinforcement learning algorithms for educational and experimental use. The repository focuses on clarity and modularity so users can study how different RL approaches are built and compare their behavior across environments. It includes implementations of well-known algorithms such as Deep Q-Networks (DQN), policy gradients, and related variants, demonstrating how neural networks can be trained through interaction with simulated environments. The project is commonly used by learners who want to move beyond theory and understand the practical mechanics of training RL agents. Visualization utilities and training scripts help users monitor learning progress and debug experiments.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Deep learning time series forecasting

    Deep learning time series forecasting

    Deep learning PyTorch library for time series forecasting

    Example image Flow Forecast (FF) is an open-source deep learning for time series forecasting framework. It provides all the latest state-of-the-art models (transformers, attention models, GRUs) and cutting-edge concepts with easy-to-understand interpretability metrics, cloud provider integration, and model serving capabilities. Flow Forecast was the first time series framework to feature support for transformer-based models and remains the only true end-to-end deep learning for time series forecasting framework. Currently, Task-TS from CoronaWhy primarily maintains this repository. Pull requests are welcome. Historically, this repository provided open-source benchmarks and codes for flash flood and river flow forecasting. Full transformer (SimpleTransformer in model_dict): The full original transformer with all 8 encoder and decoder blocks. Requires passing the target in at inference.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    Deep-Learning-Interview-Book

    Deep-Learning-Interview-Book

    Interview guide for machine learning, mathematics, and deep learning

    Deep-Learning-Interview-Book collects structured notes, Q&A, and concept summaries tailored to deep-learning interviews, turning scattered study into a coherent playbook. It spans the core math (linear algebra, probability, optimization) and the practitioner topics candidates actually face, like CNNs, RNNs/Transformers, attention, regularization, and training tricks. Explanations emphasize intuition first, then key formulas and common pitfalls, so you can reason through unseen questions rather than memorize trivia. Many entries connect theory to implementation details, including how choices in activation, initialization, or normalization affect convergence and stability. The content is organized for fast review before an interview loop but is also deep enough for systematic study over weeks. Because it’s text-first and modular, it works equally well as a quick refresher or a backbone for a full study plan.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    DeepCTR-Torch

    DeepCTR-Torch

    Easy-to-use,Modular and Extendible package of deep-learning models

    DeepCTR-Torch is an easy-to-use, Modular and Extendible package of deep-learning-based CTR models along with lots of core components layers that can be used to build your own custom model easily.It is compatible with PyTorch.You can use any complex model with model.fit() and model.predict(). With the great success of deep learning, DNN-based techniques have been widely used in CTR estimation tasks. The data in the CTR estimation task usually includes high sparse,high cardinality categorical features and some dense numerical features. Low-order Extractor learns feature interaction through product between vectors. Factorization-Machine and it’s variants are widely used to learn the low-order feature interaction. High-order Extractor learns feature combination through complex neural network functions like MLP, Cross Net, etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    DeepChem

    DeepChem

    Democratizing Deep-Learning for Drug Discovery, Quantum Chemistry, etc

    DeepChem aims to provide a high-quality open-source toolchain that democratizes the use of deep learning in drug discovery, materials science, quantum chemistry, and biology. DeepChem currently supports Python 3.7 through 3.9 and requires these packages on any condition. DeepChem has a number of "soft" requirements. If you face some errors like ImportError: This class requires XXXX, you may need to install some packages. Deepchem provides support for TensorFlow, PyTorch, JAX and each requires an individual pip Installation. The DeepChem project maintains an extensive collection of tutorials. All tutorials are designed to be run on Google collab (or locally if you prefer). Tutorials are arranged in a suggested learning sequence that will take you from beginner to proficient at molecular machine learning and computational biology more broadly.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    DeepLearn

    DeepLearn

    Implementation of research papers on Deep Learning+ NLP+ CV in Python

    Welcome to DeepLearn. This repository contains an implementation of the following research papers on NLP, CV, ML, and deep learning. The required dependencies are mentioned in requirement.txt. I will also use dl-text modules for preparing the datasets. If you haven't use it, please do have a quick look at it. CV, transfer learning, representation learning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    DeepLearningProject

    DeepLearningProject

    An in-depth machine learning tutorial

    This tutorial tries to do what most Most Machine Learning tutorials available online do not. It is not a 30 minute tutorial that teaches you how to "Train your own neural network" or "Learn deep learning in under 30 minutes". It's a full pipeline which you would need to do if you actually work with machine learning - introducing you to all the parts, and all the implementation decisions and details that need to be made. The dataset is not one of the standard sets like MNIST or CIFAR, you will make you very own dataset. Then you will go through a couple conventional machine learning algorithms, before finally getting to deep learning! In the fall of 2016, I was a Teaching Fellow (Harvard's version of TA) for the graduate class on "Advanced Topics in Data Science (CS209/109)" at Harvard University. I was in charge of designing the class project given to the students, and this tutorial has been built on top of the project I designed for the class.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    Deepo

    Deepo

    Set up deep learning environment in a single command line

    Deepo is a series of Docker images that allows you to quickly set up your deep learning research environment, supports almost all commonly used deep learning frameworks, supports GPU acceleration (CUDA and cuDNN included), also works in CPU-only mode, and works on Linux (CPU version/GPU version), Windows (CPU version) and OS X (CPU version). Their Dockerfile generator that allows you to customize your own environment with Lego-like modules, and automatically resolves the dependencies for you. For users in China who may suffer from slow speeds when pulling the image from the public Docker registry, you can pull deepo images from the China registry mirror by specifying the full path, including the registry, in your docker pull command. This should work and enables Deepo to use the GPU from inside a docker container.
    Downloads: 0 This Week
    Last Update:
    See Project