Showing 2 open source projects for "npp-compare"

View related business solutions
  • Loan management software that makes it easy. Icon
    Loan management software that makes it easy.

    Ideal for lending professionals who are looking for a feature rich loan management system

    Bryt Software is ideal for lending professionals who are looking for a feature rich loan management system that is intuitive and easy to use. We are 100% cloud-based, software as a service. We believe in providing our customers with fair and honest pricing. Our monthly fees are based on your number of users and we have a minimal implementation charge.
    Learn More
  • AestheticsPro Medical Spa Software Icon
    AestheticsPro Medical Spa Software

    Our new software release will dramatically improve your medspa business performance while enhancing the customer experience

    AestheticsPro is the most complete Aesthetics Software on the market today. HIPAA Cloud Compliant with electronic charting, integrated POS, targeted marketing and results driven reporting; AestheticsPro delivers the tools you need to manage your medical spa business. It is our mission To Provide an All-in-One Cutting Edge Software to the Aesthetics Industry.
    Learn More
  • 1
    bufferline.nvim

    bufferline.nvim

    A snazzy bufferline for Neovim

    A snazzy buffer line (with tab page integration) for Neovim built using Lua. This plugin shamelessly attempts to emulate the aesthetics of GUI text editors/Doom Emacs. It is advised that you specify either the latest tag or a specific tag and bump them manually if you'd prefer to inspect changes before updating.
    Downloads: 5 This Week
    Last Update:
    See Project
  • 2
    char-rnn

    char-rnn

    Multi-layer Recurrent Neural Networks (LSTM, GRU, RNN)

    char-rnn is a classic codebase for training multi-layer recurrent neural networks on raw text to build character-level language models that learn to predict the next character in a sequence. It supports common recurrent architectures including vanilla RNNs as well as LSTM and GRU variants, letting users compare behavior and output quality across model types. It is straightforward: you provide a single text file, train the model to minimize next-character prediction loss, then sample from the trained network to generate new text one character at a time in the style of the dataset. The project is designed for experimentation, offering tunable settings for depth, hidden size, dropout, sequence length, and sampling temperature to control creativity and coherence. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB