• Turn traffic into pipeline and prospects into customers Icon
    Turn traffic into pipeline and prospects into customers

    For account executives and sales engineers looking for a solution to manage their insights and sales data

    Docket is an AI-powered sales enablement platform designed to unify go-to-market (GTM) data through its proprietary Sales Knowledge Lake™ and activate it with intelligent AI agents. The platform helps marketing teams increase pipeline generation by 15% by engaging website visitors in human-like conversations and qualifying leads. For sales teams, Docket improves seller efficiency by 33% by providing instant product knowledge, retrieving collateral, and creating personalized documents. Built for GTM teams, Docket integrates with over 100 tools across the revenue tech stack and offers enterprise-grade security with SOC 2 Type II, GDPR, and ISO 27001 compliance. Customers report improved win rates, shorter sales cycles, and dramatically reduced response times. Docket’s scalable, accurate, and fast AI agents deliver reliable answers with confidence scores, empowering teams to close deals faster.
    Learn More
  • Award-Winning Medical Office Software Designed for Your Specialty Icon
    Award-Winning Medical Office Software Designed for Your Specialty

    Succeed and scale your practice with cloud-based, data-backed, AI-powered healthcare software.

    RXNT is an ambulatory healthcare technology pioneer that empowers medical practices and healthcare organizations to succeed and scale through innovative, data-backed, AI-powered software.
    Learn More
  • 1
    Aida Lib

    Aida Lib

    Aida is a language agnostic library for text generation

    Aida is a language-agnostic library for text generation. When using Aida, first you compose a tree of operations on your text that includes conditions via branches and other control flow. Later, you fill the tree with data and render the text. A building block is a variable class: Var. Use it to represent a value that you want to control later. A variable can hold numbers (e.g. float, int) or strings. You can create branches and complex logic with Branch. The context, represented by the class Ctx, is useful to create rules that depends on what has been written before. Each object or literal that is passed to Aida is remembered by the context. Creating a reference expression is a common use-case, so we have a helper function called create_ref. You can compose operations on your text with some handy operators.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    BlogWizard

    BlogWizard

    Generate blog articles from video or audio

    BlogWizard is a demo/utility project built on top of Groq’s LLM infrastructure that converts video or audio content into well-structured blog posts, enabling creators to repurpose multimedia content into text — useful for SEO, accessibility, or reaching audiences that prefer reading. The tool uses transcription (e.g. via Whisper) to extract text from audio/video, then runs an LLM-based generation pipeline to transform that content into coherent, readable blog-format posts — with sections, formatting, and possibly metadata. This bridges the gap between modern multimedia content (podcasts, YouTube videos, interviews) and traditional written content, making cross-format publishing more efficient. For content creators, educators, or businesses producing audio/video content, blogwizard automates the tedious, manual process of transcription + blog writing, saving time while ensuring output quality.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    CPT

    CPT

    CPT: A Pre-Trained Unbalanced Transformer

    A Pre-Trained Unbalanced Transformer for Both Chinese Language Understanding and Generation. We replace the old BERT vocabulary with a larger one of size 51271 built from the training data, in which we 1) add missing 6800+ Chinese characters (most of them are traditional Chinese characters); 2) remove redundant tokens (e.g. Chinese character tokens with ## prefix); 3) add some English tokens to reduce OOV. Position Embeddings We extend the max_position_embeddings from 512 to 1024. We initialize the new version of models with the old version of checkpoints with vocabulary alignment. Token embeddings found in the old checkpoints are copied. And other newly added parameters are randomly initialized. We further train the new CPT & Chinese BART 50K steps with batch size 2048, max-seq-length 1024, peak learning rate 2e-5, and warmup ratio 0.1. Aiming to unify both NLU and NLG tasks, We propose a novel Chinese Pre-trained Un-balanced Transformer (CPT).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    CRSLab

    CRSLab

    CRSLab is an open-source toolkit

    CRSLab is an open-source toolkit for building Conversational Recommender System (CRS). It is developed based on Python and PyTorch. CRSLab has the following highlights. Comprehensive benchmark models and datasets: We have integrated commonly-used 6 datasets and 18 models, including graph neural network and pre-training models such as R-GCN, BERT and GPT-2. We have preprocessed these datasets to support these models, and release for downloading. Extensive and standard evaluation protocols: We support a series of widely-adopted evaluation protocols for testing and comparing different CRS. General and extensible structure: We design a general and extensible structure to unify various conversational recommendation datasets and models, in which we integrate various built-in interfaces and functions for quickly development. Easy to get started: We provide simple yet flexible configuration for new researchers to quickly start in our library. Human-machine interaction interfaces.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Failed Payment Recovery for Subscription Businesses Icon
    Failed Payment Recovery for Subscription Businesses

    For subscription companies searching for a failed payment recovery solution to grow revenue, and retain customers.

    FlexPay’s innovative platform uses multiple technologies to achieve the highest number of retained customers, resulting in reduced involuntary churn, longer life span after recovery, and higher revenue. Leading brands like LegalZoom, Hooked on Phonics, and ClinicSense trust FlexPay to recover failed payments, reduce churn, and increase customer lifetime value.
    Learn More
  • 5

    Companionem Linguae

    Ultra Large Language Model

    Companionem Linguae is an ultra large language model in early stages of development. Companionem Linguae is being reworked. I have uploaded a part of the new training data (Latein.json). Although English is the standard for international communication, I decided to train the model with Latin and German first, because these languages have many grammatical features the English language doesn't have, but they could be useful for translations into French, Portuguese, Spanish, or other languages that have similar grammatical features.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    Dickinson

    Dickinson

    Text generation language

    Dickinson is a text-generation language. You can try out the language on the web without installing anything. Binaries for some platforms are available on the releases page. There is an install script that will try to download the right release for your computer.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    GPT-2 FR

    GPT-2 FR

    GPT-2 French demo | Démo française de GPT-2

    OpenAI GPT-2 model trained on four different datasets in French. Books in French, French film scripts, reports of parliamentary debates, Tweet by Emmanuel Macron, allowing to generate text. Tensorflow and gpt-2-simple are required in order to fine-tune GPT-2. Create an environment then install the two packages pip install tensorflow==1.14 gpt-2-simple. A script and a notebook are available in the src folder to fine-tune GPT-2 on your own datasets. The output of each workout, i.e. the folder checkpoint/run1, is to be put ingpt2-model/model1 model2 model3 etc. You can run the script deploy_cloudrun.shto deploy all your different models (into gpt2-model) at once. However, you must have already initialized the gcloud CLI tool (Cloud SDK).
    Downloads: 0 This Week
    Last Update:
    See Project
  • 8
    GenBook
    # GenBook ## ✨ Introducción Este proyecto utiliza Large Language Models (LLMs) para generar novelas de fantasía completas con transparencia del proceso creativo. > **⚠️ AVISO**: Proyecto en fase temprana de desarrollo. Pueden existir bugs y liimitaciones.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    Insert Text

    Insert Text

    Extends the media pool with an effect for outputting text in the image

    This addon expands the media manager with the effect Bild. The effect can be used, for example, to display a copyright notice, creation date, image title, etc. on images. The values ​​set for the effect are considered "default" and can be changed individually for each image if necessary via the effect parameters in meta data. The text source of the effect can be selected here. inputfor the field Textausgabe or any meta field from the media pool. A text area can also be selected from the media pool, which opens up even more possibilities for this effect. The values ​​set in the effect are considered "default" for all images to which this media type is applied. If a "text area" from the meta data has been selected as the text source, an individual setting for the effect can be applied to each image.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Inventory and Order Management Software for Multichannel Sellers Icon
    Inventory and Order Management Software for Multichannel Sellers

    Avoid stockouts, overselling, and losing control as your business grows.

    We are the most powerful inventory and order management platform for Amazon, Walmart, and multichannel product sellers. Centralize orders, product information, and fulfillment operations to run more efficiently, sell more products, and stay compliant with marketplace requirements so you can grow profitably.
    Learn More
  • 10
    Market Reporter

    Market Reporter

    Automatic Generation of Brief Summaries of Time-Series Data

    Market Reporter automatically generates short comments that describe time series data of stock prices, FX rates, etc. This is an implementation of Murakami et al. This tool stores data to Amazon S3. Ask the manager to give you AmazonS3FullAccess and issue a credential file. For details, please read AWS Identity and Access Management. Install Docker and Docker Compose. Edit envs/docker-compose.yaml according to your environment. Then, launch containers by docker-compose. We recommend to use pipenv to make a Python environment for this project. Suppose you have a database named master on your local machine. Prediction submodule generates a single comment of a financial instrument at specified time by loading a trained model.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 11
    Minimal text diffusion

    Minimal text diffusion

    A minimal implementation of diffusion models for text generation

    A minimal implementation of diffusion models of text: learns a diffusion model of a given text corpus, allowing to generate text samples from the learned model. The main idea was to retain just enough code to allow training a simple diffusion model and generating samples, remove image-related terms, and make it easier to use. To train a model, run scripts/train.sh. By default, this will train a model on the simple corpus. However, you can change this to any text file using the --train_data argument. Note that you may have to increase the sequence length (--seq_len) if your corpus is longer than the simple corpus. The other default arguments are set to match the best setting I found for the simple corpus.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Pipeline for training Language Models

    Pipeline for training Language Models

    Pipeline for training Language Models using PyTorch.

    Pipeline for training Language Models using PyTorch. Inspired by Yandex Data School NLP Course (week 03: Language Modeling) Prepared text file with space-separated words on each line.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    ShortGPT Lite

    ShortGPT Lite

    Get short and concise answers from GPT 3/GPT 4

    Short GPT Lite is a simple tool for Windows/Linux based on OpenAI's GPT3/GPT4 large language model. The main focus is to get quick and concise answers from GPT. ShortGPT is now available on Android : https://play.google.com/store/apps/details?id=io.github.rupeshs.shortgpt_lite ShortGPT basic web version is now available try it for free: https://nolowiz.com/shortgpt-get-short-and-concise-answers-from-gpt-for-free/
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    TFKit

    TFKit

    Handling multiple nlp task in one pipeline

    TFKit is a tool kit mainly for language generation. It leverages the use of transformers on many tasks with different models in this all-in-one framework. All you need is a little change of config. You can use tfkit for model training and evaluation with tfkit-train and tfkit-eval. The key to combine different task together is to make different task with same data format. All data will be in csv format - tfkit will use csv for all task, normally it will have two columns, first columns is the input of models, the second column is the output of models. Plane text with no tokenization - there is no need to tokenize text before training, or do re-calculating for tokenization, tfkit will handle it for you. No header is needed.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Texar-PyTorch

    Texar-PyTorch

    Integrating the Best of TF into PyTorch, for Machine Learning

    Texar-PyTorch is a toolkit aiming to support a broad set of machine learning, especially natural language processing and text generation tasks. Texar provides a library of easy-to-use ML modules and functionalities for composing whatever models and algorithms. The tool is designed for both researchers and practitioners for fast prototyping and experimentation. Texar-PyTorch was originally developed and is actively contributed by Petuum and CMU in collaboration with other institutes. A mirror of this repository is maintained by Petuum Open Source. Texar-PyTorch integrates many of the best features of TensorFlow into PyTorch, delivering highly usable and customizable modules superior to PyTorch native ones. Texar-PyTorch (this repo) and Texar-TF have mostly the same interfaces. Both further combine the best design of TF and PyTorch. Data processing, model architectures, loss functions, training and inference algorithms, evaluation, etc.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Text Gen

    Text Gen

    Almost state of art text generation library

    Almost state of art text generation library. Text gen is a python library that allow you build a custom text generation model with ease. Something sweet built with Tensorflow and Pytorch(coming soon). Load your data, your data must be in a text format. Download the example data from the example folder. Tune your model to know the best optimizer, activation method to use.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17

    Text Image Creator Kovalenko

    Text Image creator mobile app

    Text Image Creator Kovalenko is the ultimate text generation mobile app, designed to streamline your content creation process and inspire your creativity. Whether you're a writer, marketer, student, or simply someone looking to spice up your social media posts, Text Image Creator Kovalenko is your go-to tool for generating high-quality and engaging text content on the fly. Content Marketing: Craft compelling blog posts, social media updates, and email newsletters effortlessly. Academic Writing: Generate research paper abstracts, summaries, or creative writing assignments. Business Communication: Create professional reports, proposals, and marketing materials with ease. Social Media: Craft engaging captions, tweets, and Facebook posts to captivate your followers. Creative Writing: Use as a writing prompt generator to kickstart your storytelling. Indispensable tool for generating text content that stands out. Unlock your
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    TextBox

    TextBox

    A text generation library with pre-trained language models github.com

    TextBox 2.0 is an up-to-date text generation library based on Python and PyTorch focusing on building a unified and standardized pipeline for applying pre-trained language models to text generation. From a task perspective, we consider 13 common text generation tasks such as translation, story generation, and style transfer, and their corresponding 83 widely-used datasets. From a model perspective, we incorporate 47 pre-trained language models/modules covering the categories of general, translation, Chinese, dialogue, controllable, distilled, prompting, and lightweight models (modules). From a training perspective, we support 4 pre-training objectives and 4 efficient and robust training strategies, such as distributed data parallel and efficient generation. Compared with the previous version of TextBox, this extension mainly focuses on building a unified, flexible, and standardized framework for better supporting PLM-based text generation models.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Travesty

    Travesty

    Parody text generator

    A parody text generator. This is taken from the article published in BYTE Magazine in 1984. Literary critic Hugh Kenner and computer scientist Joseph O'Rourke introduced their text scrambler "Travesty" in an issue of BYTE magazine 1984. See the Wikipedia page for more information. The code has been mostly preserved, I've just added a GUI to make it easier to play around with the options and included a copy of Alice in Wonderland. A Windows binary is available on the releases page. Parody generators are computer programs which generate text that is syntactically correct, but usually meaningless, often in the style of a technical paper or a particular writer. They are also called travesty generators and random text generators. Their purpose is often satirical, intending to show that there is little difference between the generated text and real examples.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    amrlib

    amrlib

    A python library that makes AMR parsing, generation and visualization

    A python library that makes AMR parsing, generation and visualization simple. amrlib is a python module designed to make processing for Abstract Meaning Representation (AMR) simple by providing the following functions. Sentence to Graph (StoG) parsing to create AMR graphs from English sentences. Graph to Sentence (GtoS) generation for turning AMR graphs into English sentences. A QT-based GUI to facilitate the conversion of sentences to graphs and back to sentences. Methods to plot AMR graphs in both the GUI and as library functions. Training and test code for both the StoG and GtoS models. A SpaCy extension that allows direct conversion of SpaCy Docs and Spans to AMR graphs. Sentence to Graph alignment routines FAA_Aligner (Fast_Align Algorithm), based on the ISI aligner code detailed in this paper. RBW_Aligner (Rule Based Word) for a simple, single token to single node alignment.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 21
    artikelschreiber

    artikelschreiber

    Frontend and Backend Code for ArtikelSchreiber.com and UNAIQUE.NET

    Frontend and Backend Code for ArtikelSchreiber.com and UNAIQUE.NET Text Generator deutsch - Dein KI Text Generator kostenlos mit Künstlicher Intelligenz The Software as a Service can be found here: SEO Optimizer: Ghost Writer - Hausarbeiten schreiben mit KI and KI Text Generator This product includes software developed by Sebastian Enger, M.Sc. Copyright (c) 2023, Sebastian Enger, M.Sc. All rights reserved. Frontend and Backend Source Code for Project: https://github.com/sebastianenger1981/ https://www.artikelschreiber.com/ https://www.artikelschreiben.com/ https://www.unaique.net/ https://www.artikelschreiber.com/opensource/ https://www.unaique.com/
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    commit-autosuggestions

    commit-autosuggestions

    A tool that AI automatically recommends commit messages

    This is implementation of CommitBERT: Commit Message Generation Using Pre-Trained Programming Language Model. CommitBERT is accepted in ACL workshop : NLP4Prog. Have you ever hesitated to write a commit message? Now get a commit message from Artificial Intelligence! CodeBERT: A Pre-Trained Model for Programming and Natural Languages introduces a pre-trained model in a combination of Program Language and Natural Language(PL-NL). It also introduces the problem of converting code into natural language (Code Documentation Generation). We can use CodeBERT to create a model that generates a commit message when code is added. However, most code changes are not made only by add of the code, and some parts of the code are deleted. We plan to slowly conquer languages that are not currently supported. To run this project, you need a flask-based inference server (GPU) and a client (commit module). If you don't have a GPU, don't worry, you can use it through Google Colab.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    hebrew-gpt_neo

    hebrew-gpt_neo

    Hebrew text generation models based on EleutherAI's gpt-neo

    Hebrew text generation models based on EleutherAI's gpt-neo. Each was trained on a TPUv3-8 which was made available to me via the TPU Research Cloud Program. The Open Super-large Crawled ALMAnaCH coRpus is a huge multilingual corpus obtained by language classification and filtering of the Common Crawl corpus using the goclassy architecture.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    node-markov-generator

    node-markov-generator

    Generates simple sentences based on given text corpus

    This simple generator emits short sentences based on the given text corpus using a Markov chain. To put it simply, it works kinda like word suggestions that you have while typing messages in your smartphone. It analyzes which word is followed by which in the given corpus and how often. And then, for any given word it tries to predict what the next one might be. Here you create an instance of TextGenerator passing an array of strings to it - it represents your text corpus which will be used to "train" the generator. The more strings/sentences you pass, the more diverse results you get, so you'd better pass like hundreds of them, or even more! If you have your texts in an external file, you can pass the path to it as an argument for TextGenerator's constructor.
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB