Ango Hub
Ango Hub is a quality-focused, enterprise-ready data annotation platform for AI teams, available on cloud and on-premise. It supports computer vision, medical imaging, NLP, audio, video, and 3D point cloud annotation, powering use cases from autonomous driving and robotics to healthcare AI.
Built for AI fine-tuning, RLHF, LLM evaluation, and human-in-the-loop workflows, Ango Hub boosts throughput with automation, model-assisted pre-labeling, and customizable QA while maintaining accuracy. Features include centralized instructions, review pipelines, issue tracking, and consensus across up to 30 annotators. With nearly twenty labeling tools—such as rotated bounding boxes, label relations, nested conditional questions, and table-based labeling—it supports both simple and complex projects. It also enables annotation pipelines for chain-of-thought reasoning and next-gen LLM training and enterprise-grade security with HIPAA compliance, SOC 2 certification, and role-based access controls.
Learn more
Vertex AI
Build, deploy, and scale machine learning (ML) models faster, with fully managed ML tools for any use case.
Through Vertex AI Workbench, Vertex AI is natively integrated with BigQuery, Dataproc, and Spark. You can use BigQuery ML to create and execute machine learning models in BigQuery using standard SQL queries on existing business intelligence tools and spreadsheets, or you can export datasets from BigQuery directly into Vertex AI Workbench and run your models from there. Use Vertex Data Labeling to generate highly accurate labels for your data collection.
Vertex AI Agent Builder enables developers to create and deploy enterprise-grade generative AI applications. It offers both no-code and code-first approaches, allowing users to build AI agents using natural language instructions or by leveraging frameworks like LangChain and LlamaIndex.
Learn more
Nemotron 3 Ultra
Nemotron 3 Nano is a compact, open large language model in NVIDIA’s Nemotron 3 family, designed for efficient agentic reasoning, conversational AI, and coding tasks. It uses a hybrid Mixture-of-Experts Mamba-Transformer architecture that activates only a small subset of parameters per token, enabling low-latency inference while maintaining strong accuracy and reasoning performance. It has approximately 31.6 billion total parameters with around 3.2 billion active (3.6 billion including embeddings), allowing it to achieve higher accuracy than previous Nemotron 2 Nano while using less computation per forward pass. Nemotron 3 Nano supports long-context processing of up to one million tokens, enabling it to handle large documents, multi-step workflows, and extended reasoning chains in a single pass. It is designed for high-throughput, real-time execution, excelling in multi-turn conversations, tool calling, and agent-based workflows where tasks require planning, reasoning, and more.
Learn more