Yardi Voyager
Yardi Voyager is a web-based, fully integrated end-to-end platform with mobile access for larger portfolios to manage operations, execute leasing, run analytics, and provide innovative resident, tenant, and investor services. With a solution and best-of-breed product suite designed for every real estate market including commercial (office, retail, industrial), multifamily, affordable, senior, PHA and military housing, Voyager helps you meet all your property management and accounting needs using a single database to run your entire business. Voyager automates workflows and provides system-wide transparency that enables you to work more productively and collaboratively than ever before. Using any browser and mobile device, Voyager gives you instant access to your data. And as a SaaS platform, Voyager frees you from managing your software — so you can focus on your business.
Learn more
Mistral AI
Mistral AI is a pioneering artificial intelligence startup specializing in open-source generative AI. The company offers a range of customizable, enterprise-grade AI solutions deployable across various platforms, including on-premises, cloud, edge, and devices. Flagship products include "Le Chat," a multilingual AI assistant designed to enhance productivity in both personal and professional contexts, and "La Plateforme," a developer platform that enables the creation and deployment of AI-powered applications. Committed to transparency and innovation, Mistral AI positions itself as a leading independent AI lab, contributing significantly to open-source AI and policy development.
Learn more
Gemini Embedding
Gemini Embedding’s first text model (gemini-embedding-001) is now generally available via the Gemini API and Vertex AI, having held a top spot on the Massive Text Embedding Benchmark Multilingual leaderboard since its experimental launch in March, thanks to superior performance across retrieval, classification, and other embedding tasks compared to both legacy Google and external proprietary models. Exceptionally versatile, it supports over 100 languages with a 2,048‑token input limit and employs the Matryoshka Representation Learning (MRL) technique to let developers choose output dimensions of 3072, 153,6, or 768 for optimal quality, performance, and storage efficiency. Developers can access it through the existing embed_content endpoint in the Gemini API, and while legacy experimental versions will be deprecated later in 2025, migration requires no re‑embedding of existing content.
Learn more
Pinecone
The AI Knowledge Platform.
The Pinecone Database, Inference, and Assistant make building high-performance vector search apps easy. Developer-friendly, fully managed, and easily scalable without infrastructure hassles.
Once you have vector embeddings, manage and search through them in Pinecone to power semantic search, recommenders, and other applications that rely on relevant information retrieval.
Ultra-low query latency, even with billions of items. Give users a great experience. Live index updates when you add, edit, or delete data. Your data is ready right away. Combine vector search with metadata filters for more relevant and faster results.
Launch, use, and scale your vector search service with our easy API, without worrying about infrastructure or algorithms. We'll keep it running smoothly and securely.
Learn more