Tiger Cloud: Performance, Scale, Enterprise, Free

Self-hosted products

MST

You can build and deploy AI Assistants that understand, analyze, and act on your organizational data using Tiger Data. Whether you're building semantic search applications, recommendation systems, or intelligent agents that answer complex business questions, Tiger Data provides the tools and infrastructure you need.

Tiger Data's AI ecosystem combines Postgres with advanced vector capabilities, intelligent agents, and seamless integrations. Your AI Assistants can:

  • Access organizational knowledge from Slack, GitHub, Linear, and other data sources
  • Understand context using advanced vector search and embeddings across large datasets
  • Execute tasks, generate reports, and interact with your Tiger Cloud services through natural language
  • Scale reliably with enterprise-grade performance for concurrent conversations

Tiger Eon automatically integrates Tiger Agents for Work with your organizational data. You can:

  • Get instant access to company knowledge from Slack, GitHub, and Linear
  • Process data in real-time as conversations and updates happen
  • Store data efficiently with time-series partitioning and compression
  • Deploy quickly with Docker and an interactive setup wizard

Use Eon when you want to unlock knowledge from your communication and development tools.

Tiger Agents for Work provides enterprise-grade Slack-native AI agents. You get:

  • Durable event handling with Postgres-backed processing
  • Horizontal scalability across multiple Tiger Agent instances
  • Flexibility to choose AI models and customize prompts
  • Integration with specialized data sources through MCP servers
  • Complete observability and monitoring with Logfire

Use Tiger Agents for Work when you need reliable, customizable AI agents for high-volume conversations.

The Tiger MCP integrates directly with popular AI Assistants. You can:

  • Work with Claude Code, Cursor, VS Code, and other editors
  • Manage services and optimize queries through natural language
  • Access comprehensive Tiger Data documentation during development
  • Use secure authentication and access control

Use the Tiger MCP when you want to manage Tiger Data resources from your AI Assistant.

Pgvector is a popular open source extension for vector storage and similarity search in Postgres and pgvectorscale adds advanced indexing capabilities to pgvector. pgai on Tiger Cloud offers both extensions so you can use all the capabilities already available in pgvector (like HNSW and ivfflat indexes) and also make use of the StreamingDiskANN index in pgvectorscale to speed up vector search.

This makes it easy to migrate your existing pgvector deployment and take advantage of the additional performance features in pgvectorscale. You also have the flexibility to create different index types suited to your needs. See the vector search indexing section for more information.

Embeddings offer a way to represent the semantic essence of data and to allow comparing data according to how closely related it is in terms of meaning. In the database context, this is extremely powerful: think of this as full-text search on steroids. Vector databases allow storing embeddings associated with data and then searching for embeddings that are similar to a given query.

  • Semantic search: transcend the limitations of traditional keyword-driven search methods by creating systems that understand the intent and contextual meaning of a query, thereby returning more relevant results. Semantic search doesn't just seek exact word matches; it grasps the deeper intent behind a user's query. The result? Even if search terms differ in phrasing, relevant results are surfaced. Taking advantage of hybrid search, which marries lexical and semantic search methodologies, offers users a search experience that's both rich and accurate. It's not just about finding direct matches anymore; it's about tapping into contextually and conceptually similar content to meet user needs.

  • Recommendation systems: imagine a user who has shown interest in several articles on a singular topic. With embeddings, the recommendation engine can delve deep into the semantic essence of those articles, surfacing other database items that resonate with the same theme. Recommendations, thus, move beyond just the superficial layers like tags or categories and dive into the very heart of the content.

  • Retrieval augmented generation (RAG): supercharge generative AI by providing additional context to Large Language Models (LLMs) like OpenAI's GPT-4, Anthropic's Claude 2, and open source modes like Llama 2. When a user poses a query, relevant database content is fetched and used to supplement the query as additional information for the LLM. This helps reduce LLM hallucinations, as it ensures the model's output is more grounded in specific and relevant information, even if it wasn't part of the model's original training data.

  • Clustering: embeddings also offer a robust solution for clustering data. Transforming data into these vectorized forms allows for nuanced comparisons between data points in a high-dimensional space. Through algorithms like K-means or hierarchical clustering, data can be categorized into semantic categories, offering insights that surface-level attributes might miss. This surfaces inherent data patterns, enriching both exploration and decision-making processes.

On a high level, embeddings help a database to look for data that is similar to a given piece of information (similarity search). This process includes a few steps:

  • First, embeddings are created for data and inserted into the database. This can take place either in an application or in the database itself.
  • Second, when a user has a search query (for example, a question in chat), that query is then transformed into an embedding.
  • Third, the database takes the query embedding and searches for the closest matching (most similar) embeddings it has stored.

Under the hood, embeddings are represented as a vector (a list of numbers) that capture the essence of the data. To determine the similarity of two pieces of data, the database uses mathematical operations on vectors to get a distance measure (commonly Euclidean or cosine distance). During a search, the database should return those stored items where the distance between the query embedding and the stored embedding is as small as possible, suggesting the items are most similar.

pgai on Tiger Cloud works with the most popular embedding models that have output vectors of 2,000 dimensions or less.:

And here are some popular choices for image embeddings:

Keywords

Found an issue on this page?Report an issue or Edit this page in GitHub.