# Glossary
Source: https://docs.pinecone.io/guides/get-started/glossary
This page defines concepts in Pinecone and how they relate to each other.
## Organization
A organization is a group of one or more [projects](#project) that use the same billing. Organizations allow one or more [users](#user) to control billing and permissions for all of the projects belonging to the organization.
For more information, see [Understanding organizations](/guides/organizations/understanding-organizations).
## Project
A project belongs to an [organization](#organization) and contains one or more [indexes](#index). Each project belongs to exactly one organization, but only [users](#user) who belong to the project can access the indexes in that project. [API keys](#api-key) and [Assistants](#assistant) are project-specific.
For more information, see [Understanding projects](/guides/projects/understanding-projects).
## Index
There are two types of [serverless indexes](/guides/index-data/indexing-overview), dense and sparse.
For more information, see [Use namespaces](/guides/index-data/indexing-overview#namespaces).
## Record
A record is a basic unit of data and consists of a [record ID](#record-id), a [dense vector](#dense-vector) or a [sparse vector](#sparse-vector) (depending on the type of index), and optional [metadata](#metadata).
For more information, see [Upsert data](/guides/index-data/upsert-data).
### Record ID
A record ID is a record's unique ID. [Use ID prefixes](/guides/manage-data/manage-document-chunks#use-id-prefixes) to segment your data beyond namespaces.
### Dense vector
A dense vector, also referred to as a vector embedding or simply a vector, is a series of numbers that represent the meaning and relationships of data. Each number in a dense vector corresponds to a point in a multidimensional space. Vectors that are closer together in that space are semantically similar.
Dense vectors are stored in [dense indexes](#dense-index).
You use a dense embedding model to convert data to dense vectors. The embedding model can be external to Pinecone or [hosted on Pinecone infrastructure](/guides/index-data/create-an-index#embedding-models) and integrated with an index.
For more information about dense vectors, see [What are vector embeddings?](https://www.pinecone.io/learn/vector-embeddings/).
### Sparse vector
A sparse vector, also referred to as a sparse vector embedding, has a large number of dimensions, but only a small proportion of those values are non-zero. Sparse vectors are often used to represent documents or queries in a way that captures keyword information. Each dimension in a sparse vector typically represents a word from a dictionary, and the non-zero values represent the importance of these words in the document.
Sparse vectors are store in [sparse indexes](#sparse-index).
You use a sparse embedding model to convert data to sparse vectors. The embedding model can be external to Pinecone or [hosted on Pinecone infrastructure](/guides/index-data/create-an-index#embedding-models) and integrated with an index.
For more information about sparse vectors, see [Sparse retrieval](https://www.pinecone.io/learn/sparse-retrieval/).
### Metadata
Metadata is additional information that can be attached to vector embeddings to provide more context and enable additional [filtering capabilities](/guides/index-data/indexing-overview#metadata). For example, the original text of the embeddings can be stored in the metadata.
## Other concepts
Although not represented in the diagram above, Pinecone also contains the following concepts:
* [API key](#api-key)
* [User](#user)
* [Backup or collection](#backup-or-collection)
* [Pinecone Inference](#pinecone-inference)
### API key
An API key is a unique token that [authenticates](/reference/api/authentication) and authorizes access to the [Pinecone APIs](/reference/api/introduction). API keys are project-specific.
### User
A user is a member of organizations and projects. Users are assigned specific roles at the organization and project levels that determine the user's permissions in the [Pinecone console](https://app.pinecone.io).
For more information, see [Manage organization members](/guides/organizations/manage-organization-members) and [Manage project members](/guides/projects/manage-project-members).
### Backup or collection
A backup is a static copy of a serverless index.
Backups only consume storage. They are non-queryable representations of a set of records. You can create a backup from an index, and you can create a new index from that backup. The new index configuration can differ from the original source index: for example, it can have a different name. However, it must have the same number of dimensions and similarity metric as the source index.
For more information, see [Understanding backups](/guides/manage-data/backups-overview).
### Pinecone Inference
Pinecone Inference is an API service that provides access to [embedding models](/guides/index-data/create-an-index#embedding-models) and [reranking models](/guides/search/rerank-results#reranking-models) hosted on Pinecone's infrastructure.
## Learn more
* [Vector database](https://www.pinecone.io/learn/vector-database/)
* [Pinecone APIs](/reference/api/introduction)
* [Approximate nearest neighbor (ANN) algorithms](https://www.pinecone.io/learn/a-developers-guide-to-ann-algorithms/)
* [Retrieval augmented generation (RAG)](https://www.pinecone.io/learn/retrieval-augmented-generation/)
* [Image search](https://www.pinecone.io/learn/series/image-search/)
* [Tokenization](https://www.pinecone.io/learn/tokenization/)
# Pinecone Database
Source: https://docs.pinecone.io/guides/get-started/overview
Pinecone is the leading vector database for building accurate and performant AI applications at scale in production.
{text}
; }; export const ExampleCard = ({title, text, link, children, arrow, vectors, namespaces}) => { return{text}
{children &&{text}
; }; export const ExampleCard = ({title, text, link, children, arrow, vectors, namespaces}) => { return{text}
{children &&Seamlessly integrate, transform, and load data into Pinecone from hundreds of systems, including databases, data warehouses, and SaasS products.
Integrate your enterprise data into Amazon Bedrock using Pinecone to build highly performant GenAI applications.
Integrate machine learning models seamlessly with a fully-managed service that enables easy deployment and scalability.
Focus on building applications powered by LLMs without the need to worry about the underlying infrastructure.
Integrate results from web scrapers or crawlers into a vector database to support RAG or semantic search over web content.
Process complex, unstructured documents with a purpose-built ETL system for RAG and GenAI applications.
Access Pinecone through our Amazon Web Services marketplace listing.
Access Pinecone through our Microsoft Azure marketplace listing.
Connect a Box account to a Pinecone vector database.
Vector embedding, RAG, and semantic search at scale.
Leverage cutting-edge natural language processing tools for enhanced text understanding and generation in your applications.
Connect and process all of your data in real time with a cloud-native and complete data streaming platform.
Create end-to-end data flows that connect data sources to Pinecone.
Combine the power of a unified analytics platform with Pinecone for scalable data processing and AI insights.
Monitor and secure your applications by integrating with a cloud-scale monitoring service that provides real-time analytics.
Source, transform, and enrich data in a continuous, composable and customizable manner.
Source data from hundreds systems and push data to Pinecone, for an always up-to-date view.
Build customized LLM apps with an open source, low-code tool for developing orchestration flow & AI agents.
Build, deploy, and manage complex workflows with a low-code platform for AI-assisted ML and LLM transformations.
Build and operationalize data and AI-driven solutions at scale.
Access Pinecone through our Google Cloud Marketplace listing.
Build AI powered applications and agents.
Get personalized recommendations that enable you to retrieve relevant data and collaborate effectively with Copilot.
Implement an end-to-end search pipeline for efficient retrieval and question answering over large datasets.
Clearly visualize your execution traces and spans.
Deploy state-of-the-art machine learning models on scalable infrastructure, streamlining the path from prototype to production.
Streamline AI development with a low-code full-stack infrastructure tool for data, model, and pipeline orchestration.
Leverage powerful AI models to generate high-quality text embeddings, fine-tuned to both domain- and language-specific use cases.
Combine language models with chain-of-thought reasoning for advanced problem solving and decision support.
Access rich and high cardinal tracing for Pinecone API calls, ingestible into your observability tool of choice.
Leverage Llama for indexing and retrieving information at scale, improving data access and analysis.
Easily create and maintain data pipelines, build custom connectors for any source, and enjoy AI and high-code options to suit any need.
Implement monitoring and integrate your Pinecone application with New Relic for performance analysis and insights.
Ingest data from 500+ connectors with Nexla's low-code/no-code AI integration platform.
Nuclia RAG-as-a-Service automatically indexes files and documents from both internal and external sources.
Harness value from the latest AI innovations by delievering efficient, reliable, and customizable AI systems for your apps.
Access powerful AI models like GPT for innovative applications and services, enhancing user experiences with AI capabilities.
Manage your Pinecone collections and indexes using any language of Pulumi Infrastructure as Code.
Connect existing data sources to Pinecone with a Kafka-compatible streaming data platform built for data-intensive applications.
Run Pinecone with Snowpark Container Services, designed to deploy, manage, and scale containerized applications within the Snowflake ecosystem.
A scalable, resilient, and secure messaging and event streaming platform.
Manage your infrastructure using configuration files for a consistent workflow.
Produce traces and metrics that can be viewed in any OpenTelemetry-based platform.
Gain insights into your machine learning models' decisions, improving interpretability and trustworthiness.
Create high-quality multimodal embeddings that capture the rich context and interactions between different modalities in videos.
Load data into Pinecone with a single click.
Use Pinecone as the long-term memory for your Vercel AI projects, and easily scale to support billions of data points.
Cutting-edge embedding models and rerankers for semantic search and RAG.
{company.toUpperCase()}
{description}
Task | {task} |
Modality | {modality} |
Max Input Tokens | {maxInputTokens} |
Price | {price} |
Build end-to-end faster with models hosted by Pinecone.
Request a model{text}
; }; export const ExampleCard = ({title, text, link, children, arrow, vectors, namespaces}) => { return{text}
{children &&