Domain
Machine Learning & AI
Skill Profile
Pinecone, Weaviate, Qdrant, Milvus, pgvector, embedding indexing and search
Roles
3
where this skill appears
Levels
5
structured growth path
Mandatory requirements
13
the other 2 optional
Machine Learning & AI
LLM & Generative AI
3/17/2026
Choose your current level and compare expectations. The items below show what to cover to advance to the next level.
The table shows how skill depth grows from Junior to Principal. Click a row to see details.
| Role | Required | Description |
|---|---|---|
| AI Product Engineer | Understands the fundamentals of Vector Databases. Applies basic practices in daily work. Follows recommendations from the team and documentation. | |
| LLM Engineer | Required | Knows vector database basics: what is vector index, ANN search, distance metrics. Works with ChromaDB or Faiss for storing and searching embeddings in simple RAG applications. |
| NLP Engineer | Required | Knows vector database basics: embeddings, similarity search, ANN algorithms. Uses Pinecone/Weaviate/Qdrant for storing text embeddings and semantic search. |
| Role | Required | Description |
|---|---|---|
| AI Product Engineer | Independently applies Vector Databases in practice. Understands trade-offs of different approaches. Solves typical tasks independently. | |
| LLM Engineer | Required | Independently administers vector databases in production: Pinecone, Weaviate, Qdrant. Configures indexes (HNSW, IVF), optimizes recall vs latency, manages collections and metadata. |
| NLP Engineer | Required | Independently designs vector search for NLP: embedding model selection, index configuration, metadata filtering. Optimizes recall and latency for production semantic search. |
| Role | Required | Description |
|---|---|---|
| AI Product Engineer | Required | Has deep expertise in Vector Databases. Designs solutions for production systems. Optimizes and scales. Mentors the team. |
| LLM Engineer | Required | Designs scalable vector DB infrastructure: sharding, replication, hybrid search. Optimizes index parameters for trade-offs between recall, latency, and memory with millions of vectors. |
| NLP Engineer | Required | Designs production vector search infrastructure for NLP: multi-tenant architecture, embedding model selection, index sharding. Optimizes for scale and cost-effectiveness. |
| Role | Required | Description |
|---|---|---|
| AI Product Engineer | Required | Defines Vector Databases strategy at the team/product level. Establishes standards and best practices. Conducts reviews. |
| LLM Engineer | Required | Defines vector database strategy for the LLM platform. Establishes guidelines for vector DB selection, schema design, indexing strategy, monitoring. Coordinates migration and upgrades. |
| NLP Engineer | Required | Defines vector search strategy for the NLP team. Establishes embedding pipeline standards, index management, and evaluation metrics for semantic search systems. |
| Role | Required | Description |
|---|---|---|
| AI Product Engineer | Required | Defines Vector Databases strategy at the organizational level. Establishes enterprise approaches. Mentors leads and architects. |
| LLM Engineer | Required | Shapes enterprise vector database strategy. Defines approaches to centralized vector infrastructure, multi-tenant architecture, and cost optimization for billions of vectors at organizational scale. |
| NLP Engineer | Required | Shapes enterprise vector search strategy for the NLP platform. Defines shared embedding infrastructure architecture and semantic retrieval standards at organizational level. |