Domain
Machine Learning & AI
Skill Profile
HuggingFace Transformers, BERT, GPT, fine-tuning, tokenization
Roles
3
where this skill appears
Levels
5
structured growth path
Mandatory requirements
11
the other 4 optional
Machine Learning & AI
Natural Language Processing
3/17/2026
Choose your current level and compare expectations. The items below show what to cover to advance to the next level.
The table shows how skill depth grows from Junior to Principal. Click a row to see details.
| Role | Required | Description |
|---|---|---|
| Data Scientist | Understands core Transformer architecture (attention mechanisms, positional encoding) and basic NLP pipelines. Uses pre-trained models from Hugging Face for text classification and NER tasks. Follows team guidelines for data preprocessing and tokenization workflows. | |
| LLM Engineer | Understands core Transformer architecture and attention mechanisms. Works with pre-trained LLMs using Hugging Face Transformers library for inference and basic prompt engineering. Follows established patterns for model loading, tokenization, and API integration. | |
| NLP Engineer | Required | Knows transformer architecture basics for NLP: self-attention, positional encoding, BERT/GPT. Uses Hugging Face transformers for inference: text classification, NER, summarization. |
| Role | Required | Description |
|---|---|---|
| Data Scientist | Fine-tunes Transformer models for domain-specific NLP tasks including sentiment analysis, summarization, and question answering. Evaluates trade-offs between model size, inference speed, and accuracy. Implements custom training loops with proper evaluation metrics and cross-validation. | |
| LLM Engineer | Applies fine-tuning techniques (LoRA, QLoRA, PEFT) to adapt LLMs for specific tasks. Implements RAG pipelines combining retrieval with generation. Understands trade-offs between model quantization, context window size, and output quality for production use cases. | |
| NLP Engineer | Required | Independently fine-tunes transformer models for NLP: BERT, RoBERTa, T5 for domain-specific tasks. Configures tokenizers, training arguments, evaluation metrics via Hugging Face Trainer. |
| Role | Required | Description |
|---|---|---|
| Data Scientist | Required | Designs end-to-end NLP systems with Transformer models for production: custom architectures, distributed training, model distillation and quantization. Optimizes inference latency and throughput for real-time applications. Mentors team on advanced techniques like multi-task learning and domain adaptation. |
| LLM Engineer | Required | Architects production LLM systems with optimized serving (vLLM, TGI), model parallelism, and efficient batching strategies. Designs evaluation frameworks for model quality, safety, and bias detection. Implements advanced techniques: RLHF, constitutional AI, and chain-of-thought optimization. |
| NLP Engineer | Required | Designs advanced NLP solutions with transformers: adapter-based fine-tuning, model merging, efficient inference. Optimizes through quantization, pruning, Flash Attention for production. |
| Role | Required | Description |
|---|---|---|
| Data Scientist | Required | Defines NLP and Transformer strategy at team level: selects model architectures, establishes training infrastructure, and sets quality benchmarks. Conducts architectural reviews of ML pipelines. Drives adoption of best practices for experiment tracking, model versioning, and reproducibility. |
| LLM Engineer | Required | Defines LLM platform strategy: model selection criteria, fine-tuning pipelines, and serving infrastructure standards. Establishes evaluation benchmarks, safety guardrails, and cost optimization practices. Reviews architectural decisions for RAG systems, agent frameworks, and multi-model orchestration. |
| NLP Engineer | Required | Defines transformer strategy for the NLP team. Establishes model selection guidelines, fine-tuning approaches, and optimization techniques. Evaluates new architectures and their applicability. |
| Role | Required | Description |
|---|---|---|
| Data Scientist | Required | Defines organizational NLP and Transformer strategy: foundation model investments, build-vs-buy decisions for language AI, and cross-team knowledge sharing. Establishes enterprise standards for model governance, responsible AI, and data pipeline architecture. Mentors leads on scaling ML organizations. |
| LLM Engineer | Required | Shapes organizational LLM strategy: evaluates foundation model providers, defines enterprise AI governance policies, and architects cross-team LLM infrastructure. Establishes standards for model risk management, cost forecasting, and responsible AI deployment. Mentors leads on building scalable AI platforms. |
| NLP Engineer | Required | Shapes enterprise transformer strategy for the NLP platform. Defines model hub standards, shared fine-tuned models, and research-to-production workflow at organizational level. |