What you can do with Pioneer
NER fine-tuning
Train GLiNER encoder models on your own entity types. Extract named entities, classify text, and produce structured JSON output from unstructured text.
LLM fine-tuning
Fine-tune decoder models — Qwen, Llama, DeepSeek, and others — on your domain data using LoRA. Download the weights when training is complete.
Inference
Run predictions against any base model or your own fine-tuned model via
POST /inference. Pioneer also exposes OpenAI- and Anthropic-compatible endpoints for drop-in compatibility.Synthetic data generation
Generate labeled training examples for NER and classification tasks without manual annotation. Describe your domain and Pioneer produces ready-to-use datasets.
Supported model families
Pioneer supports two classes of models: encoder models for structured extraction tasks, and decoder models for generative tasks. Encoder models- GLiNER — A small, efficient model purpose-built for named entity recognition, text classification, and structured JSON extraction. GLiNER is the recommended starting point for agent text processing, document parsing, and routing workflows.
- Qwen — Strong at coding, multilingual tasks, and complex multi-step reasoning. Ideal for global products.
- Llama — Meta’s open-source model family. Well-suited for RAG, summarization, and general-purpose chat.
- DeepSeek — Capable at code generation, structured reasoning, and agentic planning tasks.
GET /base-models. You can filter by task type or inference support.
Pioneer supports Adaptive Inference on the Research plan — a continuous improvement loop that automatically evaluates live traffic, generates training data, and promotes improved checkpoints. See Adaptive Inference to learn more.
How it fits into your workflow
Pioneer follows a straightforward lifecycle: upload or generate data → start a training job → evaluate the result → run inference with your trained model.Choose a base model
Pick a GLiNER encoder for extraction and classification, or a decoder LLM for generative tasks. Use
GET /base-models to browse what’s available.Prepare your data
Upload a labeled dataset or use Pioneer’s synthetic data generation to create training examples from a domain description and label list.
Fine-tune
Start a training job with
POST /felix/training-jobs. Pioneer runs LoRA fine-tuning and returns F1, precision, and recall metrics on completion.Next steps
Quickstart
Make your first API call in under five minutes.
Authentication
Generate your API key and authenticate requests.

