Skip to main content

Getting Started

Protean AI is an end-to-end enterprise AI platform designed to help teams fine-tune, deploy, and operate domain-specific models in a sovereign, cost-efficient, and secure way. This section introduces the core concepts and the typical workflow, enabling a structured path from raw data to deployed AI agents.

What Protean AI Provides

Protean AI brings together all essential building blocks required for enterprise AI development:

  • Node Registration & Management – Secure registration and management of compute nodes (CPU/GPU) that are authorized to run training jobs and serve inference workloads.
  • Model Registry – Centralized management of base models, quantized variants, versions, and metadata.
  • Fine-Tuning Engine – Opinionated and efficient training pipelines with strict alignment between dataset format, objective, loss function, and evaluation strategy.
  • Adapter Registry – A dedicated registry for managing adapters (LoRA/QLoRA) independent of base models. Supports uploading adapters in standard HuggingFace format or publishing fine-tuned checkpoints as reusable adapters. Enables versioning, metadata tracking, and flexible composition of multiple adapters with base models at deployment time.
  • Model Runtime – Lightweight, scalable inference with support for adapters, quantization, and streaming responses.
  • Knowledge set Management – Native support for preparing, versioning, and validating Retrieval-Augmented Generation datasets, including document ingestion, chunking strategies, and embedding generation.
  • RAG APIs – Configurable retrieval pipelines that combine vector search, ranking, and prompt assembly in a reproducible way.
  • MCP Server – Support for both internal and external Model Context Protocol (MCP) servers. External MCP servers can be registered with managed authorization and access control, while Protean AI provides a built-in SQL MCP server for secure, structured access to relational data.
  • Agent Framework – Tools for building, deploying, and operating agents that combine models, RAG pipelines, MCP server integration, tool execution, and memory management.
  • Governance & Control – Built-in support for data sovereignty, security, auditability, and operational safety.

Workflow

A standard Protean AI workflow follows a predictable and reproducible sequence:

  1. Register a Node
    Register a compute node with Protean AI before any deployment or training can occur.
    Node registration establishes trust, advertises available hardware (CPU, GPU, memory), and defines where models are allowed to run.

  2. Select a Base Model
    Choose a foundational or instruction-tuned model that matches the task, domain complexity, and registered hardware constraints.

  3. Prepare Training Datasets
    Structure supervised training data using supported dataset formats aligned with the training objective (chat, classification, embedding, reranking, etc.).

  4. Configure and Run Fine-Tuning
    Define objective, evaluation metrics, and efficiency settings such as LoRA/QLoRA, gradient accumulation, etc.
    Execute training with frequent evaluation and reproducible results.

  5. Publish the Adapter
    Publish the fine-tuned checkpoint as an adapter for reuse across multiple base models deployments. Upload an existing adapter in HuggingFace format into the Adapter Registry. Adapters are managed independently of base models, enabling flexible composition and reuse. Adapter Registry manages versioning, lineage, and metadata of adapters.

  6. Prepare Knowledge Sets Ingest enterprise documents and structure them into Knowledge Sets. Apply deterministic chunking strategies, embedding models, and metadata enrichment to ensure predictable retrieval behavior.

  7. Configure RAG
    Define retrieval, ranking, and prompt-assembly behavior using the RAG APIs.
    Validate retrieval quality independent of generation.

  8. Deploy Models for Inference
    Deploy models and adapters to registered nodes using the runtime, with optional quantization to minimize memory footprint and infrastructure cost.

  9. Build and Deploy Agents
    Compose agents that combine deployed models, RAG pipelines, and tools.
    Deploy agents as long-running services that can reason, retrieve context, and act within controlled boundaries.

Note: Access to all the resources and actions are strictly controlled by the authorization policies defined at the individual resource level.

Hardware and Efficiency First

Protean AI is designed for lean and sovereign deployments:

  • Quantization support to significantly reduce VRAM and hardware requirements.
  • Adapter-based fine-tuning to avoid retraining full model weights.
  • Efficient runtimes suitable for on-prem, private cloud, VPC, and edge environments.

This enables expert models to run close to enterprise data without risks of data exfiltration.

Developer Experience

Protean AI integrates naturally into existing software development workflows:

  • Explicit configuration with clear intent.
  • Constrained dataset formats that prevent training-time ambiguity.
  • Reproducible experiments and comparable results across runs.
  • Production-oriented APIs for models, RAG, and agents.

Next Steps

To continue:

  • Register a Node and verify available hardware.
  • Explore the Model Registry for base models and checkpoints.
  • Upload or publish an Adapter for efficient specialization.
  • Review Training Dataset Formats to align data with objectives.
  • Follow the Fine-Tuning guide to train a first model or adapter.
  • Prepare a Knowledge set and configure retrieval pipelines.
  • Build and deploy an Agent as the final integration step.

This foundation ensures every step—from raw data, infrastructure to deployed agents—is intentional, efficient, and enterprise-ready.