arc-atlas


Namearc-atlas JSON
Version 0.1.8 PyPI version JSON
download
home_pageNone
SummaryTeacher/Student orchestration toolkit for Bring-Your-Own-Agent workflows.
upload_time2025-10-24 01:08:38
maintainerNone
docs_urlNone
authorNone
requires_python>=3.10
licenseNone
keywords autonomy agents langchain langgraph reasoning reward-models
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage No coveralls.
            # Atlas SDK — PyPI Quickstart

Atlas wraps your Bring-Your-Own-Agent (BYOA) in a guided Teacher → Student → Reward loop. Install the SDK from PyPI, point it at your agent, and Atlas handles planning, orchestration, evaluation, and optional persistence for you.

> Atlas defaults to an in-memory workflow—leave `storage: null` in your config for quick experiments. You can add PostgreSQL later if you want durable telemetry.

## What's New in v0.1.8

- **Autodiscovery & CLI Upgrades** – `atlas env init` now scaffolds full configs, auto-loads `.env`/`PYTHONPATH`, and can replay discoveries with `atlas run --config` or the fake LLM smoke-test path (`ATLAS_FAKE_LLM=1`) to validate stacks offline ([#52](https://github.com/Arc-Computer/atlas-sdk/pull/52), [#70](https://github.com/Arc-Computer/atlas-sdk/pull/70), [#74](https://github.com/Arc-Computer/atlas-sdk/pull/74), [#75](https://github.com/Arc-Computer/atlas-sdk/pull/75)).
- **Learning Playbooks in Runtime** – Student and Teacher personas fetch hashed “learning playbooks”, inject them into every planner/synthesizer/executor prompt, and track metadata so cached prompts stay in sync when playbooks change ([#76](https://github.com/Arc-Computer/atlas-sdk/pull/76)).
- **Persistent Telemetry & Learning Reports** – Discovery and runtime sessions log directly to Postgres, and the new learning evaluation harness can filter by project/task/tags while generating model-level breakdowns in Markdown/JSON reports ([#72](https://github.com/Arc-Computer/atlas-sdk/pull/72), [#73](https://github.com/Arc-Computer/atlas-sdk/pull/73)).
- **Safety Guardrails & Approvals** – Session exports require explicit approval, with CLI tooling to review/approve/quarantine runs and drift alerts captured alongside reward metadata ([#63](https://github.com/Arc-Computer/atlas-sdk/pull/63)).
- **Expanded Evaluation Suites** – Added capability probe updates (xAI Grok support), dual-agent runtime benchmarking, and a reward model harness with packaged datasets and docs to keep offline validation comprehensive ([#55](https://github.com/Arc-Computer/atlas-sdk/pull/55), [#56](https://github.com/Arc-Computer/atlas-sdk/pull/56), [#57](https://github.com/Arc-Computer/atlas-sdk/pull/57)).
- **Lean Learning History Payloads** – Capability probe history now respects an operator-defined cap, trims noisy fields, and keeps streak stats lightweight for faster probes ([#54](https://github.com/Arc-Computer/atlas-sdk/pull/54)).

## What's New in v0.1.7

- **Adaptive Runtime** – Capability probe selects execution mode (`auto`, `paired`, `coach`, `escalate`) per request based on task complexity and historical performance.
- **Persistent Learning Memory** – Guidance from each episode is tagged by reward and automatically reused on similar tasks.
- **Fingerprint-Based Certification** – First-run tasks get certified, enabling auto mode on future similar requests when confidence is high.

## Install in Minutes

```bash
python -m venv .venv
source .venv/bin/activate  # Windows: .venv\Scripts\activate
pip install --upgrade pip
pip install arc-atlas
```

- Python 3.10 or newer is required (3.13 recommended).
- For development tooling and tests, install extras with `pip install arc-atlas[dev]`.

## Configure Your Environment

Set API keys before running Atlas:

```bash
export OPENAI_API_KEY=sk-... # your api key
export GEMINI_API_KEY=... # for reward system
```

Prefer storing secrets in a `.env` file? The SDK automatically loads it on startup (via `python-dotenv`), so CLI commands and examples pick up those values without manual exports.

Atlas reads additional provider keys from adapter-specific `llm.api_key_env` fields.

## Create a Minimal Config

Save the following as `atlas_quickstart.yaml` (storage disabled by default):

```yaml
agent:
  type: openai
  name: quickstart-openai-agent
  system_prompt: |
    You are an Agent. Follow instructions carefully and keep responses concise.
  tools: []
  llm:
    provider: openai
    model: gpt-4o-mini
    api_key_env: OPENAI_API_KEY
    temperature: 0.1
    max_output_tokens: 1024
student:
  max_plan_tokens: 1024
  max_step_tokens: 1024
  max_synthesis_tokens: 1024
teacher:
  llm:
    provider: openai
    model: gpt-4o-mini
    api_key_env: OPENAI_API_KEY
    temperature: 0.1
    max_output_tokens: 768
orchestration:
  max_retries: 1
  step_timeout_seconds: 600
  emit_intermediate_steps: true
rim:
  small_model:
    provider: gemini
    model: gemini/gemini-2.5-flash
    api_key_env: GEMINI_API_KEY
    max_output_tokens: 8096
  large_model:
    provider: gemini
    model: gemini/gemini-2.5-flash
    api_key_env: GEMINI_API_KEY
    max_output_tokens: 8096
  judge_prompt: 'reward the agent for attending the issues mentioned in the task'
  variance_threshold: 0.15
  uncertainty_threshold: 0.3
storage: null
```

## Run Your First Task

```python
from atlas import core

result = core.run(
    task="Summarise the latest Atlas SDK updates",
    config_path="atlas_quickstart.yaml",
    stream_progress=True,
)

print(result.final_answer)
```

`result` is an `atlas.types.Result` containing the final answer, reviewed plan, and per-step evaluations. Set `stream_progress=True` to mirror planner/executor telemetry in your terminal.
The console summary includes the adaptive mode, confidence, certification flag, and session reward so you can watch the J-curve without any database setup.

Need the structured metadata? Access `ExecutionContext.get().metadata` after the run or export later via the CLI once storage is configured.

## Wrap Your Existing Agent

### OpenAI-Compatible Chat Agent

```python
from atlas import core
from atlas.connectors import create_adapter
from atlas.config.models import OpenAIAdapterConfig

adapter = create_adapter(OpenAIAdapterConfig(
    type="openai",
    name="my-openai-agent",
    system_prompt="You are a helpful assistant.",
    tools=[],
    llm={
        "provider": "openai",
        "model": "gpt-4o-mini",
        "api_key_env": "OPENAI_API_KEY",
    },
))

result = core.run(
    task="Draft a product brief for Atlas",
    config_path="atlas_quickstart.yaml",
    adapter_override=adapter,
)
```

Override the adapter to reuse the same orchestration settings with different agents.

### Local Python Function

```python
# my_agent.py
def respond(prompt: str, metadata: dict | None = None) -> str:
    return f"echo: {prompt}"
```

Update the config’s `agent` block:

```yaml
agent:
  type: python
  name: local-function-agent
  system_prompt: |
    You call a local Python function named respond.
  import_path: my_agent
  attribute: respond
  tools: []
```

Atlas imports your callable (optionally from `working_directory`), handles async execution, generator outputs, and metadata passing.

### HTTP Endpoint

```yaml
agent:
  type: http_api
  name: http-agent
  system_prompt: |
    You delegate work to a REST endpoint that accepts {"prompt": "..."}.
  transport:
    base_url: https://your-agent.example.com/v1/atlas
    timeout_seconds: 60
  payload_template:
    prompt: "{{ prompt }}"
  result_path: ["data", "output"]
  tools:
    - name: web_search
      description: Search the web.
      parameters:
        type: object
        properties:
          query:
            type: string
        required: [query]
```

Atlas retries requests based on the adapter’s `retry` policy and normalises JSON responses using `result_path`.

## Optional: Persist Runs with PostgreSQL

```bash
# Start a local Postgres via Docker (installs Docker if missing)
atlas init  # writes atlas-postgres.yaml, starts Postgres, and applies the schema

# Or run docker compose yourself if you prefer:
# docker compose -f docker/docker-compose.yaml up -d postgres

# Point Atlas at the database
export STORAGE__DATABASE_URL=postgresql://atlas:atlas@localhost:5433/atlas
```

Add a `storage` section to your config when you want Atlas to log plans, attempts, and telemetry into Postgres for later inspection. If Docker isn’t available, install Postgres manually and provide the same connection URL.

## Observe and Export

- Set `stream_progress=True` in `core.run` to stream planner/executor/judge events alongside the adaptive summary.
- Export stored sessions with `arc-atlas --database-url postgresql://... --output traces.jsonl`—the JSONL includes `adaptive_summary`, `session_reward`, per-session learning notes, the consolidated `learning_state`, and aggregated history.
- Explore `docs/examples/` for telemetry and export walkthroughs.

## Train with Atlas Core

Use the SDK CLI to bridge runtime traces into the Atlas Core training pipeline:

```bash
git clone https://github.com/Arc-Computer/ATLAS ~/src/ATLAS
export ATLAS_CORE_PATH=~/src/ATLAS
export STORAGE__DATABASE_URL=postgresql://atlas:atlas@localhost:5433/atlas

atlas train --config-name offline/base --dry-run
# inspect the command, then rerun without --dry-run to execute training
```

`atlas train` writes a JSONL export to `<atlas-core-path>/exports/<timestamp>.jsonl` and then executes `scripts/run_offline_pipeline.py` from that directory. You can point `--output` at a custom path, forward Hydra overrides with repeated `--override` flags, or use `--output-dir` / `--wandb-project` to steer checkpoints and logging. Pass `--use-sample-dataset` to copy the bundled sample dataset when you just want to validate the workflow without hitting Postgres.

## Next Steps

- Browse `configs/examples/` for richer orchestration templates.
- Enable RIM judges by toggling `rim.active_judges`.
- Integrate Atlas into async services with `core.arun`.

            

Raw data

            {
    "_id": null,
    "home_page": null,
    "name": "arc-atlas",
    "maintainer": null,
    "docs_url": null,
    "requires_python": ">=3.10",
    "maintainer_email": null,
    "keywords": "autonomy, agents, langchain, langgraph, reasoning, reward-models",
    "author": null,
    "author_email": null,
    "download_url": "https://files.pythonhosted.org/packages/4f/4e/4f4853289224305f81d95bb871518acd419aac431e6f1aaec87aac953ca4/arc_atlas-0.1.8.tar.gz",
    "platform": null,
    "description": "# Atlas SDK \u2014 PyPI Quickstart\n\nAtlas wraps your Bring-Your-Own-Agent (BYOA) in a guided Teacher \u2192 Student \u2192 Reward loop. Install the SDK from PyPI, point it at your agent, and Atlas handles planning, orchestration, evaluation, and optional persistence for you.\n\n> Atlas defaults to an in-memory workflow\u2014leave `storage: null` in your config for quick experiments. You can add PostgreSQL later if you want durable telemetry.\n\n## What's New in v0.1.8\n\n- **Autodiscovery & CLI Upgrades** \u2013 `atlas env init` now scaffolds full configs, auto-loads `.env`/`PYTHONPATH`, and can replay discoveries with `atlas run --config` or the fake LLM smoke-test path (`ATLAS_FAKE_LLM=1`) to validate stacks offline ([#52](https://github.com/Arc-Computer/atlas-sdk/pull/52), [#70](https://github.com/Arc-Computer/atlas-sdk/pull/70), [#74](https://github.com/Arc-Computer/atlas-sdk/pull/74), [#75](https://github.com/Arc-Computer/atlas-sdk/pull/75)).\n- **Learning Playbooks in Runtime** \u2013 Student and Teacher personas fetch hashed \u201clearning playbooks\u201d, inject them into every planner/synthesizer/executor prompt, and track metadata so cached prompts stay in sync when playbooks change ([#76](https://github.com/Arc-Computer/atlas-sdk/pull/76)).\n- **Persistent Telemetry & Learning Reports** \u2013 Discovery and runtime sessions log directly to Postgres, and the new learning evaluation harness can filter by project/task/tags while generating model-level breakdowns in Markdown/JSON reports ([#72](https://github.com/Arc-Computer/atlas-sdk/pull/72), [#73](https://github.com/Arc-Computer/atlas-sdk/pull/73)).\n- **Safety Guardrails & Approvals** \u2013 Session exports require explicit approval, with CLI tooling to review/approve/quarantine runs and drift alerts captured alongside reward metadata ([#63](https://github.com/Arc-Computer/atlas-sdk/pull/63)).\n- **Expanded Evaluation Suites** \u2013 Added capability probe updates (xAI Grok support), dual-agent runtime benchmarking, and a reward model harness with packaged datasets and docs to keep offline validation comprehensive ([#55](https://github.com/Arc-Computer/atlas-sdk/pull/55), [#56](https://github.com/Arc-Computer/atlas-sdk/pull/56), [#57](https://github.com/Arc-Computer/atlas-sdk/pull/57)).\n- **Lean Learning History Payloads** \u2013 Capability probe history now respects an operator-defined cap, trims noisy fields, and keeps streak stats lightweight for faster probes ([#54](https://github.com/Arc-Computer/atlas-sdk/pull/54)).\n\n## What's New in v0.1.7\n\n- **Adaptive Runtime** \u2013 Capability probe selects execution mode (`auto`, `paired`, `coach`, `escalate`) per request based on task complexity and historical performance.\n- **Persistent Learning Memory** \u2013 Guidance from each episode is tagged by reward and automatically reused on similar tasks.\n- **Fingerprint-Based Certification** \u2013 First-run tasks get certified, enabling auto mode on future similar requests when confidence is high.\n\n## Install in Minutes\n\n```bash\npython -m venv .venv\nsource .venv/bin/activate  # Windows: .venv\\Scripts\\activate\npip install --upgrade pip\npip install arc-atlas\n```\n\n- Python 3.10 or newer is required (3.13 recommended).\n- For development tooling and tests, install extras with `pip install arc-atlas[dev]`.\n\n## Configure Your Environment\n\nSet API keys before running Atlas:\n\n```bash\nexport OPENAI_API_KEY=sk-... # your api key\nexport GEMINI_API_KEY=... # for reward system\n```\n\nPrefer storing secrets in a `.env` file? The SDK automatically loads it on startup (via `python-dotenv`), so CLI commands and examples pick up those values without manual exports.\n\nAtlas reads additional provider keys from adapter-specific `llm.api_key_env` fields.\n\n## Create a Minimal Config\n\nSave the following as `atlas_quickstart.yaml` (storage disabled by default):\n\n```yaml\nagent:\n  type: openai\n  name: quickstart-openai-agent\n  system_prompt: |\n    You are an Agent. Follow instructions carefully and keep responses concise.\n  tools: []\n  llm:\n    provider: openai\n    model: gpt-4o-mini\n    api_key_env: OPENAI_API_KEY\n    temperature: 0.1\n    max_output_tokens: 1024\nstudent:\n  max_plan_tokens: 1024\n  max_step_tokens: 1024\n  max_synthesis_tokens: 1024\nteacher:\n  llm:\n    provider: openai\n    model: gpt-4o-mini\n    api_key_env: OPENAI_API_KEY\n    temperature: 0.1\n    max_output_tokens: 768\norchestration:\n  max_retries: 1\n  step_timeout_seconds: 600\n  emit_intermediate_steps: true\nrim:\n  small_model:\n    provider: gemini\n    model: gemini/gemini-2.5-flash\n    api_key_env: GEMINI_API_KEY\n    max_output_tokens: 8096\n  large_model:\n    provider: gemini\n    model: gemini/gemini-2.5-flash\n    api_key_env: GEMINI_API_KEY\n    max_output_tokens: 8096\n  judge_prompt: 'reward the agent for attending the issues mentioned in the task'\n  variance_threshold: 0.15\n  uncertainty_threshold: 0.3\nstorage: null\n```\n\n## Run Your First Task\n\n```python\nfrom atlas import core\n\nresult = core.run(\n    task=\"Summarise the latest Atlas SDK updates\",\n    config_path=\"atlas_quickstart.yaml\",\n    stream_progress=True,\n)\n\nprint(result.final_answer)\n```\n\n`result` is an `atlas.types.Result` containing the final answer, reviewed plan, and per-step evaluations. Set `stream_progress=True` to mirror planner/executor telemetry in your terminal.\nThe console summary includes the adaptive mode, confidence, certification flag, and session reward so you can watch the J-curve without any database setup.\n\nNeed the structured metadata? Access `ExecutionContext.get().metadata` after the run or export later via the CLI once storage is configured.\n\n## Wrap Your Existing Agent\n\n### OpenAI-Compatible Chat Agent\n\n```python\nfrom atlas import core\nfrom atlas.connectors import create_adapter\nfrom atlas.config.models import OpenAIAdapterConfig\n\nadapter = create_adapter(OpenAIAdapterConfig(\n    type=\"openai\",\n    name=\"my-openai-agent\",\n    system_prompt=\"You are a helpful assistant.\",\n    tools=[],\n    llm={\n        \"provider\": \"openai\",\n        \"model\": \"gpt-4o-mini\",\n        \"api_key_env\": \"OPENAI_API_KEY\",\n    },\n))\n\nresult = core.run(\n    task=\"Draft a product brief for Atlas\",\n    config_path=\"atlas_quickstart.yaml\",\n    adapter_override=adapter,\n)\n```\n\nOverride the adapter to reuse the same orchestration settings with different agents.\n\n### Local Python Function\n\n```python\n# my_agent.py\ndef respond(prompt: str, metadata: dict | None = None) -> str:\n    return f\"echo: {prompt}\"\n```\n\nUpdate the config\u2019s `agent` block:\n\n```yaml\nagent:\n  type: python\n  name: local-function-agent\n  system_prompt: |\n    You call a local Python function named respond.\n  import_path: my_agent\n  attribute: respond\n  tools: []\n```\n\nAtlas imports your callable (optionally from `working_directory`), handles async execution, generator outputs, and metadata passing.\n\n### HTTP Endpoint\n\n```yaml\nagent:\n  type: http_api\n  name: http-agent\n  system_prompt: |\n    You delegate work to a REST endpoint that accepts {\"prompt\": \"...\"}.\n  transport:\n    base_url: https://your-agent.example.com/v1/atlas\n    timeout_seconds: 60\n  payload_template:\n    prompt: \"{{ prompt }}\"\n  result_path: [\"data\", \"output\"]\n  tools:\n    - name: web_search\n      description: Search the web.\n      parameters:\n        type: object\n        properties:\n          query:\n            type: string\n        required: [query]\n```\n\nAtlas retries requests based on the adapter\u2019s `retry` policy and normalises JSON responses using `result_path`.\n\n## Optional: Persist Runs with PostgreSQL\n\n```bash\n# Start a local Postgres via Docker (installs Docker if missing)\natlas init  # writes atlas-postgres.yaml, starts Postgres, and applies the schema\n\n# Or run docker compose yourself if you prefer:\n# docker compose -f docker/docker-compose.yaml up -d postgres\n\n# Point Atlas at the database\nexport STORAGE__DATABASE_URL=postgresql://atlas:atlas@localhost:5433/atlas\n```\n\nAdd a `storage` section to your config when you want Atlas to log plans, attempts, and telemetry into Postgres for later inspection. If Docker isn\u2019t available, install Postgres manually and provide the same connection URL.\n\n## Observe and Export\n\n- Set `stream_progress=True` in `core.run` to stream planner/executor/judge events alongside the adaptive summary.\n- Export stored sessions with `arc-atlas --database-url postgresql://... --output traces.jsonl`\u2014the JSONL includes `adaptive_summary`, `session_reward`, per-session learning notes, the consolidated `learning_state`, and aggregated history.\n- Explore `docs/examples/` for telemetry and export walkthroughs.\n\n## Train with Atlas Core\n\nUse the SDK CLI to bridge runtime traces into the Atlas Core training pipeline:\n\n```bash\ngit clone https://github.com/Arc-Computer/ATLAS ~/src/ATLAS\nexport ATLAS_CORE_PATH=~/src/ATLAS\nexport STORAGE__DATABASE_URL=postgresql://atlas:atlas@localhost:5433/atlas\n\natlas train --config-name offline/base --dry-run\n# inspect the command, then rerun without --dry-run to execute training\n```\n\n`atlas train` writes a JSONL export to `<atlas-core-path>/exports/<timestamp>.jsonl` and then executes `scripts/run_offline_pipeline.py` from that directory. You can point `--output` at a custom path, forward Hydra overrides with repeated `--override` flags, or use `--output-dir` / `--wandb-project` to steer checkpoints and logging. Pass `--use-sample-dataset` to copy the bundled sample dataset when you just want to validate the workflow without hitting Postgres.\n\n## Next Steps\n\n- Browse `configs/examples/` for richer orchestration templates.\n- Enable RIM judges by toggling `rim.active_judges`.\n- Integrate Atlas into async services with `core.arun`.\n",
    "bugtrack_url": null,
    "license": null,
    "summary": "Teacher/Student orchestration toolkit for Bring-Your-Own-Agent workflows.",
    "version": "0.1.8",
    "project_urls": {
        "Documentation": "https://docs.arc.computer/introduction",
        "Issues": "https://github.com/Arc-Computer/atlas-sdk/issues",
        "Source": "https://github.com/Arc-Computer/atlas-sdk"
    },
    "split_keywords": [
        "autonomy",
        " agents",
        " langchain",
        " langgraph",
        " reasoning",
        " reward-models"
    ],
    "urls": [
        {
            "comment_text": null,
            "digests": {
                "blake2b_256": "e2228ad3f3902977788891ca6df1615c2b36cec8ef7b0ec2936844b53476dc76",
                "md5": "dc0640d39246a0cbc90a36b36b8df463",
                "sha256": "d4a639c98f18c0e6854264d83c811b9726c710a407f55beffaee2bc93361c24c"
            },
            "downloads": -1,
            "filename": "arc_atlas-0.1.8-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "dc0640d39246a0cbc90a36b36b8df463",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": ">=3.10",
            "size": 493814,
            "upload_time": "2025-10-24T01:08:37",
            "upload_time_iso_8601": "2025-10-24T01:08:37.324350Z",
            "url": "https://files.pythonhosted.org/packages/e2/22/8ad3f3902977788891ca6df1615c2b36cec8ef7b0ec2936844b53476dc76/arc_atlas-0.1.8-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": null,
            "digests": {
                "blake2b_256": "4f4e4f4853289224305f81d95bb871518acd419aac431e6f1aaec87aac953ca4",
                "md5": "25c95e0cf9f5480de30f3e353f0c72dc",
                "sha256": "2c8669eb8320a9c3b338913b86da72e3a50e77ae32e7a7086c4b9feb68e34221"
            },
            "downloads": -1,
            "filename": "arc_atlas-0.1.8.tar.gz",
            "has_sig": false,
            "md5_digest": "25c95e0cf9f5480de30f3e353f0c72dc",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.10",
            "size": 466690,
            "upload_time": "2025-10-24T01:08:38",
            "upload_time_iso_8601": "2025-10-24T01:08:38.896648Z",
            "url": "https://files.pythonhosted.org/packages/4f/4e/4f4853289224305f81d95bb871518acd419aac431e6f1aaec87aac953ca4/arc_atlas-0.1.8.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2025-10-24 01:08:38",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "Arc-Computer",
    "github_project": "atlas-sdk",
    "travis_ci": false,
    "coveralls": false,
    "github_actions": true,
    "lcname": "arc-atlas"
}
        
Elapsed time: 1.95158s