From cd537c1af6731965817eed2ae32b8dd8ea9d0b5e Mon Sep 17 00:00:00 2001 From: Timothy Kassis Date: Wed, 21 Jan 2026 09:16:34 -0500 Subject: [PATCH] Replace all instances of Sonnet to Opus --- scientific-skills/markitdown/SKILL.md | 6 +++--- .../markitdown/assets/example_usage.md | 8 ++++---- .../markitdown/references/api_reference.md | 13 +++++-------- .../markitdown/scripts/convert_with_ai.py | 19 ++++++++----------- .../references/AI_CURATION.md | 4 ++-- .../references/api_reference.md | 4 ++-- 6 files changed, 24 insertions(+), 30 deletions(-) diff --git a/scientific-skills/markitdown/SKILL.md b/scientific-skills/markitdown/SKILL.md index fb42f29..eb42c4c 100644 --- a/scientific-skills/markitdown/SKILL.md +++ b/scientific-skills/markitdown/SKILL.md @@ -139,7 +139,7 @@ client = OpenAI( md = MarkItDown( llm_client=client, - llm_model="anthropic/claude-sonnet-4.5", # recommended for scientific vision + llm_model="anthropic/claude-opus-4.5", # recommended for scientific vision llm_prompt="Describe this image in detail for scientific documentation" ) @@ -262,7 +262,7 @@ client = OpenAI( md = MarkItDown( llm_client=client, - llm_model="anthropic/claude-sonnet-4.5", # recommended for presentations + llm_model="anthropic/claude-opus-4.5", # recommended for presentations llm_prompt="Describe this slide image in detail, focusing on key visual elements and data" ) @@ -419,7 +419,7 @@ client = OpenAI( md_ai = MarkItDown( llm_client=client, - llm_model="anthropic/claude-sonnet-4.5", + llm_model="anthropic/claude-opus-4.5", llm_prompt="Describe scientific figures with technical precision" ) ``` diff --git a/scientific-skills/markitdown/assets/example_usage.md b/scientific-skills/markitdown/assets/example_usage.md index 8eef213..ffd966e 100644 --- a/scientific-skills/markitdown/assets/example_usage.md +++ b/scientific-skills/markitdown/assets/example_usage.md @@ -101,7 +101,7 @@ from openai import OpenAI client = OpenAI() md = MarkItDown( llm_client=client, - llm_model="anthropic/claude-sonnet-4.5", + llm_model="anthropic/claude-opus-4.5", llm_prompt="Describe this scientific slide, focusing on data and key findings" ) @@ -143,7 +143,7 @@ Be technical and precise. md = MarkItDown( llm_client=client, - llm_model="anthropic/claude-sonnet-4.5", # recommended for scientific vision + llm_model="anthropic/claude-opus-4.5", # recommended for scientific vision llm_prompt=scientific_prompt ) @@ -167,14 +167,14 @@ client = OpenAI( # Scientific papers - use Claude for technical analysis scientific_md = MarkItDown( llm_client=client, - llm_model="anthropic/claude-sonnet-4.5", + llm_model="anthropic/claude-opus-4.5", llm_prompt="Describe scientific figures with technical precision" ) # Presentations - use GPT-4o for visual understanding presentation_md = MarkItDown( llm_client=client, - llm_model="anthropic/claude-sonnet-4.5", + llm_model="anthropic/claude-opus-4.5", llm_prompt="Summarize slide content and key visual elements" ) diff --git a/scientific-skills/markitdown/references/api_reference.md b/scientific-skills/markitdown/references/api_reference.md index 90ac446..bade33d 100644 --- a/scientific-skills/markitdown/references/api_reference.md +++ b/scientific-skills/markitdown/references/api_reference.md @@ -23,7 +23,7 @@ md = MarkItDown( | Parameter | Type | Default | Description | |-----------|------|---------|-------------| | `llm_client` | OpenAI client | `None` | OpenAI-compatible client for AI image descriptions | -| `llm_model` | str | `None` | Model name (e.g., "anthropic/claude-sonnet-4.5") for image descriptions | +| `llm_model` | str | `None` | Model name (e.g., "anthropic/claude-opus-4.5") for image descriptions | | `llm_prompt` | str | `None` | Custom prompt for image description | | `docintel_endpoint` | str | `None` | Azure Document Intelligence endpoint | | `enable_plugins` | bool | `False` | Enable 3rd-party plugins | @@ -235,7 +235,7 @@ client = OpenAI( # Create MarkItDown with AI support md = MarkItDown( llm_client=client, - llm_model="anthropic/claude-sonnet-4.5", # recommended for scientific vision + llm_model="anthropic/claude-opus-4.5", # recommended for scientific vision llm_prompt="Describe this image in detail for scientific documentation" ) @@ -246,11 +246,8 @@ result = md.convert("presentation.pptx") ### Available Models via OpenRouter Popular models with vision support: -- `anthropic/claude-sonnet-4.5` - **Recommended for scientific vision** -- `anthropic/claude-opus-4.5` - Advanced vision model -- `openai/gpt-4o` - GPT-4 Omni -- `openai/gpt-4-vision` - GPT-4 Vision -- `google/gemini-pro-vision` - Gemini Pro Vision +- `anthropic/claude-opus-4.5` - **Recommended for scientific vision** +- `google/gemini-3-pro-preview` - Gemini Pro Vision See https://openrouter.ai/models for the complete list. @@ -269,7 +266,7 @@ Be precise and technical. md = MarkItDown( llm_client=client, - llm_model="anthropic/claude-sonnet-4.5", + llm_model="anthropic/claude-opus-4.5", llm_prompt=scientific_prompt ) ``` diff --git a/scientific-skills/markitdown/scripts/convert_with_ai.py b/scientific-skills/markitdown/scripts/convert_with_ai.py index 47dcebc..9eb8440 100755 --- a/scientific-skills/markitdown/scripts/convert_with_ai.py +++ b/scientific-skills/markitdown/scripts/convert_with_ai.py @@ -71,7 +71,7 @@ def convert_with_ai( input_file: Path, output_file: Path, api_key: str, - model: str = "anthropic/claude-sonnet-4.5", + model: str = "anthropic/claude-opus-4.5", prompt_type: str = "general", custom_prompt: str = None ) -> bool: @@ -82,7 +82,7 @@ def convert_with_ai( input_file: Path to input file output_file: Path to output Markdown file api_key: OpenRouter API key - model: Model name (default: anthropic/claude-sonnet-4.5) + model: Model name (default: anthropic/claude-opus-4.5) prompt_type: Type of prompt to use custom_prompt: Custom prompt (overrides prompt_type) @@ -154,10 +154,10 @@ Examples: python convert_with_ai.py paper.pdf output.md --prompt-type scientific # Convert a presentation with custom model - python convert_with_ai.py slides.pptx slides.md --model anthropic/claude-sonnet-4.5 --prompt-type presentation + python convert_with_ai.py slides.pptx slides.md --model anthropic/claude-opus-4.5 --prompt-type presentation # Use custom prompt with advanced vision model - python convert_with_ai.py diagram.png diagram.md --model anthropic/claude-sonnet-4.5 --custom-prompt "Describe this technical diagram" + python convert_with_ai.py diagram.png diagram.md --model anthropic/claude-opus-4.5 --custom-prompt "Describe this technical diagram" # Set API key via environment variable export OPENROUTER_API_KEY="sk-or-v1-..." @@ -167,11 +167,8 @@ Environment Variables: OPENROUTER_API_KEY OpenRouter API key (required if not passed via --api-key) Popular Models (use with --model): - anthropic/claude-sonnet-4.5 - Recommended for scientific vision - anthropic/claude-opus-4.5 - Advanced vision model - openai/gpt-4o - GPT-4 Omni (vision support) - openai/gpt-4-vision - GPT-4 Vision - google/gemini-pro-vision - Gemini Pro Vision + anthropic/claude-opus-4.5 - Recommended for scientific vision + google/gemini-3-pro-preview - Gemini Pro Vision """ ) @@ -183,8 +180,8 @@ Popular Models (use with --model): ) parser.add_argument( '--model', '-m', - default='anthropic/claude-sonnet-4.5', - help='Model to use via OpenRouter (default: anthropic/claude-sonnet-4.5)' + default='anthropic/claude-opus-4.5', + help='Model to use via OpenRouter (default: anthropic/claude-opus-4.5)' ) parser.add_argument( '--prompt-type', '-t', diff --git a/scientific-skills/neuropixels-analysis/references/AI_CURATION.md b/scientific-skills/neuropixels-analysis/references/AI_CURATION.md index 0d60274..b0d8650 100644 --- a/scientific-skills/neuropixels-analysis/references/AI_CURATION.md +++ b/scientific-skills/neuropixels-analysis/references/AI_CURATION.md @@ -67,7 +67,7 @@ result = npa.analyze_unit_visually( analyzer, unit_id=0, api_client=client, - model='claude-3-5-sonnet-20241022', + model='claude-opus-4.5', task='quality_assessment' ) @@ -287,7 +287,7 @@ Currently supported APIs: | Provider | Client | Model Examples | |----------|--------|----------------| -| Anthropic | `anthropic.Anthropic()` | claude-3-5-sonnet-20241022 | +| Anthropic | `anthropic.Anthropic()` | claude-opus-4.5 | | OpenAI | `openai.OpenAI()` | gpt-4-vision-preview | | Google | `google.generativeai` | gemini-pro-vision | diff --git a/scientific-skills/neuropixels-analysis/references/api_reference.md b/scientific-skills/neuropixels-analysis/references/api_reference.md index 395ba44..01f9032 100644 --- a/scientific-skills/neuropixels-analysis/references/api_reference.md +++ b/scientific-skills/neuropixels-analysis/references/api_reference.md @@ -284,7 +284,7 @@ npa.analyze_unit_visually( analyzer: SortingAnalyzer, unit_id: int, api_client: Any = None, - model: str = 'claude-3-5-sonnet-20241022', + model: str = 'claude-opus-4.5', task: str = 'quality_assessment', custom_prompt: str = None, ) -> dict @@ -304,7 +304,7 @@ npa.batch_visual_curation( analyzer: SortingAnalyzer, unit_ids: list = None, api_client: Any = None, - model: str = 'claude-3-5-sonnet-20241022', + model: str = 'claude-opus-4.5', output_dir: str = None, progress_callback: callable = None, ) -> dict