Overview
Marco-o1 not only focuses on disciplines with standard answers, such as mathematics, physics, and coding—which are well-suited for reinforcement learning (RL)—but also places greater emphasis on open-ended resolutions. We aim to address the question: "Can the o1 model effectively generalize to broader domains where clear standards are absent and rewards are challenging to quantify?"
Currently, Marco-o1 Large Language Model (LLM) is powered by Chain-of-Thought (CoT) fine-tuning, Monte Carlo Tree Search (MCTS), reflection mechanisms, and innovative reasoning strategies—optimized for complex real-world problem-solving tasks.
Variants
No | Variant | Cortex CLI command |
---|---|---|
1 | Marco-o1-8b | cortex run marco-o1:8b |
Use it with Jan (UI)
- Install Jan using Quickstart
- Use in Jan model Hub:
cortexhub/marco-o1
Use it with Cortex (CLI)
- Install Cortex using Quickstart
- Run the model with command:
cortex run marco-o1
Credits
- Downloads last month
- 566
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The model has no library tag.