Arc Intelligence’s cover photo
Arc Intelligence

Arc Intelligence

Software Development

Brooklyn, NY 230 followers

Infrastructure For Agents That Learn

About us

Arc is a research‑driven team building the learning layer for agents. We believe the next decade belongs to systems that learn with you and for you, capturing what works, transferring it across teams, and getting better with every outcome. With Arc, agents build experience. They learn your standards, push back when a choice conflicts with past wins, and share distilled tactics across your organization, creating a self‑improving execution layer. We’re building for an open future where experience is auditable, transferable, and compounding. If this resonates, follow our work and get in touch.

Website
https://www.arc.computer
Industry
Software Development
Company size
2-10 employees
Headquarters
Brooklyn, NY
Type
Privately Held
Founded
2025
Specialties
AI, Machine Learning, and Developer Tools

Locations

Employees at Arc Intelligence

Updates

  • Last night we hosted a hackathon with HackerSquad by Developer Events at betaworks! Huge shout-out to Brian Matzelle for building something that perfectly captures what we're excited about: making agents smarter over time at inference time i.e. getting better on the job. Here's what Brian built with our SDK: https://lnkd.in/erSdNwY4 He had an agent with 10 MCP tools (search, edit, analyze, etc.) and wanted it to get better at tool selection and learn which tools work best for different tasks, without constant retraining. ATLAS wrapped his MCP agent, judged tool efficacy via our reward system, and built learning playbooks from successful patterns. The agent automatically preferred high-reward tool sequences and received Teacher guidance when deviating from proven approaches. Shout-out to Adam Chan for bringing it all together 🚀

    • No alternative text description for this image
  • We're excited to be participating in Hack Night at Betaworks tonight! If you're in the neighbourhood, come say hi 👋

    View profile for Adam Chan

    Bringing developers together to build epic projects with epic tools!

    The fun never ends. Tonight, I'm grateful to host at the betaworks office with the support from their team (I'm incredibly grateful for all of you helping make this event come to life). We'll first hear from Tiger Data (creators of TimescaleDB) where 🤖 Jacky Liang will take the stage first demonstrate their newest capabilities with Agentic Postgres on Tiger Data. Next up we'll hear from Arc Intelligence where Jarrod Barnes and Aman Jaglan will show us demonstrate the power of Infrastructure for Agents That Learn through Atlas. I'm planning to build end to end with the crowd. It's going to be awesome and maybe I'll get to demo something too 😳 This is one night you're not going to want to miss in New York City! Let's goooo!

    • No alternative text description for this image
  • At Arc, similar to other folks at the frontier research of AI, we have been thinking deeply about continual learning. We, too, have followed the research and capabilities of reinforcement learning, continual learning, and the implications for world models. Over the past five months, our team has been particularly obsessed about the following question: What does it actually mean for AI agents to genuinely learn and improve over time? As a result, we’ve explored what might it look like for agents to adapt and learn at inference-time. Instead of asking "How do we update weights without forgetting?" We ask: "Can we skip weight updates entirely?" What did we discover? When we approach continual learning and adaptation as an orchestration problem, we are able to dynamically adjust supervision levels, retrieve distilled experience, and coordinate dual-agent handoffs based on task history. No gradients, no retraining, no specialized hardware required. We evaluated our system on a real cyber-threat investigation benchmark by Microsoft, and proved SOTA performance: 1️⃣ 54.1% task success with GPT-5-mini vs GPT-5 (High)'s 48.0% 2️⃣ 86% lower cost per task 3️⃣ Cross-incident transfer: Frozen artifacts from one scenario improve accuracy 28% → 41% on new scenarios with zero retraining We’re super proud of our approach to continual learning, the results of our benchmark, the open-source system we designed, and what’s next. We’d love for you to read the full paper, check out our blog, and try out the code in the comments below.

  • Arc Intelligence reposted this

    View profile for Jarrod Barnes

    Building Adaptive ML Systems @ Arc

    If you've been spending time at the frontier of AI research, you've likely been flooded with buzzwords like reinforcement learning, continual learning, and or context engineering. What does this all actually mean in practice? We've anchored our entire research at Arc Intelligence on optimizing the $ cost per correct unit of work (increasing the value per token generated by an LLM). That's informed our beliefs of what organizations "hire" AI to do - reliable work done right, at a lower total cost. In order to get there, what needs to be true? Today, that looks like agents adapting and learning at inference-time. Tomorrow, that looks like agents being able to imagine actions, simulate outcomes, and self-correct in a tight, low-cost internal loop. When we approach continual learning and adaptation as an orchestration problem, we are able to dynamically adjust supervision levels, retrieve distilled experience, and coordinate dual-agent handoffs based on task history. No gradients, no retraining, no specialized hardware required. We evaluated our system on a real cyber-threat investigation benchmark by Microsoft, and proved SOTA performance: 54.1% task success with GPT-5-mini vs GPT-5 (High) at 86% lower cost per task. Additionally, we proved cross-task learning. Insight from one scenario improved accuracy 28% → 41% on new scenarios with zero retraining. We’re super proud of our approach to continual learning, the results of our benchmark, the open-source system we designed, and what’s next. We’d love for you to read the full paper, check out our blog, and try out the code below. Read the Blog: https://lnkd.in/ejXi9Yt7 Code and dataset: https://lnkd.in/eZ24EBSZ

  • Getting started with the Atlas SDK? Check out our new demo video 👇 Atlas is a drop-in learning harness that enables your agent to learn from experience, adapt to new challenges, and become more efficient over time - all without modifying your existing agent code. We put the SDK to the test with a supply chain agent, and the results demonstrate a clear "J-curve" of learning. ▶️ How it Works: Atlas wraps your existing agent with an adaptive dual-agent reasoning loop guided by reward signals. This lets agents stay fast on familiar work while escalating supervision on new or risky tasks. The SDK records rich telemetry from these interactions, allowing the agent to learn and adapt over time. ▶️ The Results: The impact is immediate and measurable. After an initial learning dip, the agent's runtime efficiency skyrocketed. By scenario 3, it achieved a +52.8% improvement. After just 5 scenarios, it was running +25.8% more efficiently than its baseline. This translates to tangible outcomes: what was once a verbose, unstructured report from a frontier model becomes a perfectly structured, efficient JSON output. The agent gets faster, cheaper, and more effective with every interaction. Explore the Atlas SDK 👇 GitHub: https://lnkd.in/erSdNwY4 Documentation: https://docs.arc.computer/ #ContinualLearning #Agents

  • 👉 Kicking off our first public release notes for the Atlas SDK, which we'll be sharing regularly. This week: v0.1.3, a continual-learning harness for production agents that turns every task into a structured learning episode. Full post: https://lnkd.in/echUe3Zw Technical Highlights for v0.1.3: 1️⃣ Adaptive Runtime: A core runtime routes each request through a capability probe that selects one of four execution modes: auto, paired, coach, or escalate. This architecture balances autonomous execution with graduated supervision, ensuring low latency on familiar tasks while capturing rich data on complex ones. 2️⃣ Persistent Learning & Memory: The SDK now includes a persistent memory system that captures agent instructions and human guidance from each run, tagged by reward signals. This compounding knowledge base can be automatically reused in future tasks. An optional Postgres backend is supported for production. 3️⃣ Structured Data for Training: Full learning trajectories, including plans, step attempts, guidance notes, and reward payloads, can be exported as structured JSONL. This provides clean, telemetry-rich data ready for evals or training pipelines. 4️⃣ Compatibility: The atlas CLI provides scaffolding for triage adapters (atlas triage init). Agents can be configured via YAML with support for Python, HTTP, and OpenAI-compatible endpoints. The full source code, architecture diagrams, and technical documentation are available on GitHub 👉 https://lnkd.in/erSdNwY4

  • ATLAS SDK is now LIVE: the learning harness for your agents. We enable your agents to learn and adapt in real-time on your most critical, long context workflows. Add our harness to boost performance, reduce token costs and build persistent memory. All you need to do is BYOA: Bring Your Own Agent and your agents evolve from static to dynamic, instead of just executing tasks, they build skills and get better with every run. Here's how to get started: 1. pip install arc-atlas 2. BYOA: plug in your agent (config file) 3. Define what you want to optimize for (rewards) 4. Execute your workflow Arc is fully open-source and deployable on your own infrastructure, giving you complete control over your environment. You own everything, the code, the data, and the skills your system develops and can export datasets at any time for fine-tuning or transferring knowledge elsewhere. A managed cloud version is on the way, so stay tuned. We want to see what you build. Our team is ready to forward-deploy to help you integrate and ship. Try SDK: https://lnkd.in/erSdNwY4 Docs: https://lnkd.in/eFZ6hD4V Talk to us: agent@arc.computer

    • No alternative text description for this image
  • Arc Intelligence reposted this

    View profile for Jarrod Barnes

    Building Adaptive ML Systems @ Arc

    What is continual learning? How is it different from reinforcement learning? Why does it matter for the future of AI systems? Reinforcement Learning optimizes actions in a fixed world, while Continual Learning adapts the model when the world changes. This shift from policy optimization to model adaptation is the key to building agents that can handle dynamic, real-world environments. See our deep dive into the key terms, what matters, and why CL is so critical for creating truly autonomous systems. If you are building or conducting research in this space, would love to connect! 👉 Full Blog: https://lnkd.in/eWpjzrqb

    • No alternative text description for this image
  • Arc Intelligence reposted this

    View profile for Jarrod Barnes

    Building Adaptive ML Systems @ Arc

    I had the chance to present our work on real-time continual learning for agents at Cloudflare's HQ last week, and wanted to share the core ideas, slides, and full thesis here. Our research is driven by one question: "If an AI agent were to learn like a human, what would its architecture look like?" Model capability will continue to improve - the gap for translating that capability is better learning environments. As a former educator and coach, our thesis is that there is no better classroom than a live production environment. We built the infrastructure to turn it into one. We first looked at the related works in hierarchical reasoning emerging in physical AI and the proven principles of human pedagogy to determine the most efficient way to achieve this is by first decoupling planning from execution. We began experimentation with a dual-agent system - a "teacher" model that guides any agent (the student). It diagnoses the student's reasoning before offering help. This pedagogical approach is the key to making learning reliable. However, a teacher is only as good as its ability to evaluate and assess learning. A reliable reward signal (feedback) is what scales a learning loop, acting as the trusted verifier for what an agent has actually learned. We then built a novel reward system that set a new state-of-the-art on RewardBench V2. Which enables developers to configure what learning they want the system to optimize for. To scale this capability, we designed a hybrid reinforcement learning engine that separates training from adaptation: Offline for reliability: Hitting a 97% non-degradation rate. Online for speed: Seeing +165% performance boosts in under 2 hours. The novel long tail here is true autonomous skill transfer - the same way a master can teach an apprentice. Train one agent, the entire network can learn. How? We trained a teacher on hard math problems and watched it help an agent get 6x better at an unrelated enterprise task by transferring its abstract problem-solving process. Effectively take full trajectories from “Agent A”, distill the learned behavior, and transfer it to Agent B (and C) with non-degradation guarantees. This is the foundation for moving from disposable AI to durable systems with persistent memory and skills. Building this with a model-agnostic infrastructure gives developers the control to build compounding intelligence on their own terms, without vendor lock-in. We're building this in the open at Arc Intelligence. All our research, models, and code are available. Research: https://lnkd.in/ecyd78Eg ATLAS Technical Report: https://lnkd.in/e3SVbijQ "Bridging The Judgment Gap" Paper: https://lnkd.in/efrqP_-s GitHub (Code & Models): https://lnkd.in/eC2EYRi2 If you're building agents or thinking about this space, I'd love to hear your perspective.

Similar pages