← Nick0
AI Agent · Computer Vision · 2025

Friday

An autonomous agent that controls your Mac. Sees the screen, understands context, and executes tasks with full keyboard and mouse autonomy.

2025 · Solo project · macOS

AI assistants can talk but can't act

Every AI assistant in 2025 works the same way: you type, it responds with text. But what if you want it to actually do things on your computer? Open apps, fill forms, navigate interfaces, execute multi-step workflows?

The gap between "AI that answers questions" and "AI that does work" is enormous. Bridging it requires an agent that can see, understand, and act in a real desktop environment.

See the screen, control the machine

Friday is an autonomous agent that operates your Mac the way a human would: by looking at the screen and using the keyboard and mouse.

Give Friday a natural language instruction and it figures out the steps, navigates the interface, and completes the task — handling errors and unexpected states along the way.

What Friday can do

Any
macOS app
Full
Autonomy
Multi
LLM backend

The bridge to physical robots

Friday isn't just a productivity tool. It's a proof of concept for a deeper idea: if an AI can perceive a visual environment, understand context, plan actions, and execute them through physical interfaces — that's exactly what a robot needs to do in the real world.

The perception-action loop in Friday (screen → understanding → plan → mouse/keyboard) maps directly to the loop a robot needs (cameras → understanding → plan → actuators). Building Friday was a stepping stone toward 20n.

How it works

Built with

Python Computer Vision OCR macOS Accessibility API GPT-4 / Claude PyAutoGUI AppKit