Greg Baugues thumbnail

📝 Greg Baugues Blog

Bringing Claude Code to Life with Custom Sounds: A Fun Introduction to Hooks

When Enthropic announced hooks for Claude Code, it opened a whole new world of possibilities for customizing how Claude interacts with your workflow. Inspired by a playful comment on the Claude AI subreddit—“I’m going to use this to make Claude meow. I’ve always wanted a cat, but I’m allergic”—I decided to dive in and create a unique sound-based notification system for Claude’s various events. The result? Claude meowing, beeping, and even speaking as it runs commands in the background.

Why Use Hooks in Claude Code?

Hooks are a powerful addition that let you run custom commands triggered by different Claude events. These events include:

  • Pre-tool use: Before Claude uses a tool like bash, file editing, or fetching data from the web.
  • Post-tool use: After Claude completes a task.
  • Notification: When Claude is waiting for your approval.
  • Stop: When Claude finishes its task.
  • Pre-compact: Before autocompacting your session history.

By assigning sounds or commands to these events, you get real-time feedback about what Claude is doing, making it easier to understand and control its behavior.

Setting Up Hooks: Where to Start

To set up hooks, you edit your settings.json file within your project. This ensures your hooks are version-controlled and consistent across different work trees. Although you can also configure hooks in user-level settings, keeping them project-specific is a good starting point.

Here’s the gist of the setup:

  • Create a hooks directory inside your .cloud folder.
  • Define a list of hooks in your settings.json, specifying the event type and the command to run.
  • Use a single Python script (hook_handler.py) to handle all hook events. This centralizes your logic and simplifies debugging.

Example snippet from settings.json:

json "hooks": [ { "event": "notification", "command": ["python3", "/path/to/hook_handler.py"] }, { "event": "stop", "command": ["python3", "/path/to/hook_handler.py"] }, { "event": "pre-tool-use", "command": ["python3", "/path/to/hook_handler.py"] } ]

The Magic Behind the Scenes: The Python Hook Handler

The hook_handler.py script receives JSON data from Claude via standard input whenever an event is triggered. This data includes session info, event name, tool name, and tool input, which you can log and analyze to determine what action to take.

I used Claude itself to generate and iteratively refine the Python script, making it as readable and maintainable as possible. The script’s responsibilities include:

  • Logging incoming event data for debugging.
  • Mapping specific events and commands to corresponding sounds or voice notifications.
  • Playing sounds using predefined audio files stored in categorized directories (beeps, voices, meows).

Curating the Sounds: From Beeps to Meows

To keep things fun and engaging, I sourced sounds from Epidemic Sound, a platform I often use for my YouTube videos. They offer a wide range of sound effects, including beeps, cat meows, and even voice actors providing customizable voice clips.

Some tips for working with Epidemic Sound:

  • Sounds often include multiple effects in one file, so use their segment tool to extract just the part you want.
  • You can assign specific sounds to particular actions, for example:
  • A "committing" voice clip when Claude commits code.
  • Different beeps for bash commands, file edits, or pull request creation.
  • Sad meows or playful cat sounds for other events.

This mapping helped me instantly recognize what Claude was doing just by listening.

Lessons Learned and Practical Uses for Hooks

Building this sound notification system was more than just a fun experiment; it was a fantastic way to understand Claude’s inner workings and the power of hooks. Here are some insights and practical applications:

  • Understanding Cloud Code behavior: Assigning sounds to events revealed how often Claude updates to-dos or runs bash commands.
  • Custom safeguards: You can create hooks that prevent dangerous commands like rm -rf from executing accidentally.
  • Automation enforcement: Hooks can ensure tests run before opening pull requests or run linters automatically.
  • Better notifications: Replace or supplement default notifications with customized alerts that better fit your workflow.

Getting Started Yourself

If you want to explore this yourself, check out the full code repository at hih high.ai/hooks. The Python script and example sounds are all there for you to experiment with and customize.

I’d love to hear how you’re using hooks in Claude Code—whether for fun like me or to build more pragmatic workflows. Hooks unlock a new layer of control and creativity, and I hope this post inspires you to dive in!


Summary

  • Hooks allow running custom commands on Claude Code events.
  • Setting hooks in settings.json keeps them project-specific and version-controlled.
  • A Python script can handle multiple hook events for better maintainability.
  • Assigning sounds to events helps understand Claude’s behaviors.
  • Hooks can be used for both fun notifications and practical workflow controls.
  • Check out the full project and share your hook ideas!

Happy coding—and meowing—with Claude Code! đŸ±đŸŽ¶

đŸ“č Video Information:

Title: I used HOOKS to make CLAUDE CODE Meow, Beep, and Talk
Duration: 14:28

How Custom Sounds Help You Understand Claude Code Hooks (and Why You Should Try It!)

When Anthropic announced hooks for Claude Code, one Reddit user joked, “I’m going to use this to make Claude meow. I’ve always wanted a cat, but I’m allergic.” That comment sparked an unexpectedly powerful way to get started with hooks: adding custom sounds to different Claude events. Not only is it fun and whimsical, but it’s also an educational deep-dive into how Claude Code operates under the hood.

Let’s walk through how this sound-based project works, why it’s useful, and how it can inspire more advanced, pragmatic uses of hooks.


Understanding the Power of Claude Code Hooks

Claude Code hooks let you assign custom commands to different events in the coding workflow. These events include things like:

  • Pre-tool use: Before Claude uses a tool (often Bash commands, editing files, reading files, web requests, etc.)
  • Post-tool use: After Claude finishes using a tool.
  • Notifications: When Claude is waiting for user input or approval.
  • Stop: When the current task is completed.
  • Pre-compact: Before Claude auto-compacts session history.

By assigning sounds to each event, you get immediate, sensory feedback as these events are triggered. This helps you see (or hear!) what’s happening in the background, making hooks less abstract and more tangible.


Setting Up Your Project: Where to Configure Hooks

To get started, you’ll edit your settings.json—specifically, the one in your project directory (not your user or local settings). This ensures your hook configuration is committed to your repository and applies across all work trees.

Within your project, create a hooks directory to store all the scripts and sound files. If you eventually want these hooks to work across all projects, you can migrate them to your user settings, but localizing them per project is best for experimentation.


Defining Hooks in settings.json

In your settings.json, hooks are defined as a list, where each hook specifies:

  • Type of event (e.g., pre-tool use, stop, notification)
  • Command to run (in this case, a Python script)

For simplicity and maintainability, it’s best to keep the JSON configuration minimal and put most of the logic inside your Python script. This allows for easier debugging and flexibility.


Building the Python Hook Handler

The Python script acts as the core logic center. Here’s how to approach it:

  1. Log Incoming Data: Whenever Claude triggers a hook, JSON data is piped into your script via standard input. This data contains session information, the event name, the tool being used, and any tool-specific input. Logging this is crucial for understanding what’s happening and for debugging.

  2. Map Events to Sounds: Create directories for different types of sounds (beeps, meows, voices, etc.). You can use sound effect services like Epidemic Sound to download fun beeps, cat meows, or even AI-generated voice snippets.

  3. Assign Sounds to Actions: Either assign random sounds or, more effectively, map specific sounds to specific events or Bash commands. For example, use a “meow” for file edits, a “beep” for notifications, or a British-accented “committing” for git actions.

  4. Optional Patterns: Fine-tune the sound mapping for more granular feedback. For example, distinguish between editing files, reading files, or running specific CLI commands by matching against the command name.


Why Start with Sounds?

Assigning sounds to hooks isn’t just playful—it’s surprisingly educational. You’ll quickly discover:

  • Which events are triggered most often (e.g., how many actions are actually Bash commands)
  • How Claude interacts with your files and tools
  • Opportunities for more advanced hook logic (like intercepting dangerous commands or ensuring tests run before a pull request)

Making abstract processes audible helps demystify Claude’s inner workings and gives you confidence to try more serious customizations.


Beyond Sounds: Unlocking the Full Potential of Hooks

Once you’re comfortable, hooks enable all sorts of productivity and safety improvements:

  • Preventing dangerous commands: Block risky Bash operations like rm -rf before they execute.
  • Running Linters: Automatically trigger code quality checks after edits.
  • Enforcing Test Runs: Ensure tests pass before allowing pull requests.
  • Custom Notifications: Replace unreliable system beeps with tailored sounds or even spoken messages.

Hooks give you deterministic control over Claude Code’s behavior—making your coding environment smarter, safer, and more responsive.


Ready to Try? Resources and Next Steps

You can find all the code and examples from this project at hihigh.ai/hooks.

Whether you want to make Claude meow, beep, or speak in a proper British accent, starting with custom sounds is a delightful way to understand hooks. Once you grasp the basics, you’ll be well-equipped to use hooks for more complex, pragmatic workflows.

What creative uses have you found for Claude Code hooks? Share your ideas and let’s build smarter tools together!

Unlocking Superpowers in Web App Development with AI Coding Workflow: A Deep Dive into Cloud Code and GitHub Integration

Over the past few weeks, I embarked on a journey to build a new web app using a powerful AI-assisted development workflow integrating Cloud Code and GitHub. This approach has truly unlocked new superpowers in my coding process, allowing for rapid, iterative, and well-managed software development. In this blog post, I’ll share the high-level workflow, the rationale behind it, and detailed insights into each phase, including planning, creating, testing, and deploying code with AI assistance.


The AI Coding Workflow at a Glance

The workflow centers around managing work through GitHub issues and leveraging Cloud Code's custom slash commands to process these issues methodically. Here's how it works:

  1. Issue Creation: Every feature or bug fix starts as a GitHub issue.
  2. Planning: Cloud Code uses detailed instructions to break down big issues into small, atomic tasks using scratchpads.
  3. Code Generation: Once planned, Cloud Code writes the code.
  4. Testing: The AI runs test suites and uses Puppeteer to verify UI changes by simulating browser interactions.
  5. Commit & Pull Requests: After successful testing, Cloud Code commits changes and opens a pull request (PR) for review.
  6. Review & Integration: PRs are reviewed either by me or by Claude Code itself, with continuous integration (CI) systems running tests on every commit.
  7. Cycle Repeats: After clearing the context window with /clear, the system moves to the next issue.

This cycle aligns closely with the classic software development life cycle (SDLC) phases: plan, create, test, and deploy.


Why Adopt This Workflow?

Even with advanced AI coding assistants, writing code is only one part of delivering quality software. The software industry has long recognized that managing the entire lifecycle — from planning to deployment — is crucial. The workflow I use is heavily based on GitHub Flow, a proven process designed for small teams and perfectly suited for a human paired with an AI assistant.

This structured approach ensures that AI-generated code integrates well with existing systems, remains maintainable, and is thoroughly tested before deployment.


Detailed Breakdown of the Workflow

1. Creating and Refining GitHub Issues

My process starts with dictation and transforming ideas into a requirements document using Claude. From there, Claude helps generate GitHub issues.

Key Lessons:
- Initially, I created 30–40 issues right away, which was overly optimistic.
- The issues needed to be granular, specific, and atomic to set the AI up for success.
- I had to take on a managerial role: refining issues, prioritizing tasks, and clarifying specifications.
- This phase is critical and involves a lot of human involvement to ensure the AI understands the scope and requirements precisely.

2. Setting Up the Foundation

Before AI can be effective, a solid foundation is essential:

  • Test Suite and Continuous Integration (CI): I prioritized setting up a robust test suite using Rails’ integrated testing framework and connected it with GitHub Actions for automatic testing on commits.
  • Puppeteer for UI Testing: Puppeteer allows Cloud Code to interact with the app in a real browser, clicking buttons and filling forms to validate UI changes, which is surprisingly satisfying to watch.
  • The foundation enables rapid, confident feature development with automated safety nets.

3. Planning with Custom Slash Commands

I created a custom slash command that accepts an issue number and guides Cloud Code through four stages: plan, create code, test, and deploy.

Planning Highlights:
- Cloud Code uses the GitHub CLI to fetch issue details.
- It searches scratchpads and previous PRs to gather context and avoid duplicate work.
- I utilize Anthropic’s “think harder” prompt to encourage deeper reasoning.
- The AI breaks down issues into manageable subtasks and documents them.

4. Create, Test, and Deploy

Creating and Testing:
- Claude writes the code and commits changes.
- While some teams prefer manually reviewing all commits, I’ve allowed Claude to handle commits and focused on reviewing pull requests.
- Tests serve as a confidence booster, ensuring changes don’t break existing features.
- I encourage thorough code reviews and sometimes have Claude perform PR reviews in the style of coding experts like Sandy Mets, which surfaces maintainability improvements.

Deploying:
- I deploy using Render, which automatically updates the app on merges to the main branch.
- Merging a PR essentially equates to a production deployment in this workflow.


Human vs. AI Responsibilities

  • Human Focus: I spend most of my time during the planning phase, crafting precise issues and refining specs.
  • AI Focus: Claude Code handles much of the coding, testing, and deploying, with occasional human code reviews.
  • This balance leverages the strengths of both parties effectively.

Managing Context with /clear

After merging a PR, I use the /clear command to wipe Cloud Code’s context window, ensuring that each issue is tackled fresh without previous context pollution. This keeps token usage efficient and forces issues to be self-contained with all necessary information.


Using Claude via GitHub Actions: Pros and Cons

Anthropic introduced Claude integration directly in GitHub Actions, allowing AI to work on GitHub issues via tagging. However:

  • This method is metered and can incur additional API costs, even on unlimited plans.
  • I found better results and more insightful interactions using Cloud Code in the console.
  • GitHub Actions are ideal for small tweaks or copy edits but less practical for large, complex changes.

Running Parallel Agents with Work Trees

Work trees allow running multiple AI agents in parallel on different branches, akin to multitabling in online poker. However, I encountered some challenges:

  • Permission re-approval was required for each new session, leading to time-consuming babysitting.
  • Managing multiple work trees increased merge conflicts and overhead.
  • Due to the early stage of my app and iterative nature of development, parallel work had limited benefit.
  • For now, a single Cloud Code instance suffices for my needs.

Final Thoughts

This AI-powered workflow combining Cloud Code, GitHub issues, custom slash commands, automated testing, and CI has transformed how I build web apps. It requires a shift in mindset — embracing a managerial role in planning while delegating coding and testing to AI. The process is iterative and involves continuous refinement of issues and commands, but it holds immense promise for boosting productivity and code quality.

If you’re interested in learning more, I recommend checking out my video on Claude Code pro tips for deeper insights into maximizing AI coding assistants.


Further Reading


Harness the power of AI coding assistants and structured workflows to supercharge your software development today!

Mastering Claude Code: Pro Tips to Supercharge Your Development Workflow

As developers, we're always on the lookout for tools that not only simplify our coding process but also enhance productivity. Over the past few months, Claude Code has emerged as a game-changing assistant for writing code. Developed by Boris Churnney and the Enthropic team, Claude Code offers a versatile, AI-powered command-line interface (CLI) tool that integrates seamlessly into your development environment.

In this blog post, we'll explore some essential pro tips to help you harness the full potential of Claude Code, based on insights from Boris’s comprehensive guide and real-world experience.


1. Command-Line Interface (CLI) Basics

Claude Code operates primarily as a CLI tool, familiar to anyone accustomed to bash-based environments. You can:

  • Pass command-line arguments that execute on startup.
  • Use the -p flag to run Claude Code in headless mode.
  • Chain Claude Code with other CLI tools or pipe data into it.
  • Run multiple instances simultaneously.
  • Even have Claude Code launch sub-instances of itself, useful for complex, multi-agent workflows.

This flexibility means Claude Code fits naturally into existing shell-based workflows.


2. Working with Images and Screenshots

Claude Code supports image inputs, which can be a powerful asset for UI/UX development:

  • On macOS, drag and drop images directly into the terminal.
  • Use Shift + Command + Control + 4 to take a screenshot, then paste it with Control + V (note: it's Control, not Command).
  • Leverage images for two key use cases:
  • Mockups: Paste UI mockups into Claude and ask it to generate corresponding interface code.
  • Feedback Loop: After Claude builds something, take a screenshot of the result, feed it back into Claude, and request iterative improvements.

For automated screenshots, you can integrate Puppeteer — a headless browser tool — with Claude Code, enabling it to open your app, take screenshots, and save them locally without manual intervention.


3. Harnessing MCP Servers and Clients

Claude Code functions both as an MCP (Multi-Channel Protocol) server and client, allowing it to interact with various data sources and services:

  • Use MCP servers like Postgres to directly connect Claude Code to your databases.
  • Employ MCP wrappers around APIs to provide Claude with up-to-date external information.
  • Some companies (e.g., Cloudflare) expose documentation via MCP servers, enabling Claude to fetch and use the latest docs during development.
  • Claude can also fetch URLs you provide, extracting knowledge from online resources to inform coding tasks. For example, referencing official game rules to build accurate game logic.

4. The Power of claude.md

One of the most critical features is the claude.md file—a persistent prompt loaded with every interaction Claude has with your project. It typically contains:

  • Project-specific instructions.
  • Coding style and linting guidelines.
  • Test commands.
  • Repository etiquette.

You can auto-generate this file by typing /init when launching Claude in a project directory; it scans and summarizes the project structure for you. Some tips for managing claude.md files:

  • Use a global claude.md in your home directory for universal rules.
  • Add project- or subdirectory-specific files to cover nuanced needs.
  • Regularly refactor to keep prompts concise and specific—avoid redundant or excessive information.
  • Use Anthropic’s prompt optimizer tool to fine-tune your prompts for better AI performance.

5. Slash Commands for Custom Shortcuts

Slash commands are customizable prompt templates stored in the cloud/commands folder. They allow you to:

  • Create shortcuts for common tasks like refactoring, linting, or reviewing pull requests.
  • Pass command-line arguments to dynamically fill in prompt templates.

This feature streamlines repetitive tasks and integrates tightly with your workflow.


6. UI Tips for Efficient Sessions

To maximize session efficiency:

  • Use Tab to autocomplete files and directories, helping Claude focus on specific code areas.
  • Don’t hesitate to interrupt Claude by pressing Escape if it goes off-track.
  • Use the undo feature (Escape followed by an undo command) to revert unwanted changes quickly.

Interrupting Claude early prevents wasted effort and keeps your project clean.


7. Version Control Integration

One common pitfall is letting Claude make too many ambitious changes that break your codebase. To avoid this:

  • Use version control diligently alongside Claude Code.
  • After every significant change, commit with messages generated by Claude—these tend to be impressively clear and descriptive.
  • Revert often if needed and clear Claude’s conversation history to reset context before trying again.
  • Install the GitHub CLI to enable Claude to manage GitHub interactions like filing pull requests and conducting code reviews.
  • Alternatively, you can use the GitHub MCP server for GitHub integration without installing CLI tools.

8. Managing Context and Token Limits

Claude Code sessions have token limits and auto-compacting behaviors that can affect performance:

  • Monitor the auto-compacting indicator to know when Claude will reduce memory.
  • Compact context at natural breakpoints, such as after completing a task or making a commit, to free up tokens.
  • Consider clearing context entirely when appropriate to work with fresh memory.
  • Use scratchpads or GitHub issues to plan and organize Claude’s work outside the main conversation.
  • For cost-conscious users paying per token, aggressively manage context and use external memory tools.

9. Controlling Costs Effectively

Claude Code can be expensive, but there are strategies to manage costs:

  • Use open telemetry support to track usage metrics. For example, integrate with monitoring tools like DataDog to visualize token consumption across your team.
  • Consider upgrading to Claude Max plans ($100 or $200/month), which bundle tokens and offer better value.
  • Monitor usage closely and use cost-saving practices like context management and external memory.

Final Thoughts

Claude Code is a powerful tool that, when used with these pro tips, can revolutionize your coding workflow. From seamless CLI integration and image handling to smart prompt management and version control synergy, the possibilities are vast.

For an in-depth dive, check out Boris Churnney’s original post and related resources linked below. Whether you’re building games, APIs, or complex applications, Claude Code is poised to be your indispensable coding partner.


Resources:

Happy coding! 🚀


Written by Greg, Developer and Claude Code Enthusiast

How to Build a Real-Time Web Research Bot Using Anthropic’s New Web Search API

In the rapidly evolving world of AI, staying updated with the latest information is crucial. Traditional AI models often rely on static training data, which can be months or even years old. That’s why Anthropic’s recent announcement of web search capabilities via their cloud API is a game changer. It enables developers to build AI-powered research bots that fetch real-time data from the web — without relying on external scraping tools or additional servers.

In this blog post, we’ll explore how to leverage Anthropic’s new web search API to build a real-time research assistant in Python. We’ll cover a demo example, key implementation details, useful options, and pricing considerations.


What’s New with Anthropic’s Web Search API?

Before this update, building a research bot that accessed up-to-date information meant integrating third-party tools or managing your own scraping infrastructure. Anthropic’s web search API simplifies this by allowing direct querying of the web through their cloud API, keeping everything streamlined in one place.

For example, imagine wanting to know about a breaking news event that happened just hours ago — such as the recent selection of the first American Pope (recorded May 8th, 2025). Since this information isn’t part of the AI’s training data, it needs to perform a live web search to generate an accurate and current report.


Building Your First Web Search Request: A “Hello World” Example

Getting started is straightforward. Here’s an overview of the basic steps using the Anthropic Python client:

  1. Set your Anthropic API Key as an environment variable. This allows the client to authenticate requests seamlessly.

  2. Instantiate the Anthropic client in Python.

  3. Send a message to the model with your question. For example, “Who is the new pope?”

  4. Add the tools parameter with web_search enabled. This tells the model to access live web data.

Here’s a snippet summarizing this:

```python
from anthropic import Anthropic, HUMAN_PROMPT, AI_PROMPT

client = Anthropic()

response = client.completions.create(
model="claude-3",
messages=[{"role": "user", "content": "Who is the new pope?"}],
tools=["web_search"], # Enable web search tool
max_tokens=1000
)

print(response.choices[0].message.content)
```

Without web search enabled, the model might respond with outdated information (e.g., “As of the last update, it was Pope Francis”). But with web search active, it fetches the latest details, complete with citations from recent news sources.


Understanding the Web Search Response

The response from the API when using web search is richer and more complex than standard completions. It includes:

  • Initial text from the model indicating it is performing a search.
  • Tool usage details showing search queries and pages found.
  • Encrypted content blocks representing scraped snippets (Anthropic encrypts these to avoid direct scraping exposure).
  • Summarized text with citations — a distilled answer referencing URLs, page titles, and quoted text snippets.

Parsing this response can be a bit challenging. The Python client lets you convert the response to a dictionary or JSON format for easier inspection.

For example, you can iterate over the response’s message blocks, extract the main text, and gather citations like URLs and titles. This lets you assemble a report with clickable sources, ideal for building research assistants or automated reporting tools.


Improving Performance with Streaming

Waiting 20 seconds for a full response might be too slow for some applications. Anthropic supports streaming responses through an asynchronous client.

Using the async client, you can receive partial results as they become available and display them in real-time, improving user experience in chatbots or interactive assistants.


Customizing Search Domains: Allowed and Blocked Domains

Anthropic’s API offers parameters to restrict searches to certain domains (allowed_domains) or exclude others (blocked_domains). For example, if you only want information from Reuters, you can specify that in your request:

python tools=[{"name": "web_search", "allowed_domains": ["reuters.com"]}]

However, note that some domains are off-limits due to scraping restrictions (e.g., BBC.co.uk, Reddit). Trying to search those will result in an error.

You can use either allowed_domains or blocked_domains in a single request, but not both simultaneously.


Pricing Overview: How Much Does It Cost?

Anthropic’s web search API pricing stands out as very competitive:

  • $10 per 1,000 searches plus token usage for the normal API calls.
  • Compared to OpenAI’s web search pricing of $30 to $50 per 10,000 calls, Anthropic’s is more affordable.

The pricing difference might be due to different search context sizes or optimizations, but it makes Anthropic a cost-effective choice for integrating live web data.


Wrapping Up

Anthropic’s new web search API opens exciting possibilities for developers building AI applications that require fresh, real-time data from the web. With simple integration, customizable domain filters, streaming support, and competitive pricing, it’s a compelling option for research bots, news aggregators, and knowledge assistants.

If you want to try this out yourself, check out the Anthropic Python client, set your API key, and start experimenting with live web queries today!


Useful Links


Author: Greg
Recorded May 8th, 2025

Feel free to leave comments or questions below if you want help building your own web research bot!

How to Build a Remote MCP Server with Python and FastMCP: A Step-by-Step Guide

In the rapidly evolving world of AI and large language models (LLMs), new tools and integrations continue to push the boundaries of what’s possible. One of the most exciting recent developments is Enthropic’s announcement of remote MCP (Model Control Protocol) server support within Claude, an AI assistant platform. This breakthrough means that users can now connect to MCP servers simply by providing a URL—no complex local setup or developer-level skills required.

In this blog post, we’ll explore what remote MCP servers are, why they matter, and how you can build your own using Python and the FastMCP framework. Whether you’re a developer or an AI enthusiast, this guide will give you the knowledge and tools to create powerful, accessible AI integrations.


What is an MCP Server?

An MCP server serves as a bridge between large language models and external tools or data sources. This enables LLMs like Claude to perform tasks or fetch real-time information beyond their static knowledge base. Traditionally, MCP servers were local setups requiring developer expertise to configure and maintain, limiting their accessibility.


Why Remote MCP Servers are a Game-Changer

Remote MCP servers allow users to connect to MCP servers hosted anywhere via a simple URL. This innovation dramatically lowers the barrier to entry, making it easier for less technical users to enhance their AI assistants with custom tools. For example, a remote MCP server can provide up-to-the-minute data like the current time, weather, or stock prices—capabilities that standard LLMs often lack due to knowledge cutoffs.

Claude’s new integration support means you can now:

  • Add MCP servers by entering a URL in Claude’s settings.
  • Automatically discover available tools and their parameters.
  • Seamlessly invoke those tools during conversations with Claude.

This marks a significant step toward more interactive, capable AI assistants.


Demo: Adding a Current Time Tool to Claude

To illustrate the power of remote MCP servers, here’s a quick example:

  1. Problem: Claude cannot provide the current time because its knowledge is frozen at a cutoff date.
  2. Solution: Create a remote MCP server that returns the current date and time.
  3. Integration: Add the MCP server URL to Claude’s settings under “Integrations.”
  4. Usage: Ask Claude, “What is the current time?” Claude recognizes it has access to the time tool, invokes it with the correct parameters (like time zone), and returns an accurate, up-to-date answer.

This simple enhancement vastly improves Claude’s utility for real-world tasks.


Building Your Own Remote MCP Server in Python with FastMCP

Step 1: Set Up Your Environment

Begin by visiting gomcp.com or the FastMCP GitHub repository for documentation and code examples.

Install FastMCP via pip:

bash pip install fastmcp

Step 2: Create the MCP Server

Here’s a basic MCP server script that provides the current date and time:

```python
from fastmcp import MCPServer
from datetime import datetime
import pytz

server = MCPServer(name="DateTime Server", instructions="Provides the current date and time.")

@server.tool(name="current_datetime", description="Returns the current date and time given a timezone.")
def current_datetime(time_zone: str = "UTC") -> str:
try:
tz = pytz.timezone(time_zone)
now = datetime.now(tz)
return now.strftime("%Y-%m-%d %H:%M:%S %Z")
except Exception as e:
return f"Error: {str(e)}"

if name == "main":
server.run()
```

Step 3: Run Locally and Test with MCP Inspector

FastMCP offers an MCP Inspector tool for debugging and testing your server:

bash npx fastmcp inspector my_server.py

This GUI lets you invoke your tools and view responses directly, providing a deterministic way to debug interactions with your MCP server.

Step 4: Deploy as a Remote MCP Server

To make your MCP server accessible remotely, you need to use a transport protocol suitable for networking.

  • The default stdin/stdout transport works locally.
  • For remote access, use Server-Sent Events (SSE) transport or, soon, the more efficient streamable HTTP transport (currently in development).

Modify your server code to use SSE transport and deploy it on a cloud platform such as Render.com. Assign a custom domain (e.g., datetime.yourdomain.ai) pointing to your deployment.

Once deployed, add your server URL in Claude’s integrations, and it will be ready for use remotely.


The Future of MCP Servers

The adoption of remote MCP servers is poised to explode, as they become far easier to create and integrate with AI assistants like Claude. This will likely spur more companies to launch their own MCP servers, offering a diverse ecosystem of tools accessible via URLs.

For developers, this is an exciting time to dive into MCP and FastMCP development. Even those with limited coding experience can now build meaningful AI enhancements quickly.


Final Thoughts

  • MCP servers empower AI models to access real-time data and perform specialized tasks.
  • Remote MCP servers eliminate the technical hurdles of local setup.
  • FastMCP and Python provide a straightforward path to building your own MCP servers.
  • Claude’s new integrations make adding MCP servers as simple as entering a URL.
  • The future will see more widespread adoption and innovation in MCP technology.

If you want to stay ahead in AI tooling, start experimenting with FastMCP today. Build your own remote MCP server, connect it to Claude, and unlock new capabilities for your AI assistant.

Happy coding!


Resources


Have questions or want to share your MCP server projects? Drop a comment below or connect with me on Twitter!

Why Cloud Code Outshines OpenAI’s Codeex: A Developer’s Perspective

Hey there! I’m Greg, a developer who has spent hundreds of dollars experimenting with AI-powered coding assistants over the past few months. Lately, I’ve made Cloud Code my go-to coding agent, especially when starting new projects or navigating large, older codebases. With the recent launch of OpenAI’s Codeex, I was eager to give it a shot and pit it against Cloud Code in a head-to-head comparison. Spoiler alert: Codeex fell short in several key areas.

In this blog post, I’ll share my firsthand experience with both tools, highlight what OpenAI needs to improve in Codeex, and explain why developer experience is crucial for AI coding assistants to truly shine.


First Impressions Matter: The Developer Experience

Right from the start, Codeex’s developer experience felt frustrating. Although I have a Tier 5 OpenAI account—which is supposed to grant access to the latest GPT-4 models—Codeex informed me that GPT-4 was unavailable. Instead of gracefully falling back to a supported model, the system simply failed when I tried to use GPT-4 Mini.

To make matters worse, the interface for switching models was confusing. I had to use a /help command to discover a /model command with a list of options ranging from GPT-3.5 to Babbage and even DALL·E (an image generation model that doesn’t belong here). Most of these options didn’t work with the product, so I was left guessing which model to pick. This was a baffling experience—why show options that don’t actually work? It felt like a basic user experience bug that should have been caught during testing.

For developers, the first interaction with a tool should be smooth and intuitive—no guesswork, no dead ends. Sadly, Codeex made me jump through unnecessary hoops just to get started.


API Key Management: A Security and Usability Concern

Cloud Code shines in how it manages API keys. It securely authenticates you via OAuth, then automatically stores your API key in a local config file. This seamless process means you can focus on coding without worrying about environment variables or security risks.

Codeex, on the other hand, expects you to manually set your OpenAI API key as a global environment variable or in a .env file. This approach has several drawbacks:

  • Security Risk: Having a global API key in your environment exposes it to any local script or app, increasing the chances of accidental leaks.
  • Lack of Separation: You can’t easily dedicate a separate API key for Codeex usage, which complicates cost tracking and project management.
  • Inconvenience: Managing environment variables across multiple projects can become tedious.

Cloud Code’s approach is more secure, user-friendly, and better suited for developers juggling multiple projects.


Cost Management: Transparency and Control Matter

AI coding assistants can get expensive, and managing usage costs is critical. Cloud Code offers helpful features to keep your spending in check:

  • /cost Command: View your session’s spend anytime.
  • /compact Command: Summarize and compress chat history to reduce token usage and lower costs.

Codeex lacks these features entirely. There is no way to check how much you’ve spent during a session or to compact conversation history to reduce billing. This opacity can lead to unpleasant surprises on your bill and makes cost management stressful.


Project Context Awareness: Smarter by Design

One of Cloud Code’s standout features is its ability to scan your project directory on startup, building an understanding of your codebase. It lets you save this context into a claw.md file, so it doesn’t have to reanalyze your project every time you launch the tool. You can even specify project-specific preferences and coding conventions.

Codeex, by contrast, offers zero context-awareness upon startup. It simply opens a chat window with your chosen model and waits for input. This puts the burden on the developer to manually introduce project context, which is inefficient and time-consuming.

For a coding agent, understanding your existing codebase from the get-go is a game-changer that Codeex currently misses.


User Interface: Polished vs. Minimal Viable

Cloud Code’s command-line interface (CLI) is thoughtfully designed with clear separation between input and output areas, syntax highlighting, and even color schemes optimized for color-blind users. The UI feels intentional, refined, and comfortable for extended use.

Codeex feels like a bare minimum implementation. Its output logs scroll continuously without clear visual breaks, it lacks syntax highlighting, and it provides only rudimentary feedback like elapsed wait time messages. This minimalism contributes to a frustrating user experience.


Stability and Reliability: Crashes Are a Dealbreaker

Cloud Code has never crashed on me. Codeex, unfortunately, has crashed multiple times, especially when switching models. Each crash means reconfiguring preferences and losing all previous session context—a major productivity killer.

Reliability is table stakes for developer tools, and Codeex’s instability makes it feel unready for prime time.


Advanced Features: MCP Server Integration

Cloud Code supports adding MCP (Machine Control Protocol) servers, enabling advanced use cases like controlling a browser via Puppeteer to close the feedback loop by viewing changes in real-time. This kind of extensibility greatly expands what you can do with the tool.

Codeex currently lacks support for MCP servers, limiting its potential for power users.


The Origin Story: Why Polished Tools Matter

During a recent Cloud Code webinar, I learned that Cloud Code began as an internal tool at Anthropic. It gained traction within the company, prompting the team to polish it extensively before releasing it publicly. This internal usage ensured a high-quality, battle-tested product.

In contrast, Codeex feels like it was rushed to market with minimal internal adoption and testing. With just a couple of weeks of internal use and intentional polish, Codeex could improve dramatically.


Final Thoughts: Potential vs. Reality

I have not even touched on the core coding ability or problem-solving skills of Codeex’s models, such as GPT-4 Mini plus codecs. It’s possible that, once the bugs and UX issues are ironed out, Codeex could outperform Cloud Code at a lower cost.

But right now, the frustrating user experience, instability, poor key management, and lack of cost transparency prevent me from fully engaging with Codeex. A well-designed developer experience isn’t just a nice-to-have; it’s essential to unlocking the true power of AI coding assistants.


What OpenAI Needs to Do to Bring Codeex Up to Par

  1. Graceful Model Fallback: Automatically switch to a supported model if the default is unavailable.
  2. Clear and Accurate Model List: Only show models that actually work with the product.
  3. Secure and Convenient API Key Management: Implement OAuth or a dedicated API key setup for the tool.
  4. Cost Transparency: Add commands or UI elements to track session spending and manage token usage.
  5. Project Context Awareness: Automatically scan and remember project details to save time and costs.
  6. Stable, Polished UI: Improve the CLI interface with clear input/output zones, syntax highlighting, and accessibility options.
  7. Reliability: Fix crash bugs to ensure smooth, uninterrupted workflows.
  8. Advanced Feature Support: Enable MCP servers or equivalent extensibility to boost functionality.

Conclusion

AI coding assistants hold incredible promise to revolutionize software development, but only if they respect developers’ time, security, and workflows. Cloud Code exemplifies how thoughtful design and polish can make a tool truly empowering.

OpenAI’s Codeex has potential, but it needs significant improvements in developer experience and stability before it can compete. I look forward to seeing how it evolves and hope these insights help guide its growth.


Thanks for reading! If you’ve tried either Cloud Code or Codeex, I’d love to hear about your experiences in the comments below. Happy coding!

How to Supercharge Your Writing Workflow with Claude Desktop, MCP Servers, and Obsidian

Hello readers! Today, I want to share an exciting workflow that combines three powerful tools—Claude Desktop, MCP servers, and Obsidian—to dramatically boost your productivity with minimal effort. Whether you’re a writer, researcher, or knowledge worker, this setup can transform how you create and manage content. Let’s dive in!

What Are These Tools?

Claude Desktop
Claude Desktop is the downloadable app version of Claude, an AI assistant developed by Anthropic. Unlike the web version, Claude Desktop can act as an MCP client, meaning it can connect to MCP servers to access additional contextual information and tools.

MCP Servers (Model Context Protocol Servers)
MCP, or Model Context Protocol, is an open standard designed to enrich AI models with external context such as databases or programmatic functions. MCP servers provide this context, allowing the AI to perform tasks like reading and writing files, accessing directories, or calling specific functions. One simple example is the file system MCP server, which lets Claude interact directly with files on your computer.

Obsidian
Obsidian is a popular note-taking app built around a concept called a “vault,” which is essentially a folder containing markdown files. It’s highly flexible and perfect for personal knowledge management.

How Do These Tools Work Together?

By running Claude Desktop with an MCP server that grants file system access, you enable Claude to read, edit, and create files inside your Obsidian vault. This integration allows the AI to interact directly with your notes, making it a powerful assistant for writing and knowledge work.

A Real-World Example: Writing an Interview Response

Last week, after speaking at a conference, I was asked to answer 10 interview questions via email. Normally, answering such questions can be a time-consuming task prone to procrastination. So, I decided to try out this new setup.

  1. Preparation: I created a new note in Obsidian and pasted all 10 questions there.
  2. Interaction: I asked Claude Desktop if it could access the note—indeed, it could.
  3. Interview Simulation: Claude then asked me each question one at a time, simulating an interview.
  4. Dictation: Instead of typing responses, I used a speech-to-text tool called Super Whisper to dictate my answers, allowing me to speak freely without the pressure of editing on the fly.
  5. Live Updates: After each answer, I instructed Claude to update the original Obsidian note with my response. This worked smoothly, creating a living document that grew with each answer.
  6. Editing and Refinement: Once all answers were in, I asked Claude to review and clean up the responses for clarity and brevity, outputting the edits into a new document to safeguard the original text.
  7. Final Polish: I then did a traditional manual edit inside Obsidian to add finishing touches, remove redundancies, and improve flow.

The result? A polished 2,000-word response completed in about 90 minutes—a task that would have otherwise taken much longer and felt more arduous.

Why This Workflow Works So Well

  • Focused Interaction: Being asked one question at a time by Claude reduced overwhelm and helped me focus.
  • Separation of Writing and Editing: Dictating first and editing later prevented me from getting stuck in perfectionism during the initial draft.
  • Seamless File Integration: The ability for Claude to read and write directly to my Obsidian notes streamlined the entire process.
  • Time Savings: Automating the first editing pass saved a huge amount of time, leaving only the final polish for me to do.

Future Possibilities

This experiment sparked an idea: what if Claude could connect to a Git MCP server? Adding Git version control to Obsidian notes managed by Claude could unlock even more powerful workflows, like automatic versioning and rollback of your knowledge base edits. This is definitely a project I’m marking for the future.

Final Thoughts

If you’re looking to accelerate your writing or note-taking tasks, consider combining Claude Desktop with MCP servers and Obsidian. This setup allows you to leverage AI not just as a conversational partner but as an integrated tool that directly manipulates your files and notes. The potential to save time and increase output quality is significant.

Give it a try, and you might find yourself knocking out projects faster and with less friction—just like I did with my interview response!


Tools Mentioned:

Feel free to share your experiences or ask questions about setting up this workflow in the comments below!

Cursor vs. Claude Code: Which AI Coding Agent Reigns Supreme?

The emergence of AI-powered coding agents has been one of the most exciting developments for software developers recently. Just this past week, two major players—Cursor and Anthropic’s Claude Code—launched their coding agents simultaneously. Intrigued by which tool might better serve developers, I decided to put both to the test on a real-world Rails application running in production. Here’s a detailed breakdown of my experience, comparing their user experience, code quality, cost, autonomy, and integration with the software development lifecycle.


The Test Setup: A Real Rails App with Complex Needs

My project is a Rails app acting as an email "roaster" for GPTs—essentially bots that process and respond to emails with unique personalities. The codebase is moderately complex and had been untouched for nine months, making it perfect for testing AI assistance on:

  1. Cleaning up test warnings and updating gem dependencies.
  2. Replacing LangChain calls with direct OpenAI API usage.
  3. Adding support for Anthropic’s API.

Both agents used the same underlying model—Claude 3.7 Sonnet—to keep the comparison fair.


User Experience (UX): Terminal Simplicity vs. IDE Integration

Cursor:
Cursor’s agent is integrated into a fully featured IDE and has recently made the agent the primary way to interact with the code. While this offers powerful context and control, I found the interface occasionally clunky—multiple “accept” buttons, cramped terminal panes, and confusing prompts requiring manual clicks. The file editor pane often felt unnecessarily large given that I rarely needed to manually tweak files mid-action.

Claude Code:
Claude Code operates as a CLI tool right in the terminal. You run commands from your project root, and it prompts you with simple yes/no questions to confirm each action. This single-pane approach felt clean, intuitive, and perfectly suited for delegating control to the agent. The lack of a GUI was a non-issue given the agent’s autonomy.

Winner: Claude Code for its streamlined, efficient command-line interaction.


Code Quality and Capability: Documentation Search Matters

Both agents produced similar code given the same model, but Cursor’s ability to search the web for documentation gave it a notable edge. When adding Anthropic support, Claude Code struggled with API syntax and ultimately wrote its own HTTP implementation. Cursor, however, seamlessly referenced web docs to get the calls right, rescuing itself from dead ends.

Winner: Cursor, thanks to its web search integration.


Cost: Subscription vs. Metered Pricing

  • Claude Code: Approximately $8 for 90 minutes of work on these tasks. While reasonable, costs could add up quickly for frequent use.
  • Cursor: $20/month subscription includes 500 premium model requests; I used less than 10% of that for this exercise, roughly costing $2.

Winner: Cursor, offering more usage for less money and a simpler subscription pricing model.


Autonomy: Earning Trust with Incremental Permissions

Claude Code shines here with a granular permission model. Initially, it asks for approval on commands; after repeated approvals, it earns trust to perform actions autonomously. By the end of my session, it was acting independently with minimal prompts.

Cursor, in contrast, lacks this “earned trust” feature. It repeatedly asks for confirmation without a way to grant blanket permissions. Given the nature of coding agents, I believe this is a feature Cursor should adopt soon.

Winner: Claude Code for smarter incremental permissioning.


Integration with Software Development Lifecycle

I emphasize test-driven development (TDD) and version control (Git), so how each agent handled these was crucial.

  • Claude Code: Excellent at generating and running tests before coding features, ensuring quality. Its commit messages were detailed and professional—better than any I’ve written myself. Being a CLI tool, it felt natural coordinating commands and output.

  • Cursor: While it offers a nice Git UI within the IDE and can autogenerate commit messages, these were more generic and less informative. Its handling of test outputs in a small terminal pane felt awkward.

Winner: Claude Code, for superior test and version control workflow integration.


Final Verdict: Use Both, But Lean Towards Claude Code—for Now

Both agents completed all three complex tasks successfully—a testament to how far AI coding assistants have come. It’s remarkable to see agents not only write code but also tests and meaningful commit messages that improve project maintainability.

That said, this is not a binary choice. I recommend developers use both tools in tandem:

  • Use Cursor for day-to-day coding within your IDE, benefiting from its subscription model and web documentation search.
  • Use Claude Code for command-line driven tasks that require incremental permissions, superior test integration, and detailed commit management.

For now, I personally prefer Claude Code for its user experience, autonomy model, and lifecycle integration. But Cursor’s rapid iteration pace means it will likely close these gaps soon.


Takeaway for Developers

If you’re a software developer curious about AI coding agents:

  • Get the $20/month Cursor subscription to familiarize yourself with agent-assisted coding.
  • Experiment with Claude Code in your terminal to experience granular control and trust-building autonomy.
  • Use both to balance cost, control, and convenience.
  • Embrace AI coding agents as powerful collaborators that can help you break through stalled projects and increase productivity.

The future of software development is here—and these AI coding agents are just getting started.


Have you tried Cursor or Claude Code? Share your experiences and thoughts in the comments below!