Skip to content

Instantly share code, notes, and snippets.

View josephbolus's full-sized avatar

Joseph Bolus josephbolus

View GitHub Profile
@MattMatheus
MattMatheus / AGENT_QUERY_EXAMPLES.md
Created February 23, 2026 08:02
AthenaMind/AthenaWork + Codex proof-of-concept with agent query prompts

Agent Query Examples

Use these prompts directly against this gist.

1) Executive Summary

"Produce a 10-bullet executive summary of the PoC. Keep only high-confidence, evidence-backed points."

2) Metrics Extraction

"Extract every numeric metric in the report. Return as a table with columns: metric, value, context, and confidence."

@DMontgomery40
DMontgomery40 / CODEX.md
Last active March 16, 2026 13:17
Ralph audit loop: Codex CLI read-only code audit runner

Ralph Audit Agent Instructions (OpenAI Codex)


Safety Notice (Customize)

If this codebase is production, handles money, or touches sensitive data: treat this audit loop as a high-risk operation. Run with least privilege, avoid exporting long-lived credentials in your shell, and keep the agent in read-only mode.


@OmerFarukOruc
OmerFarukOruc / claude.md
Last active March 16, 2026 06:57
AI Agent Workflow Orchestration Guidelines

AI Coding Agent Guidelines (claude.md)

These rules define how an AI coding agent should plan, execute, verify, communicate, and recover when working in a real codebase. Optimize for correctness, minimalism, and developer experience.


Operating Principles (Non-Negotiable)

  • Correctness over cleverness: Prefer boring, readable solutions that are easy to maintain.
  • Smallest change that works: Minimize blast radius; don't refactor adjacent code unless it meaningfully reduces risk or complexity.
@lkarlslund
lkarlslund / glm4.7flash-opencode.md
Created January 21, 2026 15:34
Run GLM 4.7 Flash with OpenCode on RTX 5090 (or other 32GB+ setups)

Run GLM 4.7 Flash with OpenCode on RTX 5090

Grab latest llama.cpp sources and build it:

git clone https ://github.com/ggml-org/llama.cpp/
cmake llama.cpp -B llama.cpp/build -DBUILD_SHARED_LIBS=OFF -DGGML_CUDA=ON
cmake --build llama.cpp/build --config Release -j --clean-first --target llama-cli llama-mtmd-cli llama-server llama-gguf-split
@emschwartz
emschwartz / README.md
Last active March 16, 2026 21:18
The Most Popular Blogs of Hacker News in 2025

This is an OPML version of the HN Popularity Contest results for 2025, for importing into RSS feed readers.

Plug: if you want to find content related to your interests from thousands of obscure blogs and noisy sources like HN Newest, check out Scour. It's a free, personalized content feed I work on where you define your interests in your own words and it ranks content based on how closely related it is to those topics.

@raoulbia-ai
raoulbia-ai / claude-flow-codespace-setup-guide.md
Last active January 27, 2026 11:01
Complete setup guide for Claude-Flow AI agent swarms in GitHub Codespaces. Includes universal dev container config for Docker/Kubernetes/Codespaces, authentication workflow, and community-tested troubleshooting. Based on AI Code Chat community discussions.
@cablej
cablej / default.md
Created June 21, 2025 18:46
Cluely System prompt

<core_identity> You are an assistant called Cluely, developed and created by Cluely, whose sole purpose is to analyze and solve problems asked by the user or shown on the screen. Your responses must be specific, accurate, and actionable. </core_identity>

<general_guidelines>

  • NEVER use meta-phrases (e.g., "let me help you", "I can see that").
  • NEVER summarize unless explicitly requested.
  • NEVER provide unsolicited advice.
  • NEVER refer to "screenshot" or "image" - refer to it as "the screen" if needed.
  • ALWAYS be specific, detailed, and accurate.

Git Commit Assistant

You are an AI assistant tasked with analyzing git changes and creating appropriate commits. Your goal is to identify logical groupings of changes and create conventional commits one at a time.

Step 1: Review Changes

First, analyze the current state:

git status
@dhh
dhh / linux-setup.sh
Last active February 18, 2026 22:50
linux-setup.sh
# THIS LINUX SETUP SCRIPT HAS MORPHED INTO A WHOLE PROJECT: HTTPS://OMAKUB.ORG
# PLEASE CHECKOUT THAT PROJECT INSTEAD OF THIS OUTDATED SETUP SCRIPT.
#
#
# Libraries and infrastructure
sudo apt update -y
sudo apt install -y \
docker.io docker-buildx \
build-essential pkg-config autoconf bison rustc cargo clang \
@scyto
scyto / swarm-updating-shepherd.md
Created May 15, 2022 02:05
use shepherd to update images

Update Swarm Images

i wanted to update swarm images whenever there is a new image available (even at the risk of breaking something)

it seems watchtower can't do this for swarm stacks/services - only one off containers.

I found this great container from https://github.com/djmaze/shepherd

placement considerations

Must be placed on a manager node.