Add Python-based slack-cli package that reads Slack messages from local
Chromium IndexedDB cache. Includes pinned dependencies for dfindexeddb,
python-snappy, and zstd. Enable on mac-va-mbp-work home config.
Add per-item blur_radius=30 with background.clip=0.5 to create a
frosted-glass effect on each component pill while keeping the bar
itself fully transparent. Bump bg alpha from 0x20 to 0x30 so the
blur is visible through the semi-transparent backgrounds.
- Add new address-gh-review skill with SKILL.md and gh_review.sh script
for fetching and addressing unresolved GitHub PR review comments
- Update AGENTS.md comment style to require Title Case for block titles
- Add packages/nunc derivation built from Gitea with swiftPackages
- Add home module reichard.services.nunc with configurable settings
and launchd agent for macOS
- Enable nunc service on mac-va-mbp-work
- Tweak mac-va-mbp-work defaults: hide menu bar, disable auto
brightness, hide Finder desktop icons
Add a new shell function that uses an LLM API to generate shell commands
from natural language queries. The command displays the generated command
for user confirmation before execution and adds it to history.
- Generates shell commands from natural language via LLM
- Displays command for user review before execution
- Automatically adds executed commands to bash history
- Uses qwen3-coder-next-80b-instruct model
- Extract git_cmd and git_cmd_list for reusable git command execution
- Add yank helper for consistent clipboard copying with notifications
- Enhance get_git_info to include abs_path and file_dir
- Add get_blame_info to retrieve detailed blame information with file hash
- Implement copy_commit_url for copying direct commit URLs
- Convert keymaps to use function references
New: <Leader>gC copies GitHub commit URL with file hash fragment
- Upgrade llama-cpp from version 8196 to 8229
- Remove GPT OSS CSEC (20B) - Thinking model config
- Remove Qwen3 Next (80B) - Instruct model config
- Increment -ncmoe parameter for Qwen3 Coder Next (80B) model
- Bump llama-cpp from version 8157 to 8196
- Bump llama-swap from version 192 to 197
- Switch default assistant model from qwen3-coder-next-80b to qwen3.5-27b-thinking
- Remove glm-4-32b-instruct model configuration
- Update qwen3.5-27b-thinking config:
- Use bartowski quantization (IQ4_XS) instead of unsloth
- Increase context window from 131k to 196k
- Add cache type settings (q8_0) and CUDA device
- Add 1password-cli to home-manager programs
- Fix typo: 'dispay' -> 'display' in llm-config.lua
- Add Nix package for @anthropic-ai/claude-code (v2.1.59)
- Add package-lock.json with optional dependencies for all platforms
- Include update script for automated version updates
- Add module integration with custom package reference
- Rename claude shell alias to claude-custom
Refs: https://github.com/anthropics/claude-code
- Enable NSWindowShouldDragOnGesture for drag gestures
- Force 24-hour time format via AppleICUForce24HourTime
- Add Touch ID authentication for sudo commands
- Remove outdated comment about Three Finger Drag
- Add Qwen3.5-35B-A3B and Qwen3.5-27B thinking model configs
- Remove deprecated Qwen2.5-Coder-7B model
- Update synthetic models list with new HF endpoints