Skip to content

Instantly share code, notes, and snippets.

View wassname's full-sized avatar
🤖

wassname (Michael J Clark) wassname

🤖
View GitHub Profile
@wassname
wassname / logs.md
Last active April 26, 2026 02:28
Coding agent compatible logs. concise, inline guidance, token effecient

Principles

  • don't use many tokens
  • make it so a dumb summary LLM can easily 1) see problems 2) have clues to diagnose
  • timing information

example of good log

  • with single line should statement inline in log that make it clear how it should look, distinguish from subtle failures, and give principled clue for diagnosis
  • table short have longest and least important lines last, so that humans can read it even with wrap around e.g short numeric columns first. long text columns last, notes or desc last
  • use tabulate plain for token effecient, not logging each step or epoch but just table
@wassname
wassname / rep_pen.md
Created April 23, 2026 00:47
hf transformers repetition penalty
@wassname
wassname / classify_linear_sublayers.py
Created April 3, 2026 08:55
classify_linear_sublayers as residual read or write
def classify_linear_sublayers(
model,
block_layers: list[str],
) -> dict[str, list[str]]:
"""Classify all Linear sublayers in each block by their residual-stream role.
For a Linear layer with weight shape [out_features, in_features]:
- residual_write : out_features == d_model (writes TO residual stream)
- residual_read : in_features == d_model (reads FROM residual stream)
@wassname
wassname / guided.py
Last active April 23, 2026 05:40
Guided CoT Evaluation: Hybrid Teacher-Forced + On-Policy Reasoning
"""Reusable guided-rollout primitive: thinkforced-close-thinkJSON choice.
One rollout, three numbers. The same primitive backs:
- calibrate()'s coherence + format + rep measurement
- probe replay at edit time
- post-keep probe regeneration
- (future) DD eval (once _measure_logratios is ported to this)
Substrate:
<user_prompt + schema_hint>
@wassname
wassname / justfile
Last active April 3, 2026 08:55
snippet showing how to do a smoke test including beartype and jaxtyping only on smoke test
# Smoke test: demo on task 0 showing steered outputs at -1, 0, +1
smoke *ARGS:
BEARTYPE=1 {{ PY }} ssteer_v3.py --quick {{ ARGS }} 2>&1 | tee /tmp/smoke.log | tail -80
@wassname
wassname / bwap_cuda_uv_fix.md
Created March 2, 2026 04:34
claude bubblewrap with cuda and uv

GPU + uv in Claude Code sandbox (Linux/NVIDIA)

Problem: Claude Code's sandbox (bubblewrap) blocks GPU devices and the uv package cache.

Solution: bwrap wrapper at ~/.local/bin/bwrap

@wassname
wassname / how_to_get_logprobs_from_generation_v3.ipynb
Last active November 10, 2025 12:34
problem: model.generate does not return input logprobs, solution `model.forward`, then `model.generate(past_key_values)`, also get logprobs on the last token of a stopping sequences
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
@wassname
wassname / generate_with_input_logits.py
Last active November 9, 2025 00:07
generate_with_input_logits and clone_dynamic_cache
import torch
def generate_with_input_logits(model, tokenizer, batch2, **kwargs):
"""
problem: generate does not return logits for inputs, but we need them for nll
but forward -> generate with past key values does, and it doesn't recompute the input logits
so this is a helper that does both
@wassname
wassname / example_gen_usage.py
Last active November 9, 2025 11:31
Use a regexp to greedy select minimal tokens that satisfy a regexp
"""
Stopping criteria: regexp
ref:
- https://huggingface.co/docs/transformers/v4.56.1/en/main_classes/text_generation#transformers.GenerationMixin.generate.stopping_criteria
- https://github.com/huggingface/transformers/blob/e8a6eb3304033fdd9346fe3b3293309fe50de238/tests/generation/test_stopping_criteria.py#L51
"""
from transformers import StopStringCriteria, StoppingCriteriaList, EosTokenCriteria
@wassname
wassname / peft_adapter_papers.md
Last active February 22, 2026 10:12
PEFT Adapter papers

PEFT Adapter papers

Orthogonal methods: OFT, BOFT, HRA, ROAD, GOFT (Givens), OFTv2 Low-rank methods: LoRA, AdaLoRA, LoHa, LoKr, RandLoRA, VBLoRA, FourierFT, DeLoRA Scaling methods: IA3, VeRA Prompt-based: Prompt Tuning, Prefix Tuning, P-Tuning, Adaption Prompt, Multitask Prompt Tuning, CPT Specialized: MiSS, SHiRA, C3A, LN Tuning, Poly, XLoRA