so i kept pushing on shap and ended up spending days slowly working on 
finegling the ai to refactor the monstrous mess it had made into something 
organized and simple. it would have been much quicker to start over with the 
knowledge i had gained. but it was mostly a recreational thing so i just went 
at it. the current state is in the path-surface-unification branch. now i am 
out of money (it costs $6 to replenish anthropic and i hit the button so much 
that i now don't even have $6 in my account. this will change come next month.) 
and out of compute on github codespaces where i was running their free 
in-browser vscode that can run cline.

things i learned:
- cline is for people new to coding or people with psychological disorders 
involving using corporate AI needlessly. it would be relatively easy for a 
coder emotionally free to do so to make something better that is much cheaper 
or free. most of the cline work is on making a great UI. they also have tuned 
prompts in their source that are designed to work well with how the Claude 
model was trained. 
- it's really good to stay involved with it, to observe it as it functions and 
advise it whenever it does something one doesn't want. this prevents things 
from compounding. honestly the moment you don't do this and let it run on its 
own, is the moment the codebase turns into a big mess.
- anthropic is both cheaper than openrouter as well as higher quality. 
openrouter seems to be engaging poor sampling settings.
- if you started with a free model at the start, you'd get more used to the 
severe quirks associated with doing so and could find paths that engage that 
situation more successfully. it's still challenging though.
- it seems it would be good to focus on really powerful core parts of a 
project, deleting cruft files, until the basics of the architecture are really 
solid. this then gives it a good environment to build off of, which it might 
respond better to than requests or prompts around coding standards.
- it's quite useful to direct it to work with documents that describe what it 
is doing, this lets one cut the context short and handle crashes well
- i had other things to write here but don't remember them so just expanded 
upon the bits i did, maybe wasn't best time to write this

yup!

free lightweight (could implement from a cell phone) approaches might focus 
around [automated] prompt tuning for each task involved
- writing commit messages based on portions of information
- updating single files to meet small requests
- summarizing or portioning information in ways that are useful for other tasks
note that wiring these things together effectively looks similar to designing 
an overall learning architecture. prompt tuning for one thing  that impacts 
many other things might want success/failure metrics to propagate up, etc ...
a prompt tuner could also be a model selector.

it was fun but it's not cheap enough yet to be a sustainable addiction for a 
homeless chap. it leaves me feeling confused ... how did i feel satisfied 
without writing code? where are my coping attempts for still trying to write 
code myself ...?

still waiting on those products that powerfully influence on their own the user 
to become skilled and independent

Reply via email to