karl3ļ¼ writeme.com wrote:
> so i kept pushing on shap and ended up spending days slowly working on 
> finegling the ai to refactor the monstrous mess it had made into something 
> organized and simple. it would have been much quicker to start over with the 
> knowledge i had gained. but it was mostly a recreational thing so i just went 
> at it. the current state is in the path-surface-unification branch. now i am 
> out of money (it costs $6 to replenish anthropic and i hit the button so much 
> that i now don't even have $6 in my account. this will change come next 
> month.) and out of compute on github codespaces where i was running their 
> free in-browser vscode that can run cline.
> 
> things i learned:
> - cline is for people new to coding or people with psychological disorders 
> involving using corporate AI needlessly. it would be relatively easy for a 
> coder emotionally free to do so to make something better that is much cheaper 
> or free. most of the cline work is on making a great UI. they also have tuned 
> prompts in their source that are designed to work well with how the Claude 
> model was trained. 
> - it's really good to stay involved with it, to observe it as it functions 
> and advise it whenever it does something one doesn't want. this prevents 
> things from compounding. honestly the moment you don't do this and let it run 
> on its own, is the moment the codebase turns into a big mess.
> - anthropic is both cheaper than openrouter as well as higher quality. 
> openrouter seems to be engaging poor sampling settings.
> - if you started with a free model at the start, you'd get more used to the 
> severe quirks associated with doing so and could find paths that engage that 
> situation more successfully. it's still challenging though.
> - it seems it would be good to focus on really powerful core parts of a 
> project, deleting cruft files, until the basics of the architecture are 
> really solid. this then gives it a good environment to build off of, which it 
> might respond better to than requests or prompts around coding standards.
> - it's quite useful to direct it to work with documents that describe what it 
> is doing, this lets one cut the context short and handle crashes well
- i ended up having a strategy of using the highest quality model at the start 
of a task, and then reducing the model quality farther in to reduce costs and 
such. This leaves the higher-quality responses in the context of the prompts 
used for the lower quality model, which gives them more intelligent and robust 
completion patterns. they still don't quite compare though.
> - i had other things to write here but don't remember them so just expanded 
> upon the bits i did, maybe wasn't best time to write this
> 
> yup!
> 
> free lightweight (could implement from a cell phone) approaches might focus 
> around [automated] prompt tuning for each task involved
> - writing commit messages based on portions of information
> - updating single files to meet small requests
> - summarizing or portioning information in ways that are useful for other 
> tasks
> note that wiring these things together effectively looks similar to designing 
> an overall learning architecture. prompt tuning for one thing  that impacts 
> many other things might want success/failure metrics to propagate up, etc ...
> a prompt tuner could also be a model selector.
> 
> it was fun but it's not cheap enough yet to be a sustainable addiction for a 
> homeless chap. it leaves me feeling confused ... how did i feel satisfied 
> without writing code? where are my coping attempts for still trying to write 
> code myself ...?
> 
> still waiting on those products that powerfully influence on their own the 
> user to become skilled and independent

Reply via email to