Conversation

Lucy [hiatus era] 𒌋𒁯

comments don't help if you were high writing them
1
1
5
@lucy I use copilot to write comments, works very well
2
0
0
@arcana i actually tried chatgpt to generate some assembly code out of curiosity and it
- failed to even produce something i could compile
- completely messed up basic tasks such as "copy n bytes from pak rom into ewram"
- used instructions that do not exist and refused to be corrected by me

so, maybe it can generate some webshit templates or perl script or whatever but here it utterly fails.

im writing assembly code for the gba.
4
0
1
@lucy @arcana ive used chatgpt before to see if there were better ways to write some piece of code, like send it a section of code and ask if it could be done better
it can help sometimes
2
0
1

Lucy [hiatus era] 𒌋𒁯

@arcana i also fed it some code that didn't work on purpose because of memory alignment and it couldn''t resolve the problem and instead tried to gaslight me that the code works indeed lmao
0
0
1

@arcana@fedi.layer02.net @lucy@netzsphaere.xyz in my experience llms are mainly useful for: generating comments, generating tests (not the actual contents but a template of basic unit tests that you then write or finish yourself), looking up documentation (just make sure to double check the actual docs afterwards)

4
0
3

Lucy [hiatus era] 𒌋𒁯

@paula @arcana probably because webdev is repeatitive and there's an ocean of training material
0
0
0
@anemone @lucy @arcana So basically they're mostly useful in boilerplate-heavy languages that lack good abstractions and metaprogramming capabilities lisp
1
1
3

@scathach @anemone @lucy @arcana yeah was about to say, that’s a lot of processing power spent on things that can be done better by editor macros

0
1
3
@anemone @lucy yes I think so, they’re good with MVC apps too I find to quickly add things that are solved problems. I find them useful for lots of things really.
1
0
0
@lucy what model was it and when did you do this?
1
0
0
@anemone @lucy @arcana >generating comments

Given most LLMs' context-window restriction wouldn't they have a tendency to generate comments that describe the wrote-mechanics of some code rather than describing why it's doing what it's doing in the larger context of the program?
3
0
4

@arcana@fedi.layer02.net @lucy@netzsphaere.xyz yeah generating common patterns often works though it still has a fairly high error rate. I'd compare it to having the average stack overflow poster consolidated into a jr dev employee who works extremely fast but is still pretty dumb

1
0
3

@anemone @arcana @lucy I've used chatgpt to analyze code and make it explain to me what it does, with great success. Not that it was overly compelx code, but still, pretty nice

0
0
0

@scathach @anemone @lucy @arcana they also get it wrong all the time because all they do is pattern matching

3
0
1
@anemone @lucy to me, it feels basically identical to when I had two juniors working under me but it’s much much faster, and with good prompting, responds much more accurately
0
0
0

@mia@movsw.0x0.st @scathach@stereophonic.space @lucy@netzsphaere.xyz @arcana@layer02.net pattern matching can, imo be useful, but I don't expect LLMS to write code for me, just cut out a little busywork

0
0
0
@scathach @anemone @lucy the context windows are extremely huge these days and it’s not like you can’t edit them as you go
1
0
0
@lucy lmao, yeah, try o1
2
0
0

@arcana@fedi.layer02.net @lucy@netzsphaere.xyz yeah any of the gpt-4 derivatives are significantly better than gpt-3

2
0
2
@mia @scathach @anemone @lucy it’s odd I think how differently people seem to get output from them. Some seem to find them useless and to some they’re amazing. As somebody who has managed juniors, they’re very good imo, at least the latest models, and respond well to management techniques and language I find
1
0
0

@paula@netzsphaere.xyz @lucy@netzsphaere.xyz @arcana@fedi.layer02.net yeah im it's not good for writing actual code but it can be good for scaffolding for tests and comments, and so on

0
0
0
@mia @scathach @anemone @arcana very toddler tasks it works but anything complex, where you actually have to deeply understand what your code does it's useless. the worst of llms is how they ALWAYS try to generate SOMETHING and won't hesitate to present something broken with absolute confidence.
1
0
1
@anemone @lucy yeah, if its gpt3 and not even 3.5 then that’s incredibly basic.
1
0
0

@anemone @arcana@fedi.layer02.net @lucy i still don’t expect these to be any better at this task

0
0
0
@lucy @anemone @mia @scathach you can minimize that heavily as part of your system commands
0
0
0
@arcana nah not worth my time and money
1
0
0
@arcana @anemone can you even use gpt-4 without paying/signing up?
2
0
0
@lucy fair enough, for me it’s been a massive return on investment, you used to have to pay people 25k a year plus NI for this kind of stuff, if not more
0
0
0

@arcana@fedi.layer02.net @mia@movsw.0x0.st @scathach@stereophonic.space @lucy@netzsphaere.xyz it depends on what you're going for. You can massage them into writing something that works but it will rarely be done in an efficient or maintainable manner. That's still good enough for some use cases, though, and can save people time if they're not skilled programmers. I worry about the long term consequences of this use case, however.

I mostly fine it useful as a scaffolding (comments, tests) or search/research tool. I can ask it how to do something and the answer will be wrong, but it will likely point me to a useful library, something search engines are pretty bad at these days.

1
0
1
@lucy @anemone no idea, I’m a paid subscriber and have been since last April
0
0
0
@arcana @anemone @lucy How big? For context the largest codebase I'm 100% responsible for is about 180 KB and the copy of the Linux 6.10.5 source tree on my computer has about 1.3 GB of source files, including the build system and utility scripts
1
0
0
@anemone @mia @scathach @lucy that’s part of what the system instructions help with, you can set it up to be much more efficient and effective. I have multiple different setups for different use cases too, it can be really fast, you can also set up quite quickly automatic chains much like in a business where one set up to behave as a specific type of employee delegates to others and checks/tests their work. The voice mode is awesome too because I can do so much of this while working out or doing other things just by talking on the phone.

The skills I learnt in managing people come in very useful with it honestly
0
0
0
@anemone @lucy I don’t think you can use o1 or advanced voice mode, or set up custom workflows
1
0
0
@scathach @anemone @lucy 128,000 tokens for the latest GPT models
1
0
0
@arcana @anemone @lucy What's a good heuristic for judging the rough number of tokens that would be produced by some text?
1
0
0
@anemone @lucy sure, but o1 is the most advanced model currently available via OpenAI and so it’s the best if you want to assess their current capabilities in relation to what others claim
1
0
0
@anemone @lucy Also o1 is under the GPT4 family as is the advanced voice mode, so my reply was answering the question as well, as GPT4 in its entirety includes those things.
0
0
0
@lucy @arcana it will also fuck up webshit in subtle ways
0
1
1