anon 0x43e said in #2532 1mo ago:
I've been thinking about AGI stuff for over a decade, but never really engaged much with cutting edge machine learning practice except at a theoretical and philosophical level. But recently I've been getting more technical again and trying to build things, so I've been trying to learn some actual machine learning techne.
I like Geohotz's style so I'm learning ML with tinygrad. I did the quickstart tutorial[1], mnist tutorial[2], and then replicated the cutting-edge performance of the tinygrad mnist example code[3]. The quickstart tutorial was a mess, with half the stuff being either out of date or wrong and with a lot of subtle issues unremarked-upon. It was a bit of an extra IQ test to actually get a working model out of it, but I managed to pass it and get models trained, running, and optimized up to replicating the cutting edge performance in tinygrad's example code.
[1]: https://docs.tinygrad.org/quickstart/
[2]: https://docs.tinygrad.org/mnist/
[3]: https://github.com/tinygrad/tinygrad/blob/master/examples/beautiful_mnist.py
That was fun but I'm still very fresh. I think I get what's going on in these models, but if I had to design a performant one myself it all feels a bit like voodoo. I'll keep learning.
After that I got Claude to explain transformers to me and made sure I could explain it all back. Implementing a simple transformer from my own understanding feels like a good next exercise.
Ultimately where I want to go with this is to do online causality learning and control on time series data. Some of you may have heard rumor of my "worm mind hypothesis" and AGI thermostat idea. I basically want to substantiate those in actual algorithms research: build an algorithm that you can simply drop into a nonlinear domain with no training, only a goal, and it learns the causal structure of the domain sufficient to train itself to accomplish the goal. I have a paradigm hunch in mind for how to do this that follows the insights of transformers and deep learning (though not all their particular details), so that's why I'm interested in these techniques in particular. But I have a lot to learn, and my hunch may just turn out to be an excuse for learning some new skills.
More generally, as I develop towards a more technical phase in my career, I want to be fluent in the cutting edge and able to apply some of this exciting ML stuff to the problems I encounter.
Does anyone have good resources for learning how to solve problems with machine learning techniques, or good exercises to do? Do any of you work in the field and have tips?
I like Geohotz's style so I'm learning ML with tinygrad. I did the quickstart tutorial[1], mnist tutorial[2], and then replicated the cutting-edge performance of the tinygrad mnist example code[3]. The quickstart tutorial was a mess, with half the stuff being either out of date or wrong and with a lot of subtle issues unremarked-upon. It was a bit of an extra IQ test to actually get a working model out of it, but I managed to pass it and get models trained, running, and optimized up to replicating the cutting edge performance in tinygrad's example code.
[1]: https://docs.tinygrad.org/quickstart/
[2]: https://docs.tinygrad.org/mnist/
[3]: https://github.com/tinygrad/tinygrad/blob/master/examples/beautiful_mnist.py
That was fun but I'm still very fresh. I think I get what's going on in these models, but if I had to design a performant one myself it all feels a bit like voodoo. I'll keep learning.
After that I got Claude to explain transformers to me and made sure I could explain it all back. Implementing a simple transformer from my own understanding feels like a good next exercise.
Ultimately where I want to go with this is to do online causality learning and control on time series data. Some of you may have heard rumor of my "worm mind hypothesis" and AGI thermostat idea. I basically want to substantiate those in actual algorithms research: build an algorithm that you can simply drop into a nonlinear domain with no training, only a goal, and it learns the causal structure of the domain sufficient to train itself to accomplish the goal. I have a paradigm hunch in mind for how to do this that follows the insights of transformers and deep learning (though not all their particular details), so that's why I'm interested in these techniques in particular. But I have a lot to learn, and my hunch may just turn out to be an excuse for learning some new skills.
More generally, as I develop towards a more technical phase in my career, I want to be fluent in the cutting edge and able to apply some of this exciting ML stuff to the problems I encounter.
Does anyone have good resources for learning how to solve problems with machine learning techniques, or good exercises to do? Do any of you work in the field and have tips?
I've been thinking a