Jeah, even without knowing glsl language, in some occasions I managed to feed chat gpt with the right prompt that 30 shaders in a raw where bug-free, sometimes it’s harder - but learning glsl is not hard when you already have experience in c++
@@b3n70n I will but it on modular grid when I find some time, have soldered and tested about 5 modules, changed 1-2 things in the firmware and figured out a way you can pick presets you like. A matter of day’s, I guess
@@b3n70n 8 cv in, 0 to 5 volt positiv, protected, 4 pots normalised to input 5-8 ( also for menu/mode selection ), modes and functions can be assigned to different inputs, like choose effect, choose picture ( not seen in this video but you can load and process custom .bmp pictures ), Randomise on trigger input, clock input, audio detection and more in the next firmware like internal lfo, modulation matrix and so on, and, I hope soon also a video player/sampler/slicer
@@dieAnthropologischeKonstante interesting, I used OP-Z in the past for visuals. Interesting concept TE has. Sadly packs are difficult to do and the iOS app is not consistent, can crash during a live performance so it is a bummer.
@@giantmecha almost none. Is a bare metal is, I work without a Linux like similar diy projects. And some shaders can run at 200 fps, so is the idea to implement also a second layer of shaders, for dedicated glich shaders ( reset & user ) and/or blend shaders for real seamless transition between the shader:)
@@giantmecha I started on a c64 and when you have seen what they did back in the 90‘s ( Till now ) on this machine you learn a few things: Code in assembler code or c /c++. „Hardwarerestrictions“ are really only „bad programming decisions“. you don’t need a full Linux, FFmpeg or touchdesigner api and than an python interpreter ( because it is fancy) and than your code on top: you can free up about 99.99 of the processor load by looking for a low level approach;)
@@dieAnthropologischeKonstante Ah yes makes sense. There are some other shared-based modules out there that use RasPi (Structure, Hypno), but I'm not sure if they are coded in assembly or c/c++, if they are built on top of linux, etc.
This is awesome! 🔥
Amazing! Good work!
incredible potential!
Jeah, even without knowing glsl language, in some occasions I managed to feed chat gpt with the right prompt that 30 shaders in a raw where bug-free, sometimes it’s harder - but learning glsl is not hard when you already have experience in c++
Really interested to get one and mess around with my freshly finish rack ❤
Really like your work!
Thanks, hope we will get this done soon
Sounds interesting 🎉
This is great.
Thanks
Keep up the good work, but is Oculus Mendax on Modular Grid? What modulation inputs does it have?
@@b3n70n I will but it on modular grid when I find some time,
have soldered and tested about 5 modules, changed 1-2 things in the firmware and figured out a way you can pick presets you like.
A matter of day’s, I guess
@@dieAnthropologischeKonstante waiting patiently 😁
@@b3n70n 8 cv in, 0 to 5 volt positiv, protected, 4 pots normalised to input 5-8 ( also for menu/mode selection ), modes and functions can be assigned to different inputs, like choose effect, choose picture ( not seen in this video but you can load and process custom .bmp pictures ),
Randomise on trigger input, clock input, audio detection and more in the next firmware like internal lfo, modulation matrix and so on,
and, I hope soon also a video player/sampler/slicer
@@dieAnthropologischeKonstante interesting, I used OP-Z in the past for visuals. Interesting concept TE has. Sadly packs are difficult to do and the iOS app is not consistent, can crash during a live performance so it is a bummer.
modulargrid.net/e/other-unknown-oculus-mendax
very nice! looks like there's quite low latency when changing between patches?
@@giantmecha almost none. Is a bare metal is, I work without a Linux like similar diy projects.
And some shaders can run at 200 fps, so is the idea to implement also a second layer of shaders, for dedicated glich shaders ( reset & user ) and/or blend shaders for real seamless transition between the shader:)
@@dieAnthropologischeKonstante Wow, must be some efficient code. I didn’t know a Pi Zero was capable of that :)
@@giantmecha I started on a c64 and when you have seen what they did back in the 90‘s ( Till now ) on this machine you learn a few things:
Code in assembler code or c /c++.
„Hardwarerestrictions“ are really only „bad programming decisions“.
you don’t need a full Linux,
FFmpeg or touchdesigner api and than an python interpreter
( because it is fancy)
and than your code on top:
you can free up about 99.99 of the processor load by looking for a low level approach;)
@@dieAnthropologischeKonstante Ah yes makes sense. There are some other shared-based modules out there that use RasPi (Structure, Hypno), but I'm not sure if they are coded in assembly or c/c++, if they are built on top of linux, etc.
enjoyed ! cool stuff ! now i need one 😄 drop me a message when available !
I will