• Suppoze@beehaw.org
    link
    fedilink
    arrow-up
    5
    ·
    3 months ago

    I think it’s more important than ever. Software is getting slower and slower, and the bloat is ridiculous. Imho this is because we just work with abstractions over abstracions ignorant to how it will be computed on a real machine. I think a more appropriate methapor would be, that you can speak and understand language (program) while being illiterate at the same time (not having a fundamental understanding on how a computer works). Of course this is a exaggeration, you don’t need to know about these stuff to be a adequate programmer I think.

    • Syulang@aus.social
      link
      fedilink
      arrow-up
      0
      ·
      3 months ago

      @Suppoze @dubyakay one thing I liked about programming on Atari 8 bit machines was that your code could and was expected to hit the hardware directly. It was assumed the programmer understood the nature of the hardware and would directly “talk” to it to get it to perform their task. This made coding very efficient. Not a single CPU cycle of byte of RAM was wasted. A program that analysed data from multiple environmental sensors, tabulated, averaged and plotted the results and sent then to a charter plotter would run comfortably on 16kb of RAM.

      My phone take a thousand times that to fail to open my emails.