Not a bad list. Off the top of my head, I would say it is missing two things:
- Discrete Math (formal logic, sets, probability, etc)
- Theory of Computing (not just algorithms, but things like Turing machines, NFAs, DFAs, etc.). These may not be strictly the most practical courses, but I think a Computer Science degree would be incomplete without these.
The “Introduction to Operating Systems” link no longer works (redirects to “Autonomous Systems” courses). Instead, I would recommend using Operating Systems: Three Easy Pieces, which is the textbook I use in my OS course.
Finally, something like The Missing Semester of Your CS Education would also be a nice extra.
Yeah this is severely lacking in terms of theoretical compsci.
As others have said, discrete math is one of the obvious missing pieces. My uni also has C as the first language students learn as a part of their degree, and follows up with Java and Haskell to teach students about OOP and FP as paradigms. It’s useful to have something like C so students can learn about memory management. I’m also not seeing anything on Networking and Cyber Security (aside from Cryptography), which my university also taught.
Discreet math, Calculus, and Physics change the way I think about code
Discreet math is a continuation of calculus, so can’t have it without that.
I personally hate physics class myself, more of a chem person. Biology can also be an option.
You know what else we’re missing? Statistic!
Why is it important in this day and age to learn about memory management? That’s like saying it’s important to learn cursive, when it really isn’t.
Embedded. I’m currently writing software with 96 bytes of RAM. My next project I get to splurge and have 8kB of RAM and 32k of Flash.
I’m more scared with how badly I’ll handle/manage the 8k of RAM.
Also anywhere where a GC is just too slow. Like in videogame engines.
That’s a very specific usecase though that the majority of programmers likely will never have to face.
Taking courses which involve subjects that you will likely never encounter in the workforce is a thing in every discipline. Most engineers don’t need to manually solve differential equations in their day jobs, they just need to know that they exist and will often require numerical solutions.
Getting your hands dirty with the content provides a better understanding when dealing with higher level concepts.
I think it’s more important than ever. Software is getting slower and slower, and the bloat is ridiculous. Imho this is because we just work with abstractions over abstracions ignorant to how it will be computed on a real machine. I think a more appropriate methapor would be, that you can speak and understand language (program) while being illiterate at the same time (not having a fundamental understanding on how a computer works). Of course this is a exaggeration, you don’t need to know about these stuff to be a adequate programmer I think.
@Suppoze @dubyakay one thing I liked about programming on Atari 8 bit machines was that your code could and was expected to hit the hardware directly. It was assumed the programmer understood the nature of the hardware and would directly “talk” to it to get it to perform their task. This made coding very efficient. Not a single CPU cycle of byte of RAM was wasted. A program that analysed data from multiple environmental sensors, tabulated, averaged and plotted the results and sent then to a charter plotter would run comfortably on 16kb of RAM.
My phone take a thousand times that to fail to open my emails.
They sure love them some Java. It’d be nice if they focused more on C/C++/Rust, you know, actual bare metal system languages that make you think about memory management.
Edit: I used to have a roommate who was studying compsci and they were making him program a PIC18F microcontroller on a development board in assembly. It was kinda fun because while he wasn’t using it, I’d have fun with it just programming normal C and making all the blinkenlights and gizmo peripherals on his board do shit, while he was struggling to even read a sensor.
In my CS degree I would have only learned and used java if not for my optional data science courses, a single class on machine language, a single SQL course, and a c++ course at community college before going to uni.
My data science courses introduced me to matlab, bash, r, Julia, python, machine learning, docker, Linux, and aws. My uni didn’t even have a data science degree, those courses primarily counted towards my math minor since they were under statistics.
The one piece of advice I still give to every CS student I meet is to diversify your classes whenever possible, don’t just stick to the core comp sci classes and take throwaway electives
Absolutely.
I don’t have a CS degree, I have a Cybersecurity and Forensics one. But, I love programming, and between the overlap of the two degrees and and my advanced designation I ended up taking about 3/4ths of the classes needed to get a CS degree.
Diversifying helped so much with me becoming a well rounded developer. My assembly programming class, while optional for CS, was mandatory for me, made me a significantly better dev. That assembly knowledge got me to become a skilled debugger, which made my C++ classes 10x easier, and it helped me understand memory at a lower level, making the memory problems easier to diagnose and fix.
I convinced a CS friend to take one of my cyber classes, Reverse Engineering, and he found te components of the class where we analyzed a vulnerable program to find and exploit the vuln, or the bit where we tried and determined the bug based on malware that exploited it is insightful to learning to program securely.
Learning about the infrastructure used in enterprise during a Windows admin or Linux admin class will make it easier to write code for those systems.
From the cybersecurity perspective, many of my CS classes carry me hard. Knowing how programs are written, how APIs are developed, and how to design complex software lets me make more educated recommendations based on what little information I’m given by the limited logs I am given to investigate. Writing code that interfaces with linux primitives makes it easier to conceptualize what’s going on when I am debugging a broken linux system.
Interesting, in my degree we had one lesson in Java for OOP (the rest of the course was C++), Java for android programming, Python in another course, and everything else from year 1 to year 4 (that had programming) was in C/C++. Except for assembly in computer architecture.
To be fair, this is my experience in academic computer science also
Edit: didn’t click the link before commenting, this is referencing what universities already teach
To be fair, this is my experience in environmental sciences too.
But okay, it was pretty broad. Like, pretty much everything scientific, data.
But no language courses, not much programming, missing stuff here and there.
But my Uni also encouraged (actually enforced) to choose random extra courses.
If it’s compsci, then it doesn’t need to be bare metal. It should be a language that’s good at demonstrating abstractions. Java wouldn’t be my choice, here. Elixir would be a good one.
You might want bare metal as a prereq to an operating system course.
If it’s software engineering, OTOH, then yes, a bare metal language has a bigger place.
If the Computer Science Degree is open source that means I can redistribute it.
You wouldn’t download a PhD.
It reminded me how much of a time waste formal education can feel. How come we can learn things on our own 10 times faster IF the motivation clicks in
It’s so weird tbh that at some unis they learn things in a year what you can get to know in a week if you don’t follow the slug pace plan and adhd hyperfocus kicks in
I don’t know what formal education you got but after 15 years in the industry, having learned, worked and taught colleagues in most major areas, I have yet to observe anything as efficient at learning as my university’s CS program.
I could just be one special thing then because if I am honest my colleagues always seemed slow and I had to debug their stuff all the time. I didn’t want to make a pointless comment about my subjective singular experience though and hoped it would be more universal
And if the adhd hyperfocus doesn’t kick in, it will instead sabotage you into never touching it.
Formal education gives structure and timelines to meet regardless of your mental state.
An excellent discrete maths textbook for those missing the inclusion of the subject in the course: Discrete Mathematics - An Open Introduction, 3rd edition by Oscar Levin
How did I miss this five years ago? What an excellent idea. I think we should be able to just finish the coursework and then put it on our resume.
I love this idea so much. I’m established in my industry but I think I’ll start working on it and just add to my resume. Thanks for sharing.
How would you add this to your resume? What would you write it as?
At the bottom in the
Education, Professional Development, & Credentials
section
Something like: Open Source Computer Science Coursework Completed XX hours of coursework through ABCD, EFGH, HIJK Universities Relevant Coursework: Linear Algebra (Princeton); Machine Learning (Stanford); Cryptography (Stanford)
It would weigh less than my traditional degrees, but if pressed on it (unlikely), I would describe exactly what this is: an effort to liberate CS education in the spirit of the Free Software movement, using synchronous and asynchronous learning methodology in an online learning platform from accredited, reputable universities.
At this point in my career, it would show continued aptitude for growth and professional development, since it’s been close to two decades since my first degree.
Also, at this point, I’ve seen people put shit like Strayer U and ITT Tech and Liberty on their resume and get hired for very high paying jobs. Honestly I would take this over that trash.
Even 15 years ago, most lower level undergrad coursework was 150+ students in a lecture hall where the professor would pull up Blackboard and just load the slideshow. It was only at the 300+ level where class size shrunk down and interpersonal relationships sort of mattered.
My wife’s graduate degree a few years later but still over a decade ago was almost entirely online; they only met in person to discuss their progress towards the capstone. And she has a nice prestigious degree with a very expensive university name on it, walked across the stage at that University, and nowhere does that diploma read, “Online.”
I have a lot of beef with the US university system. Change has to start somewhere.
What was the university and degree program? I might want to check it out.
Cool idea.
Commenting here to save this conversation.