Subject Code/Name: COMP9417 - Machine Learning and Data MiningEquivalent postgraduate:
COMP9417 (identical course code)
Contact Hours: 2x2hr lecture, 1hr tutorial
Assumed Knowledge:Undergrad: Two pathways: (MATH1081 + either COMP1531/COMP2041), or COMP2521
Postgrad: COMP9020 + COMP9024
Data structures and algorithms (both UG and PG) and knowledge of python suffices for the computing aspect. But you really should know some calculus, linear algebra, and statistics, in preparation for the math side.
Assessment:- 1 x 1% homework
- 2 x 7% homeworks
- Weekly questions from tutorial set, best 7 out of 8 counts, 5%
- 30% project - hackathon, or comp9417 group project
- 50% 90min final exam
Lecture Recordings? Yes, on Microsoft Teams and UNSW Echo
Notes/Materials Available: Relatively detailed lecture slides and tutorial sets. Half of the labs were very in detail; presumably all labs will be in detail next term. Also some supplementary youtube recordings from the head tutor. Head tutor managed the course forum very actively. Overall surprisingly abundant set of course resources. (However, the internet is still a valuable resource for more niche concepts.)
Textbook: No single textbook recommended anymore. A list of optional textbooks for further reading provided on the course outline on webcms3, but I didn't use any of them.
Lecturer(s): Dr. Gelareh Mohammadi
Year & Trimester of completion: 22T1
Difficulty: 3/5 (however hackathon can boost this up to 4.5/5)
Overall Rating: 4/5
Your Mark/Grade: 97 HD
Comments: This is one of many Ai courses offered at UNSW. At this point I really feel "machine learning" is a buzzword, but the course outline definition is loosely speaking enough. Namely that ML is the algorithmic approach to learning from data. It can be perceived to have a similar goal to statistical modelling, but in ML prediction accuracy tends to overrule interpretability of the model.
The course introduces some classical ML techniques, but also touches on pieces of the current state-of-the-art models (e.g. ensemble learning, neural nets). There's quite a lot of content, but this is to be expected since ML is currently rapidly growing. Generally speaking it is a good overview to current ML techniques though. (Surprisingly, it's also made me appreciate neural nets more, despite only spending 1 week on it.) As a result of so much content though, the lectures were quite fast paced. For a math major like me i didn't care, but I can see it being difficult for other students.
I should direct your attention to
this review briefly, and how the final exam dragged a 3/5 down to a -5/5. Thankfully that was over. No idea if the different lecturer meant anything here, but my exam was essentially 50 MC. Not a great experience per se - the curveball questions were quite hard. But the exam didn't feel evil or bizarre at all.
What hurt the rating? Well, homework 0 was a grind for just 1%. Not a hard to get 1%, but tiresome. As the course progressed, this was kind of forgotten, because both subsequent homeworks were interesting and made up for it. Then it came to the project. In all fairness, the hackathon itself was interesting - good final goal we were aiming for, and
it gave a taster of real world data science. Have to go self learn stuff (e.g. modelling beyond 9417 scope, mastery of pandas), but that's okay. What sucked was the server and the restraints. Painfully slow to work on AWS against 50 or so groups, all trying to fit these high CPU consuming models all the time. Could only make 10 submissions a day (previously 1 submission a day which was worse), so hard to get better model performance. Some stuff just took hours to run. (Also personally salty at the final rankings.) Difficult to tune models as well. Overall killed the course experience. Nevertheless, didn't throw the course down in the dumps or anything.
Also take some caution with the last topic on learning theory. Very interesting, but much more mathematical. Might be hard if you haven't done theoretical comp sci. But it only lasts a week, so it can't torture you too much if you hate it.