I hope you've had a good start to 2021! By way of holiday travel, annual review, and additional content production, it's been more hectic for me than I would prefer, but I'm starting to get back in a rhythm of good habits.
Onto the newsletter! In this week's edition:
I originally intended for the 2020 podcast compilation to be only one episode, but after looking through all the clips I wanted to include, I decided to break it up into two by topic: one this week on ML engineering, and one next week on ML research and careers.
Featured in this episode:
Click here to listen to the episode, or find it in your podcast player of choice: https://www.mlengineered.com/listen
It’s quite amazing that the entire software industry is young enough that many of its early pioneers are just starting to retire. A blog post was shared with me recently by a past guest, swyx, with some deceptively simple lessons from someone who’s been in the industry for over four decades.
Check it out here and also see the HackerNews comments for additional insight.
On Tuesday, Open AI dropped two incredible blog posts detailing their recent work on extending GPT-3 for multi-modal learning. This is the type of stuff that makes me extremely excited to be working in this field:
![]() |
A few months back, Google made a splash in the research scene with the release of their Vision Transformers paper (which I covered in this edition of the newsletter). Since then, we’ve seen a Cambrian explosion of follow-on research, including OpenAI’s multi-modal work above and Facebook’s DeiT, a more data-efficient ViT with a new transformer-specific knowledge distillation technique (see their results on Imagenet below).
![]() |
There’s also been research into interpreting and explaining what’s going on behind the scenes of these models.
Jacob Gildenblat has an excellent blog post showing various techniques for explainability applied to one of Facebook’s released models. Check it out here.
![]() |
If you want to dive deeper into this, Facebook also has a recent paper introducing additional techniques for transformers: Transformer Interpretability Beyond Attention Visualization
And to get a broader overview of all the research that’s been done on this topic, a survey on ViTs was recently published: Transformers in Vision: A Survey
Chip Huyen is one of the best bloggers on ML engineering today and she recently published an article detailing her thoughts on the trend of more ML systems needing to make online predictions and what’s needed to accomplish that. I especially enjoyed the latter section on how companies handle model testing after online training.
Check it out here: Machine Learning is Going Real-Time
I mentioned last week that I had to skip an episode release because of guest reschedulings. Since then, I’ve recorded five episodes, which was fun but extremely tiring. Rest assured, I won’t be missing another release! Also, in case you missed it, I wrote up a study guide for aspiring ML engineers that lays out a clear starting path and contains a list of resources that I and my friends have learned from. Read the Study Guide In this week's edition: My Interview on the MLOps Community podcast...
There’s no podcast episode this week due to an unfortunate coincidence of multiple guests needing to reschedule. My apologies, I’ll be doing my best in the future to not let this happen again. That doesn’t mean I don’t have any new content for this week, though! Today I’m releasing an article that answers one of the most common questions I get: “I want to learn machine learning, where do I start / what do I do?” When I was first getting started in ML, it was pretty straightforward: there was...
After a month off from releasing original interviews on the podcast feed, I’m so excited to be sharing this episode with all of you! Aether Biomachines is one of the most interesting machine learning startups I’ve ever come across and I was thrilled to interview the founder, Pavle Jeremic. Building a Post-Scarcity Future using Machine Learning “How can we make sure that the economy is so productive that the desperation that leads people to commit atrocities never happens?” In this episode,...