Skip to content

An AI survival guide

Published:
25 min read

This blog post is my own attempt at helping any reader interested in AI learn, keep track of the fast progress and navigate the enormous amount of information without getting lost in the slop. I will provide some resources and advice I have found helpful during my own learning so far. I am only a junior in the field and I still have a lot to learn, but what has helped me can surely help anyone else, so take this post as a non-exhaustive, subjective starting point.

We’ll cover several aspects, such as curating your feed to get the right news, selecting the right resources to maximize your learning curve, exploring some helpful tools, benchmarks, and more. I am leaving the final section of this post for some resources.

Cutting through the noise: curating your feed

Curating your feed

If there’s one exciting thing about AI right now, it’s the amount of new releases. Every week or even couple of days, we get a major release. I think this is something particularly unique at this point in time and it makes the learning process even more exciting, challenging and fun. I don’t think other engineering domains get to experience this “Netflix-like” vibe. At least I didn’t see any such parallel in say civil or mechanical engineering so far, or not at this scale. However the fast-growing AI world also comes with downsides. You just have to open Linkedin to get a glimpse of it. A lot of slop, unrealistic claims, psychosis and hype and it can be hard to distinguish what matters from the clutter. So the first thing I would suggest is to get the right information, at the right time and in the right place to keep up with new releases as they come.

From my own experience, the best way to get a good feed of news right now is twofold: newsletters and X (twitter). Now the latter is highly controversial as X can be a very toxic / sloppy place too, but it’s all about following the right people and sticking to the notifications of these people and avoid the homepage like the plague.

Newsletters

I don’t know much about newsletters, but I’ve stuck with one in particular for several months and I think it might be worth your attention: AINews by smol.ai. This newsletter is an AI roundup summarizing discussions across AI communities (mainly from X and Reddit) and created by customizable research agents. It was founded by Shawn Wang, currently AI Engineer in San Francisco. I personally enjoy this format, and mainly use it to get the main headlines at a glance. I often only go through the X summary and read between the lines. However some issues I can note are : very often images are not displayed in my inbox or sometimes I do not receive the email at all. So most of the time, I would check their website directly.

X feed

As mentioned earlier, X can be a great place to keep up with the news, learn and interact with the community. The value you will get out of it only depends on your own usage. But I will try to give you some advice as a first idea for you to start. I never or almost never visit the homepage. The X algorithm is not of my taste, it would often recommend the same content over and over again, some good content would be lost in the middle of the slop, the psychosis and a bunch of people claiming bold predictions. I would recommend you to avoid the homepage if your goal is to maximize the value you get out of twitter and make much better use of your time. That’s why 99% of my usage is only spent in the notifications tab where I only get to see the posts and replies from the people I actually care about. In this feed I would get detailed blog posts regarding technical topics (e.g a recent blog post on a tiny model to prove theorems : qed-nano-blogpost), latest AI model benchmarks (e.g Designarena on X (post)), miscellaneous projects / blog posts (e.g Simon Willison on X (post)), new model releases (e.g omarsar0 on X (post)), AI adoption in the industry (e.g bcherny on X (post)) and so much more. Obviously you will still find people with bold claims and everyone is having their own opinions on the topic, commercial incentives and so on but overall it is fairly easy to build a trustworthy feed to stay up to date, find great reads, explore new ideas. Another great thing about X is that you can start interacting with the community as well by asking for explanations when in need, contributing to some projects with your own work and so on. I list some X accounts as suggested resources for you to explore in the final section of this post.

Keeping track of benchmarks

Even though your newsletter and your X feed will have already given you a good idea of which benchmarks to look at, I am dedicating a small section here to reference some benchmarks you might want to keep an eye on. First of all a good default can be Arena. This is a community driven benchmark where models get evaluated by users who prompt models on real tasks. Users vote on which model provides the most satisfactory answers to their query and accordingly models get ranked across several categories (Code, Vision, Text…). You can also chat with many of these models for free and vote to help the community assess these models.

I would also mention METR. This organization is quite famous for the task-completion time horizon evaluation which is the task duration, measured by human expert completion time, at which an AI agent is predicted to succeed with a given level of reliability.

For a broad overview of models capabilities, you can also visit LiveBench which is a challenging and contamination-free benchmark sponsored by Abacus.AI. They release new questions regularly and the evaluation spans a diverse set of tasks which are all verifiable eliminating therefore the need for any LLM as a judge.

Finally ARC-AGI is having a set of very interesting benchmarks to measure fluid intelligence (i.e focusing on the ability of models to learn new skills). There are obviously a lot more benchmarks you can follow and new ones are often released but this is not my goal to list them all here. I leave it to you to explore and choose your own favourites :)

Learning the theory

Learning the theory

If you have already curated your feed properly, you will soon realize that there is already a lot to learn from. Not necessarily with the posts themselves but rather with the content these posts will redirect you to. However there are still many resources on which you can rely to strengthen your foundations in AI.

Research papers

The first medium I can recommend for you to learn is reading research papers. It is one of the most trustworthy source of information you can find but it is also highly technical and the entry barrier knowledge-wise can be quite high. But the good news is that it has never been easier to look for the research papers you may be interested in as well as breaking them down. To find the right paper you can obviously use Arxiv or Google Scholar but what I would most of the time rather do is spin up an LLM to understand my precise intent and find the paper for me, and most often end up reading it on arxiv. Perplexity used to be my default “go to” for such tasks, but I am not sure if the gap is still real in terms of reliable source retrieval among LLM providers. On this I would rather suggest you to try different providers and stick to the one providing you the best results. Now when it comes to break down research papers, I found it also helpful to keep an LLM on the side to ask questions throughout your reading process. Lately, OpenAI Atlas stood out as a convenient option for this use case according to me. You get to have your research paper as the main page of the browser and a little chat on the right side which removes the burden of switching back and forth between tabs, copy paste content here and there… You can simply ask for clarification regarding the paper, drop your best ELI5 prompt at complicated jargon and get the gist of something you are not familiar with. Overall I would suggest to always try to leverage LLMs to deepen your understanding rather than offloading your cognitive tasks. It’s not simple, it requires efforts and our brains are lazy by nature and would often prioritise the easiest option. We have all been there…

Another useful tool worth mentioning is Deepwiki. Whenever you stumble upon a technical repository, or poorly documented, or simply want a useful and convenient way to browse and understand a repository, simply copy paste the github link and replace “github” with “deepwiki” in the url. The repository will turn into a detailed report breaking down exactly what is the repo about and let you ask follow up questions on it.

Books

Another medium for you to explore, which might be more accessible, are books. I indeed think books are a great way to learn AI because while research papers are most of the time pretty concise, highly technical and often require a prior domain knowledge, conversely books are often breaking down notions from first principles, take time to gradually build the foundations needed for the reader to follow, without compromising on the depth and expertise. Some great books also usually come with pieces of code which can be a good way for you to follow along in an IDE and some even have exercises. Here are some great sources I would recommend, which you will also find in the final resources section :

There are obviously many more good references, and many which I am not even aware of, but these are my own readings and the ones patiently waiting on my reading list for me to give them some time.

Blog posts / articles

Blog posts and articles are a great format which I enjoy a lot too. They can be very technical as well as covering miscellaneous topics. They are usually a lot more accessible than research papers but less dense than books (even though it depends on which ones). The authors generally adopt a style which is a lot more educational and they usually care about the understanding of their audiences. Here are some blogs and articles along with examples I would recommend you to check:

Youtube videos

Finally a last medium I enjoy is simply youtube videos. Once again, just like X, it all comes down to who you are following. You can find very qualitative content there, just like you can find sloppy “AI experts”… Youtube is great because it is much more entertaining and can also provide a different way to learn. You get to see some great animations, how people work, get testimonies or simply have fun. Here are some great youtube channels I personally enjoy :

more in the resources section…

Building

Building

Maximizing your learning curve

As cool as learning theory can be, it won’t take you that far unless you actually put it into code. One of my favorite ways of learning is by building things from scratch. This philosophy of learning is what Andrej Karpathy is promoting in its GitHub repositories and educational videos, and it has highly influenced me in my way of approaching AI. It’s pretty simple to understand : you are going to code every single thing, character by character, in your IDE. The tokenizer, the attention mechanism, the MLP, the LayerNorm, the training loop and so on. You split your screen in two, one of his videos on one side, your IDE on the other and you follow along. If you are already familiar with your AI fundamentals you can just build it on the first watch. If you are starting fresh; you might want to watch his videos several times: a first time without coding anything, just for you to understand the big picture and what is being built and why (you can prompt your favorite LLM to also help you understand concepts along the way) and then a second pass where you pause, code and repeat. This is an amazing way to learn and understand the concepts from first principles, rent and spin up your first GPU instance, understand how LLMs work in depth and it is very satisfying to get to build your own GPT. For GPU renting here are some options: Lambda, Vast.ai. I personally use and like Vast.ai. The best part is that you also get to have a project to add to your portfolio. You can find more content of this kind from Umar Jamil as well, who works at Mistral AI and adopts the same style of building things from scratch.

Using AI Agents

Even though the resources above are a great place for you to start building, nowadays it is ridiculously easy for you to basically build anything from scratch by using any LLM out there. My favorite way to do so is using an IDE with an AI Agent integrated (e.g I like Antigravity for this use case so I don’t have to jump back and forth between tabs, everything is in one place). You pick something you want to build and simply ask the LLM to guide you from scratch end to end. Again it is your responsibility to make the effort of writing the code character by character, ask follow up questions, dig into concepts you do not fully understand and so on. The AI Agent is helpful because it will go off on its own reading your files, reviewing your work, pinpointing mistakes but if you use it to get all the code written, I believe the educational value is a lot worse, which is the reason why I would discourage you to do so.

To dig a little more on this topic, I see three main “healthy” use cases of AI agents in my workflows :

On “Vibecoding”

Coming back on 3) I would still recommend some best practices while doing so, and here are a few: begin your project with a PRD, split it in epics, user stories and sprints. Make sure these documents are highly detailed and updated along the way of your project so that you can easily come back to it, if you use multiple agents or if you are stuck on something, run out of credits… Build your project as CLI-first so your agent can self verify its own work and close the feedback loop. While vibecoding, do not ask the agent to just go off do everything but proceed step by step, sprint by sprint. At the end of a sprint, ask for all the new features to be tested (unit tests, regression tests yadayada) and refactor the code if necessary. Ask your agent to use several “quality tools” to make sure the code produced respect at least certain standards (Prettier, ESLint, Playwright, Husky, ruff, black, pytest, mypy, pre-commit and so on). Keep the codebase up to date with the documentation, use branches or worktrees if you spin up agents in parallel or subagents. I am not diving in details but there are a few things you can leverage as well by customizing AGENTS.md / CLAUDE.md, skills and MCPs. So far I have enjoyed Codex a lot and especially the app, Claude Code with Anthropic subscription even though too costly for the ridiculous usage provided, and Minimax coding plan paired with Claude Code (you just switch the model in Claude Code), they have a plan for 10 bucks which renew your usage every 5 hours and will never leave you out of credits.

Vibecoding is fun, but the reality is that, on the job market, at least from what I have seen, the diffusion of AI agents is slow and it’s better for you to stick to 1) and 2), at least as a junior, and keep 3) for your fun weekend projects. You have high chances that your current or future workplace will either forbid the use of LLMs for security, IP reasons and so on, take forever to adopt them, have its own in house LLM which is just a chatbot and is lacking behind in coding capabilities. But it could simply be that you do not even get the chance to onboard because you failed their interview when they asked you to implement a PyTorch module using Grouped Query Attention, RoPE and all the good stuff (without access to any LLM obviously). In this case you go back to 1), because you will learn a lot more and make yourself a lot more valuable on the market instead of simply having a blast on 3) but not learning anything.

TLDR : anyone can prompt, only a few can code. Better for you to learn fundamentals.

Collaborating

Collaborating

Another great way of furnishing your portfolio and learn is to collaborate with people on some projects or get some relevant experiences as a student. It can take several forms, but here are a few :

Sharing your contributions

Sharing your contributions

The most common way to share your contributions is to wrap it as a repository on GitHub which can then be part of your portfolio of projects. If you have trained a model or curated a dataset, you can also push it to Hugging Face. This way you will contribute to the amazing world of Open Source and if you get lucky could eventually gain some traction by the community to value your work.

If you are working in research you obviously would want to get published in a conference or a journal and if possible as a first author. I am not going to dive into this as I assume if you are working in research you know that since day 1.

One way to also share your contributions is to create your own website, start blog posting about the projects you have done, sharing these blog posts on social media. You can easily build your own website if you have the skills for, obviously, but at the era of vibecoding and if you don’t care about your web dev skills, you can also get it done fairly easily by either starting fresh with your favourite AI agent or forking a website template you like on GitHub and adapt it. Vercel can be an easy deployment option for your to serve your website and which is the one I personally use. If you want to buy your own domain, I personally use Cloudflare. You can buy your own domain name for roughly 10 bucks a year if you opt for a .com. Lastly linking your projects to your CV is common sense.

Thank you so much for reading all the way till here, I hope it was helpful to you in one way or another. Do not hesitate to reach out if you have any questions, remarks or whatsoever, I would be happy to discuss!

Resources

Newsletters

X (twitter)

AI labs / product orgs

Benchmarking / evaluation / research

Researchers / builders / founders (individuals)

Books

GPU Renting

Blog posts & Articles

Youtube Channels

Discord Servers

Benchmarks

Other Tools & Platforms


Edit on GitHub