All about how to enter Full Stack AI and how to get Cracked at it. [My Guide]
Note: This is not how I learned everything. As a generalist, Iâve often been distracted and explored much of tech, learning a lot by flowing along with my curiosity. Sometimes I dive into DevOps, sometimes cloud, sometimes API integrations, and sometimes business contexts. My path here has been a complete flow of curiosity, with a distracted mind and an obsession to understand things in detail. This is a very rough and quick guide for anyone entering AI (the full stack of analytics, ML, DL, LLMs, and everything DATA) on how they should learn and flow through it all.
- Start with Math: Math is so important. Whatever the language, understand that all the data you see goes into models as numbers. And, as you know, itâs eventually converted to binary code and executed. All data can be changed into numbers, so learn linear algebra, calculus for a deeper understanding, optimization in math, and the most important ones: statistics, probability, and permutations/combinations. If you want to dig even deeper, go for information theory and game theory, so youâll understand how decisions are made in the real world and how everything is interconnected.
- Data & Domain Understanding: Understand how data is structured in the world because AI is all about data. If you donât understand how to use it, everything is useless. By domain, by use case, and by problem, the application of data changes. In finance, fraud data will always be imbalanced, with real fraud cases often below 5% or even lower for larger companiesâthen, what do you do? In eCommerce, there will be thousands of columnsâhow do you decide which ones matter? In domains like healthcare, data is so sparseâwhat do you do then? Always remember, without problems, all these models are useless. Think in terms of problems, data, domains, and use cases rather than just applying models blindly.
- Feature Engineering: This is my favorite part and where youâll spend hours of your job. If data science or finding a solution is a thing, this is literally 70-80% of it. If you mess this up, itâs over. For all the questions I asked in data and domain understanding, your answer will come into play here. How do you fill null values? What do you do when you have less data? How do you handle imbalanced data? How do you optimize data to flow smoothly into models? If you have thousands of columns, which ones do you choose? And if every column is important, how do you reduce the density of the data? From taking raw data to feeding it into the models, everything happens here, and you should be the most skilled in this area if you want to stand out in AI or data.
- Models: Basic heuristic models are completely rule-based, more like chess. Then we get into applied statistics on data, which will tell you how diverse the data is. This is called EDA (Exploratory Data Analysis) or descriptive analytics, where you learn from the data. From there, it moves into supervised tabular ML models, unsupervised tabular ML models, basic RL (Reinforcement Learning) models, DL (Deep Learning) models, CNNs (Convolutional Neural Networks), NLP (Natural Language Processing), CV (Computer Vision), speech processing, word embeddings, and voice-based DL models. Finally, it ends with transformers, where the whole arc and aura of AI changed. Basic ML models are heavily driven by math, which is where explainable ML comes inâitâs easy, intuitive, and interesting to understand. You know why itâs overfitting or underfitting, you have great metrics to validate, and the interpretability is solid. But once you get into deep learning, everything is a black box. Thereâs no interpretability or explainability. You just make educated guesses and try to optimize to get the best accuracy and fitting. With deep learning, you can either dive deep and understand everything or just grasp the basics and leave it at that (unless youâre in research). If youâre into research, the path is different, and you need to go much deeper into the math.
Note: After this point, you have two options: either go into research or apply AI. The line here splits. Many LLMs (Large Language Models) need tons of research, such as validation, architecture, data, pre/post training, and more. The applied side is all about using AI for consumer-facing solutions, fine-tuning, and quick business AI solutions. The future is AI, and both sides are important.
- LLMs, RAGs, and the Current Era: If youâve made it this far, remember that youâll either be an applied person, mastering use cases, problem-solving, and understanding where to apply what model, or a math/research person, diving deep into the architectures, researching deeper math, fine-tuning models, publishing papers, etc. I lean more toward the real-world practical side, and though I understand the deeper stuff, I havenât spent much time in the research space. Remember, LLMs are just next-word predictors on steroids. We donât know how they work in-depth. Transformers can just remember things end-to-end, and with high probability, predict the next word. We donât fully understand the architectures of many LLMs unless theyâre open-sourced. The application is all about making them work for what you need, manipulating them to fit your needs. You canât train an LLM yourselfâyou need heavy GPUs, data, and computing power. So, you adjust its character, its âcostume,â and fine-tune it. The more you dig into fine-tuning, the heavier the compute required. Understanding how they are pre-trained, post-trained, fine-tuned, and how RLHF (Reinforcement Learning with Human Feedback) makes them behave ethically and morally is essential. A lot of manipulation happens with LLMs, and youâre not hitting the core instance itself. The true power of most LLMs comes from tools like interpreters, web search, reasoning, memory, context windows, input/output length, and more.
Note: AI engineering in the space of RAGs (Retrieval-Augmented Generation), LLMs, and Agentic AI is much deeper than you expect. This industry is thriving, and Iâve just scratched the surface. Every day, a new paper, research, or technology pops up, so keep exploring.
System Design & Under the Hood: I believe anything you do, you need to understand how things work from a birdâs-eye view. Thatâs what system design is aboutâunderstanding how each part works. Today, everything is engineering, and I donât think youâll be hired if you canât save a model, create an API out of it to run its inference, and make deployment easy. You need to make your models and systems consumer-ready. Outside of AI, if thereâs one thing I understand well, itâs DevOps, data pipelines, APIs, and cloud systemsâhow big companies handle data, from sourcing to storage, streaming, pipelines, experiments, deployment, and more. Unstructured data is stored differently compared to tabular data, and real-time or batch-based training might be needed for models. Models donât stay the sameâthey evolve, trends shift, and inference accuracy changes.
Constraints in AI Overall:
- Fundamentals Are Everything: AI is contextual, trendy with business, economy, industry, domain, and new techniques, tools, and services are popping up weekly. But if your base and fundamentals are strong, youâve got a solid foundation.
- Validation Is Hard: Validation is tricky in AI. In core ML models, there are plenty of metrics and techniques, but once you get into DL, LLMs, and beyond, it gets complex, and there are no standardized validation metrics. Youâll have to create your own validation points based on your understanding.
- Data Is Everything: If the data you collect is garbage, everything is useless. If your company doesnât automate data collection or source everything digitally, the work you do will often be useless. If consumers are typing manually and it doesnât validate or benefit the data, itâll eventually corrupt your database.
- Curiosity Is Key: If youâre learning for a big paycheck or a stable job, youâre making a mistake. Shift to dev, cloud, or DevOps roles instead. Theyâre more predictable, while AI is a long game.
- Impact Is Real: AI is full of trade-offs, and thatâs why itâs called decision science. The insights you generate can change the whole business strategy.
- Donât Fall Into Resource Traps: Donât get stuck in a loop of tutorials and videos. Until you get your hands dirty, nothing matters. Play with data, try different solutions, and get things done.
What Should You Do Once You Understand Everything? Do projects, go to Kaggle, and youâll learn fast. Kaggle is the best platform for hands-on learningâlearn from others, experiment with different techniques, and get dirty with the data.
Best Resources: Thereâs never just one best resource. Some may be great for you, others may not. Maybe youâll learn better from videos, blogs, hands-on coding, or GitHub repos. Explore different resources, whether videos, blogs, Kaggle, courses, or hands-on practice.
Some resources that might help:
Even after all this, remember, everyoneâs journey is different. Donât follow advice blindlyâdo your own research, especially now with AI. You might get obsessed with one way of learning, and that might change your curve. Stay curious, stay obsessed, and never conclude. Just explore.