I am a USAF Captain working as an AI engineer at the Defense Threat Reduction Agency. My background is in nuclear engineering and physics, not computer science. The path I took into applied AI and ML was neither obvious nor straight, and that experience gives me a perspective that most "how to break into ML" guides lack: I know what it actually looks like to transition into this field from a quantitative but non-CS background, and I know what skills you genuinely need versus what the internet says you need.
This guide is honest. I will not tell you that you can become an ML engineer in 12 weeks. I will tell you what the field actually looks like, what skills separate candidates who get hired from those who do not, and how to build a portfolio that demonstrates real competence rather than tutorial completion. I will also tell you which parts of the standard advice are wrong.
Who this guide is for: Undergraduates in STEM considering an ML career; professionals in engineering, science, or data analysis looking to transition; military officers exploring post-service paths; anyone who wants an honest map of the AI/ML landscape from someone actively working in it.
The Landscape: What AI/ML Jobs Actually Exist
The first mistake most people make is treating "AI/ML" as a single career. It is not. There are at least five distinct roles, each requiring a different skill profile:
| Role | Core Skills | Output |
|---|---|---|
| ML Engineer | Python, ML frameworks, software engineering, MLOps | Production ML systems |
| Research Scientist | Deep math, paper writing, experimental design | Novel algorithms and architectures |
| Data Scientist | Statistics, SQL, Python, business communication | Insights from data, predictive models |
| Applied Scientist | Blend of research and engineering | Adapt research to product features |
| AI/ML Analyst | Domain expertise + ML literacy | ML-informed decisions in specific domain |
In the defense and national security context where I work, there is a sixth category that the private sector rarely discusses: AI systems integrator — someone who understands both the technical capabilities of ML systems and the operational requirements of defense applications. This role is in massive demand and severely undersupplied. If you have a STEM background and a security clearance or a path to one, this is one of the most valuable career positions in the current market.
The Skills That Actually Matter
The internet generates endless debates about which ML framework to learn, which certification to get, and whether you need a PhD. Here is the cleaner answer: there is a skill stack that virtually every ML practitioner needs, and within that stack there is a clear learning order.
Foundation (Non-Negotiable)
- Python fluency
- Linear algebra (vectors, matrices, eigenvalues)
- Calculus (derivatives, gradients)
- Probability and statistics
- Git and version control
Core ML Skills
- Supervised learning algorithms
- Model evaluation (cross-validation, metrics)
- Feature engineering
- Scikit-learn proficiency
- Neural networks and backprop
Deep Learning Stack
- PyTorch or TensorFlow
- CNNs, RNNs, Transformers
- Transfer learning and fine-tuning
- GPU training basics
- Model debugging and optimization
Production and MLOps
- Model deployment (APIs, containers)
- Data pipelines
- Model monitoring
- Experiment tracking (MLflow, W&B)
- Cloud platforms (AWS, GCP, Azure)
The honest sequence: master the foundation before touching ML algorithms. Master ML algorithms before touching deep learning. Most people try to skip directly to deep learning and then wonder why they cannot debug their models or explain their results. The foundation is not optional.
The Math You Need (And the Math You Do Not)
Calculus I and II: yes, essential. You need to understand gradients conceptually and be able to follow derivations in papers. Multivariable calculus: useful, especially for understanding optimization landscapes. Linear algebra: absolutely critical. If there is one math course to prioritize, it is linear algebra — matrices, eigenvectors, and linear transformations are the language of ML. Probability and statistics: essential, and frequently undertaught in ML bootcamps. You cannot evaluate a model properly without understanding bias, variance, confidence intervals, and hypothesis testing.
What you do not need before your first job: measure theory, advanced functional analysis, or graduate-level real analysis. These become relevant if you pursue a research career, but they are not prerequisites for ML engineering or data science roles.
The Learning Path: A Phase-by-Phase Approach
Python and Math Foundations
Get Python to a comfortable intermediate level: data types, functions, classes, list comprehensions, NumPy, and Pandas. Simultaneously work through a linear algebra course (3Blue1Brown's Essence of Linear Algebra on YouTube is exceptional and free). Start probability with a solid introductory textbook.
- Build two small Python projects: a data analysis script and a data visualization notebook
- Work through Khan Academy's linear algebra section completely
- Write at least 500 lines of Python that you debugged yourself
Classical Machine Learning
Work through a comprehensive ML book (Géron's Hands-On Machine Learning covers this well, chapters 1-9). Implement every example yourself — do not copy-paste. Participate in two Kaggle competitions: one tabular data competition and one classification competition. The goal is not to win but to read other participants' notebooks and understand techniques you have not encountered.
- Build a complete ML pipeline from data loading through model evaluation and reporting
- Implement linear regression and logistic regression from scratch (without scikit-learn) at least once
- Learn cross-validation, hyperparameter tuning, and model selection properly
Deep Learning
Work through deep learning content systematically — Géron's chapters 10-18 and Andrej Karpathy's neural network series on YouTube (building language models from scratch) are both excellent. Learn PyTorch as your primary framework. Train models on real datasets. Get GPU access through Google Colab, Kaggle Notebooks, or a cloud provider — GPU training experience is explicitly required in most ML engineering job descriptions.
- Fine-tune a pre-trained image classifier on a custom dataset you assembled yourself
- Build a text classification system using a pre-trained language model
- Reproduce a published ML result from a paper, even a relatively simple one
Portfolio, Community, and Job Search
Two or three strong portfolio projects outperform a dozen weak ones. Each project should have a GitHub repository with clean code, a clear README, and ideally a deployed demo. Write at least one technical blog post explaining your most interesting project. Technical writing demonstrates communication skills, which hiring managers weight heavily alongside coding ability.
- Deploy at least one model as a live web application (Gradio, Streamlit, or a FastAPI endpoint)
- Contribute to an open-source ML project, even documentation or a bug fix
- Engage with the ML community on LinkedIn or relevant Discord servers in your target domain
Portfolio Projects That Actually Impress
The weakest portfolio projects are tutorial reproductions: MNIST digit classification, iris species prediction, Titanic survival analysis. Hiring managers have seen these thousands of times. They demonstrate that you can follow instructions, not that you can solve novel problems.
The strongest portfolio projects share three properties. First, they use a dataset you assembled yourself or one that is not in every beginner tutorial. Second, they solve a problem someone actually cares about. Third, they involve at least one non-obvious technical decision that you can explain and defend in an interview.
Strong examples from my experience reviewing candidates: an NLP model that classifies scientific abstracts by methodology type; a computer vision system that detects equipment anomalies in photos taken by non-specialists; a time-series model that predicts energy consumption from building sensor data; a recommendation system for a niche domain such as scientific papers or specialized technical equipment. None of these require a massive dataset or exotic techniques — they require applying standard techniques thoughtfully to a real problem.
The credentialing question: Do you need a master's degree or PhD to get an ML job? For research scientist roles at top labs: almost certainly yes. For ML engineering at most companies: no. A strong portfolio, demonstrable Python skills, and relevant work experience will get you interviews at a wide range of companies. The degree accelerates the path and opens certain doors, but it is not the only path in.
The Job Market: What I See From the Inside
The AI job market in 2026 is unusual. The total number of ML positions has grown dramatically, but the distribution of hiring has become more bifurcated. Entry-level ML positions are competitive because of the large supply of bootcamp graduates with similar surface-level credentials. Senior ML engineering roles and applied scientist positions remain severely undersupplied.
The skills most undersupplied relative to demand: ML systems engineering (deploying and maintaining models at scale in production), ML applied to scientific and engineering domains (physics, chemistry, materials science, biology), applied ML in regulated industries (defense, healthcare, energy), and the combination of strong domain expertise with ML competence. If you come from engineering, physics, or another STEM field, your domain knowledge is a genuine competitive advantage. Do not try to hide it in favor of presenting yourself purely as a software engineer who learned ML.
On the Government and Defense AI Sector
The United States Department of Defense and affiliated agencies represent one of the largest and fastest-growing markets for AI talent. Working in this space requires a security clearance, which takes time to obtain but is achievable for qualified candidates. The compensation is not always competitive with private-sector technology companies, but the problems are uniquely impactful and the mission clarity is high. For someone who wants to work on consequential AI applications with real-world stakes, this sector is worth serious consideration alongside commercial options.
The Advice I Would Have Given Myself
Build things constantly. Not tutorial projects — real things that answer questions you are genuinely curious about. The fastest way to develop deep ML intuition is to get stuck on a real problem, debug it systematically, and figure out why your model is behaving unexpectedly. Coursework gives you the vocabulary; building and breaking things gives you the understanding.
Read papers early and often. You do not need to understand every mathematical detail on a first read. Learn to skim a paper efficiently: abstract, introduction, figures, conclusion. Understand the claim the authors are making and whether it is plausible. Then go back and read the methods. Paper reading is a skill that improves with deliberate practice, and it separates practitioners who are current with the field from those working with techniques that are three years out of date.
Do not confuse tool familiarity with conceptual depth. Many people who can run a PyTorch training loop cannot explain what gradient descent is actually doing, why their loss is not decreasing, or how to diagnose overfitting from a learning curve. The tools are the surface; the concepts are the foundation. A hiring manager who knows ML will test the concepts, not just whether you have used the framework.
Want to Navigate the AI/ML Career Path With Expert Guidance?
I offer one-on-one mentoring and tutoring in AI and machine learning, from mathematical foundations through career preparation. My experience as a working USAF AI engineer means I focus on what actually matters in the field. Book a free intro call.
Book Free 30-Min Intro →Weekly AI Career and Learning Insights
Join 500+ students and professionals receiving Dr. Preston's weekly breakdown of the AI field, learning resources, and honest career guidance.
Subscribe Free →