Этот материал находится в платной подписке. Оформи премиум подписку и смотри или слушай PyTorch for Deep Learning, а также все другие курсы, прямо сейчас!
Премиум
  • Урок 1. 00:03:34
    PyTorch for Deep Learning
  • Урок 2. 00:05:54
    Course Welcome and What Is Deep Learning
  • Урок 3. 00:03:34
    Why Use Machine Learning or Deep Learning
  • Урок 4. 00:05:40
    The Number 1 Rule of Machine Learning and What Is Deep Learning Good For
  • Урок 5. 00:06:07
    Machine Learning vs. Deep Learning
  • Урок 6. 00:09:22
    Anatomy of Neural Networks
  • Урок 7. 00:04:31
    Different Types of Learning Paradigms
  • Урок 8. 00:06:22
    What Can Deep Learning Be Used For
  • Урок 9. 00:10:13
    What Is and Why PyTorch
  • Урок 10. 00:04:16
    What Are Tensors
  • Урок 11. 00:06:06
    What We Are Going To Cover With PyTorch
  • Урок 12. 00:05:10
    How To and How Not To Approach This Course
  • Урок 13. 00:05:22
    Important Resources For This Course
  • Урок 14. 00:07:40
    Getting Setup to Write PyTorch Code
  • Урок 15. 00:13:26
    Introduction to PyTorch Tensors
  • Урок 16. 00:09:59
    Creating Random Tensors in PyTorch
  • Урок 17. 00:03:09
    Creating Tensors With Zeros and Ones in PyTorch
  • Урок 18. 00:05:18
    Creating a Tensor Range and Tensors Like Other Tensors
  • Урок 19. 00:09:25
    Dealing With Tensor Data Types
  • Урок 20. 00:08:23
    Getting Tensor Attributes
  • Урок 21. 00:06:00
    Manipulating Tensors (Tensor Operations)
  • Урок 22. 00:09:35
    Matrix Multiplication (Part 1)
  • Урок 23. 00:07:52
    Matrix Multiplication (Part 2): The Two Main Rules of Matrix Multiplication
  • Урок 24. 00:12:58
    Matrix Multiplication (Part 3): Dealing With Tensor Shape Errors
  • Урок 25. 00:06:10
    Finding the Min Max Mean and Sum of Tensors (Tensor Aggregation)
  • Урок 26. 00:03:17
    Finding The Positional Min and Max of Tensors
  • Урок 27. 00:13:41
    Reshaping, Viewing and Stacking Tensors
  • Урок 28. 00:11:56
    Squeezing, Unsqueezing and Permuting Tensors
  • Урок 29. 00:09:32
    Selecting Data From Tensors (Indexing)
  • Урок 30. 00:09:09
    PyTorch Tensors and NumPy
  • Урок 31. 00:10:47
    PyTorch Reproducibility (Taking the Random Out of Random)
  • Урок 32. 00:11:51
    Different Ways of Accessing a GPU in PyTorch
  • Урок 33. 00:07:44
    Setting up Device Agnostic Code and Putting Tensors On and Off the GPU
  • Урок 34. 00:04:50
    PyTorch Fundamentals: Exercises and Extra-Curriculum
  • Урок 35. 00:02:46
    Introduction and Where You Can Get Help
  • Урок 36. 00:07:15
    Getting Setup and What We Are Covering
  • Урок 37. 00:09:42
    Creating a Simple Dataset Using the Linear Regression Formula
  • Урок 38. 00:08:21
    Splitting Our Data Into Training and Test Sets
  • Урок 39. 00:07:46
    Building a function to Visualize Our Data
  • Урок 40. 00:14:10
    Creating Our First PyTorch Model for Linear Regression
  • Урок 41. 00:06:11
    Breaking Down What's Happening in Our PyTorch Linear regression Model
  • Урок 42. 00:06:27
    Discussing Some of the Most Important PyTorch Model Building Classes
  • Урок 43. 00:09:51
    Checking Out the Internals of Our PyTorch Model
  • Урок 44. 00:11:13
    Making Predictions With Our Random Model Using Inference Mode
  • Урок 45. 00:08:15
    Training a Model Intuition (The Things We Need)
  • Урок 46. 00:12:52
    Setting Up an Optimizer and a Loss Function
  • Урок 47. 00:13:54
    PyTorch Training Loop Steps and Intuition
  • Урок 48. 00:08:47
    Writing Code for a PyTorch Training Loop
  • Урок 49. 00:14:58
    Reviewing the Steps in a Training Loop Step by Step
  • Урок 50. 00:09:26
    Running Our Training Loop Epoch by Epoch and Seeing What Happens
  • Урок 51. 00:11:38
    Writing Testing Loop Code and Discussing What's Happening Step by Step
  • Урок 52. 00:14:43
    Reviewing What Happens in a Testing Loop Step by Step
  • Урок 53. 00:13:46
    Writing Code to Save a PyTorch Model
  • Урок 54. 00:08:45
    Writing Code to Load a PyTorch Model
  • Урок 55. 00:06:03
    Setting Up to Practice Everything We Have Done Using Device-Agnostic Code
  • Урок 56. 00:06:09
    Putting Everything Together (Part 1): Data
  • Урок 57. 00:10:08
    Putting Everything Together (Part 2): Building a Model
  • Урок 58. 00:12:41
    Putting Everything Together (Part 3): Training a Model
  • Урок 59. 00:05:18
    Putting Everything Together (Part 4): Making Predictions With a Trained Model
  • Урок 60. 00:09:11
    Putting Everything Together (Part 5): Saving and Loading a Trained Model
  • Урок 61. 00:02:57
    Exercise: Imposter Syndrome
  • Урок 62. 00:03:58
    PyTorch Workflow: Exercises and Extra-Curriculum
  • Урок 63. 00:09:42
    Introduction to Machine Learning Classification With PyTorch
  • Урок 64. 00:09:08
    Classification Problem Example: Input and Output Shapes
  • Урок 65. 00:06:32
    Typical Architecture of a Classification Neural Network (Overview)
  • Урок 66. 00:12:19
    Making a Toy Classification Dataset
  • Урок 67. 00:11:56
    Turning Our Data into Tensors and Making a Training and Test Split
  • Урок 68. 00:04:20
    Laying Out Steps for Modelling and Setting Up Device-Agnostic Code
  • Урок 69. 00:10:58
    Coding a Small Neural Network to Handle Our Classification Data
  • Урок 70. 00:06:58
    Making Our Neural Network Visual
  • Урок 71. 00:13:18
    Recreating and Exploring the Insides of Our Model Using nn.Sequential
  • Урок 72. 00:14:51
    Setting Up a Loss Function Optimizer and Evaluation Function for Our Classification Network
  • Урок 73. 00:16:08
    Going from Model Logits to Prediction Probabilities to Prediction Labels
  • Урок 74. 00:15:27
    Coding a Training and Testing Optimization Loop for Our Classification Model
  • Урок 75. 00:14:14
    Writing Code to Download a Helper Function to Visualize Our Models Predictions
  • Урок 76. 00:08:03
    Discussing Options to Improve a Model
  • Урок 77. 00:09:07
    Creating a New Model with More Layers and Hidden Units
  • Урок 78. 00:12:46
    Writing Training and Testing Code to See if Our New and Upgraded Model Performs Better
  • Урок 79. 00:08:08
    Creating a Straight Line Dataset to See if Our Model is Learning Anything
  • Урок 80. 00:10:02
    Building and Training a Model to Fit on Straight Line Data
  • Урок 81. 00:05:24
    Evaluating Our Models Predictions on Straight Line Data
  • Урок 82. 00:10:01
    Introducing the Missing Piece for Our Classification Model Non-Linearity
  • Урок 83. 00:10:26
    Building Our First Neural Network with Non-Linearity
  • Урок 84. 00:15:13
    Writing Training and Testing Code for Our First Non-Linear Model
  • Урок 85. 00:05:48
    Making Predictions with and Evaluating Our First Non-Linear Model
  • Урок 86. 00:09:35
    Replicating Non-Linear Activation Functions with Pure PyTorch
  • Урок 87. 00:11:25
    Putting It All Together (Part 1): Building a Multiclass Dataset
  • Урок 88. 00:12:28
    Creating a Multi-Class Classification Model with PyTorch
  • Урок 89. 00:06:41
    Setting Up a Loss Function and Optimizer for Our Multi-Class Model
  • Урок 90. 00:11:03
    Going from Logits to Prediction Probabilities to Prediction Labels with a Multi-Class Model
  • Урок 91. 00:16:18
    Training a Multi-Class Classification Model and Troubleshooting Code on the Fly
  • Урок 92. 00:08:00
    Making Predictions with and Evaluating Our Multi-Class Classification Model
  • Урок 93. 00:09:18
    Discussing a Few More Classification Metrics
  • Урок 94. 00:02:59
    PyTorch Classification: Exercises and Extra-Curriculum
  • Урок 95. 00:11:48
    What Is a Computer Vision Problem and What We Are Going to Cover
  • Урок 96. 00:10:09
    Computer Vision Input and Output Shapes
  • Урок 97. 00:05:03
    What Is a Convolutional Neural Network (CNN)
  • Урок 98. 00:09:20
    Discussing and Importing the Base Computer Vision Libraries in PyTorch
  • Урок 99. 00:14:31
    Getting a Computer Vision Dataset and Checking Out Its- Input and Output Shapes
  • Урок 100. 00:09:52
    Visualizing Random Samples of Data
  • Урок 101. 00:07:18
    DataLoader Overview Understanding Mini-Batch
  • Урок 102. 00:12:24
    Turning Our Datasets Into DataLoaders
  • Урок 103. 00:14:39
    Model 0: Creating a Baseline Model with Two Linear Layers
  • Урок 104. 00:10:30
    Creating a Loss Function: an Optimizer for Model 0
  • Урок 105. 00:05:35
    Creating a Function to Time Our Modelling Code
  • Урок 106. 00:21:26
    Writing Training and Testing Loops for Our Batched Data
  • Урок 107. 00:12:59
    Writing an Evaluation Function to Get Our Models Results
  • Урок 108. 00:03:47
    Setup Device-Agnostic Code for Running Experiments on the GPU
  • Урок 109. 00:09:04
    Model 1: Creating a Model with Non-Linear Functions
  • Урок 110. 00:03:05
    Mode 1: Creating a Loss Function and Optimizer
  • Урок 111. 00:08:29
    Turing Our Training Loop into a Function
  • Урок 112. 00:06:36
    Turing Our Testing Loop into a Function
  • Урок 113. 00:11:53
    Training and Testing Model 1 with Our Training and Testing Functions
  • Урок 114. 00:04:09
    Getting a Results Dictionary for Model 1
  • Урок 115. 00:08:25
    Model 2: Convolutional Neural Networks High Level Overview
  • Урок 116. 00:19:49
    Model 2: Coding Our First Convolutional Neural Network with PyTorch
  • Урок 117. 00:15:00
    Model 2: Breaking Down Conv2D Step by Step
  • Урок 118. 00:15:49
    Model 2: Breaking Down MaxPool2D Step by Step
  • Урок 119. 00:13:46
    Mode 2: Using a Trick to Find the Input and Output Shapes of Each of Our Layers
  • Урок 120. 00:02:39
    Model 2: Setting Up a Loss Function and Optimizer
  • Урок 121. 00:07:55
    Model 2: Training Our First CNN and Evaluating Its Results
  • Урок 122. 00:07:24
    Comparing the Results of Our Modelling Experiments
  • Урок 123. 00:11:40
    Making Predictions on Random Test Samples with the Best Trained Model
  • Урок 124. 00:08:11
    Plotting Our Best Model Predictions on Random Test Samples and Evaluating Them
  • Урок 125. 00:15:21
    Making Predictions Across the Whole Test Dataset and Importing Libraries to Plot a Confusion Matrix
  • Урок 126. 00:06:55
    Evaluating Our Best Models Predictions with a Confusion Matrix
  • Урок 127. 00:11:28
    Saving and Loading Our Best Performing Model
  • Урок 128. 00:06:02
    Recapping What We Have Covered Plus Exercises and Extra-Curriculum
  • Урок 129. 00:09:54
    What Is a Custom Dataset and What We Are Going to Cover
  • Урок 130. 00:05:55
    Importing PyTorch and Setting Up Device-Agnostic Code
  • Урок 131. 00:14:05
    Downloading a Custom Dataset of Pizza, Steak and Sushi Images
  • Урок 132. 00:08:42
    Becoming One With the Data (Part 1): Exploring the Data Format
  • Урок 133. 00:11:41
    Becoming One With the Data (Part 2): Visualizing a Random Image
  • Урок 134. 00:04:48
    Becoming One With the Data (Part 3): Visualizing a Random Image with Matplotlib
  • Урок 135. 00:08:54
    Transforming Data (Part 1): Turning Images Into Tensors
  • Урок 136. 00:11:31
    Transforming Data (Part 2): Visualizing Transformed Images
  • Урок 137. 00:09:19
    Loading All of Our Images and Turning Them Into Tensors With ImageFolder
  • Урок 138. 00:07:19
    Visualizing a Loaded Image From the Train Dataset
  • Урок 139. 00:09:04
    Turning Our Image Datasets into PyTorch DataLoaders
  • Урок 140. 00:08:01
    Creating a Custom Dataset Class in PyTorch High Level Overview
  • Урок 141. 00:09:07
    Creating a Helper Function to Get Class Names From a Directory
  • Урок 142. 00:17:47
    Writing a PyTorch Custom Dataset Class from Scratch to Load Our Images
  • Урок 143. 00:07:14
    Compare Our Custom Dataset Class to the Original ImageFolder Class
  • Урок 144. 00:14:19
    Writing a Helper Function to Visualize Random Images from Our Custom Dataset
  • Урок 145. 00:07:00
    Turning Our Custom Datasets Into DataLoaders
  • Урок 146. 00:14:24
    Exploring State of the Art Data Augmentation With Torchvision Transforms
  • Урок 147. 00:08:16
    Building a Baseline Model (Part 1): Loading and Transforming Data
  • Урок 148. 00:11:25
    Building a Baseline Model (Part 2): Replicating Tiny VGG from Scratch
  • Урок 149. 00:08:10
    Building a Baseline Model (Part 3): Doing a Forward Pass to Test Our Model Shapes
  • Урок 150. 00:06:39
    Using the Torchinfo Package to Get a Summary of Our Model
  • Урок 151. 00:13:04
    Creating Training and Testing loop Functions
  • Урок 152. 00:10:15
    Creating a Train Function to Train and Evaluate Our Models
  • Урок 153. 00:09:54
    Training and Evaluating Model 0 With Our Training Functions
  • Урок 154. 00:09:03
    Plotting the Loss Curves of Model 0
  • Урок 155. 00:14:14
    Discussing the Balance Between Overfitting and Underfitting and How to Deal With Each
  • Урок 156. 00:11:04
    Creating Augmented Training Datasets and DataLoaders for Model 1
  • Урок 157. 00:07:11
    Constructing and Training Model 1
  • Урок 158. 00:03:23
    Plotting the Loss Curves of Model 1
  • Урок 159. 00:10:56
    Plotting the Loss Curves of All of Our Models Against Each Other
  • Урок 160. 00:05:33
    Predicting on Custom Data (Part 1): Downloading an Image
  • Урок 161. 00:07:01
    Predicting on Custom Data (Part2): Loading In a Custom Image With PyTorch
  • Урок 162. 00:14:08
    Predicting on Custom Data (Part 3): Getting Our Custom Image Into the Right Format
  • Урок 163. 00:04:25
    Predicting on Custom Data (Part 4): Turning Our Models Raw Outputs Into Prediction Labels
  • Урок 164. 00:12:48
    Predicting on Custom Data (Part 5): Putting It All Together
  • Урок 165. 00:06:05
    Summary of What We Have Covered Plus Exercises and Extra-Curriculum
  • Урок 166. 00:11:35
    What Is Going Modular and What We Are Going to Cover
  • Урок 167. 00:07:41
    Going Modular Notebook (Part 1): Running It End to End
  • Урок 168. 00:04:51
    Downloading a Dataset
  • Урок 169. 00:13:51
    Writing the Outline for Our First Python Script to Setup the Data
  • Урок 170. 00:10:36
    Creating a Python Script to Create Our PyTorch DataLoaders
  • Урок 171. 00:09:19
    Turning Our Model Building Code into a Python Script
  • Урок 172. 00:06:17
    Turning Our Model Training Code into a Python Script
  • Урок 173. 00:06:08
    Turning Our Utility Function to Save a Model into a Python Script
  • Урок 174. 00:15:47
    Creating a Training Script to Train Our Model in One Line of Code
  • Урок 175. 00:06:00
    Going Modular: Summary, Exercises and Extra-Curriculum
  • Урок 176. 00:10:06
    Introduction: What is Transfer Learning and Why Use It
  • Урок 177. 00:05:13
    Where Can You Find Pretrained Models and What We Are Going to Cover
  • Урок 178. 00:08:06
    Installing the Latest Versions of Torch and Torchvision
  • Урок 179. 00:06:42
    Downloading Our Previously Written Code from Going Modular
  • Урок 180. 00:08:01
    Downloading Pizza, Steak, Sushi Image Data from Github
  • Урок 181. 00:14:41
    Turning Our Data into DataLoaders with Manually Created Transforms
  • Урок 182. 00:13:07
    Turning Our Data into DataLoaders with Automatic Created Transforms
  • Урок 183. 00:12:16
    Which Pretrained Model Should You Use
  • Урок 184. 00:10:57
    Setting Up a Pretrained Model with Torchvision
  • Урок 185. 00:07:12
    Different Kinds of Transfer Learning
  • Урок 186. 00:06:50
    Getting a Summary of the Different Layers of Our Model
  • Урок 187. 00:13:27
    Freezing the Base Layers of Our Model and Updating the Classifier Head
  • Урок 188. 00:07:55
    Training Our First Transfer Learning Feature Extractor Model
  • Урок 189. 00:06:27
    Plotting the Loss Curves of Our Transfer Learning Model
  • Урок 190. 00:07:58
    Outlining the Steps to Make Predictions on the Test Images
  • Урок 191. 00:10:01
    Creating a Function Predict On and Plot Images
  • Урок 192. 00:07:24
    Making and Plotting Predictions on Test Images
  • Урок 193. 00:06:22
    Making a Prediction on a Custom Image
  • Урок 194. 00:03:22
    Main Takeaways, Exercises and Extra Curriculum
  • Урок 195. 00:07:07
    What Is Experiment Tracking and Why Track Experiments
  • Урок 196. 00:08:14
    Getting Setup by Importing Torch Libraries and Going Modular Code
  • Урок 197. 00:10:24
    Creating a Function to Download Data
  • Урок 198. 00:08:31
    Turning Our Data into DataLoaders Using Manual Transforms
  • Урок 199. 00:07:48
    Turning Our Data into DataLoaders Using Automatic Transforms
  • Урок 200. 00:10:29
    Preparing a Pretrained Model for Our Own Problem
  • Урок 201. 00:13:36
    Setting Up a Way to Track a Single Model Experiment with TensorBoard
  • Урок 202. 00:04:39
    Training a Single Model and Saving the Results to TensorBoard
  • Урок 203. 00:10:18
    Exploring Our Single Models Results with TensorBoard
  • Урок 204. 00:10:45
    Creating a Function to Create SummaryWriter Instances
  • Урок 205. 00:04:58
    Adapting Our Train Function to Be Able to Track Multiple Experiments
  • Урок 206. 00:06:00
    What Experiments Should You Try
  • Урок 207. 00:06:02
    Discussing the Experiments We Are Going to Try
  • Урок 208. 00:06:32
    Downloading Datasets for Our Modelling Experiments
  • Урок 209. 00:08:29
    Turning Our Datasets into DataLoaders Ready for Experimentation
  • Урок 210. 00:15:55
    Creating Functions to Prepare Our Feature Extractor Models
  • Урок 211. 00:14:28
    Coding Out the Steps to Run a Series of Modelling Experiments
  • Урок 212. 00:03:51
    Running Eight Different Modelling Experiments in 5 Minutes
  • Урок 213. 00:13:39
    Viewing Our Modelling Experiments in TensorBoard
  • Урок 214. 00:10:33
    Loading In the Best Model and Making Predictions on Random Images from the Test Set
  • Урок 215. 00:03:45
    Making a Prediction on Our Own Custom Image with the Best Model
  • Урок 216. 00:03:57
    Main Takeaways, Exercises and Extra Curriculum
  • Урок 217. 00:07:35
    What Is a Machine Learning Research Paper?
  • Урок 218. 00:03:14
    Why Replicate a Machine Learning Research Paper?
  • Урок 219. 00:08:19
    Where Can You Find Machine Learning Research Papers and Code?
  • Урок 220. 00:08:22
    What We Are Going to Cover
  • Урок 221. 00:08:22
    Getting Setup for Coding in Google Colab
  • Урок 222. 00:04:03
    Downloading Data for Food Vision Mini
  • Урок 223. 00:09:48
    Turning Our Food Vision Mini Images into PyTorch DataLoaders
  • Урок 224. 00:03:46
    Visualizing a Single Image
  • Урок 225. 00:09:54
    Replicating a Vision Transformer - High Level Overview
  • Урок 226. 00:11:13
    Breaking Down Figure 1 of the ViT Paper
  • Урок 227. 00:10:56
    Breaking Down the Four Equations Overview and a Trick for Reading Papers
  • Урок 228. 00:08:15
    Breaking Down Equation 1
  • Урок 229. 00:10:04
    Breaking Down Equations 2 and 3
  • Урок 230. 00:07:28
    Breaking Down Equation 4
  • Урок 231. 00:11:06
    Breaking Down Table 1
  • Урок 232. 00:15:42
    Calculating the Input and Output Shape of the Embedding Layer by Hand
  • Урок 233. 00:15:04
    Turning a Single Image into Patches (Part 1: Patching the Top Row)
  • Урок 234. 00:12:34
    Turning a Single Image into Patches (Part 2: Patching the Entire Image)
  • Урок 235. 00:13:34
    Creating Patch Embeddings with a Convolutional Layer
  • Урок 236. 00:12:55
    Exploring the Outputs of Our Convolutional Patch Embedding Layer
  • Урок 237. 00:10:00
    Flattening Our Convolutional Feature Maps into a Sequence of Patch Embeddings
  • Урок 238. 00:05:04
    Visualizing a Single Sequence Vector of Patch Embeddings
  • Урок 239. 00:17:02
    Creating the Patch Embedding Layer with PyTorch
  • Урок 240. 00:13:25
    Creating the Class Token Embedding
  • Урок 241. 00:13:25
    Creating the Class Token Embedding - Less Birds
  • Урок 242. 00:11:26
    Creating the Position Embedding
  • Урок 243. 00:13:26
    Equation 1: Putting it All Together
  • Урок 244. 00:14:31
    Equation 2: Multihead Attention Overview
  • Урок 245. 00:09:04
    Equation 2: Layernorm Overview
  • Урок 246. 00:14:34
    Turning Equation 2 into Code
  • Урок 247. 00:05:41
    Checking the Inputs and Outputs of Equation
  • Урок 248. 00:09:12
    Equation 3: Replication Overview
  • Урок 249. 00:11:26
    Turning Equation 3 into Code
  • Урок 250. 00:08:51
    Transformer Encoder Overview
  • Урок 251. 00:09:17
    Combining Equation 2 and 3 to Create the Transformer Encoder
  • Урок 252. 00:15:55
    Creating a Transformer Encoder Layer with In-Built PyTorch Layer
  • Урок 253. 00:18:20
    Bringing Our Own Vision Transformer to Life - Part 1: Gathering the Pieces of the Puzzle
  • Урок 254. 00:10:42
    Bringing Our Own Vision Transformer to Life - Part 2: Putting Together the Forward Method
  • Урок 255. 00:07:14
    Getting a Visual Summary of Our Custom Vision Transformer
  • Урок 256. 00:11:27
    Creating a Loss Function and Optimizer from the ViT Paper
  • Урок 257. 00:04:30
    Training our Custom ViT on Food Vision Mini
  • Урок 258. 00:09:09
    Discussing what Our Training Setup Is Missing
  • Урок 259. 00:06:14
    Plotting a Loss Curve for Our ViT Model
  • Урок 260. 00:14:38
    Getting a Pretrained Vision Transformer from Torchvision and Setting it Up
  • Урок 261. 00:05:54
    Preparing Data to Be Used with a Pretrained ViT
  • Урок 262. 00:07:16
    Training a Pretrained ViT Feature Extractor Model for Food Vision Mini
  • Урок 263. 00:05:14
    Saving Our Pretrained ViT Model to File and Inspecting Its Size
  • Урок 264. 00:03:47
    Discussing the Trade-Offs Between Using a Larger Model for Deployments
  • Урок 265. 00:03:31
    Making Predictions on a Custom Image with Our Pretrained ViT
  • Урок 266. 00:06:51
    PyTorch Paper Replicating: Main Takeaways, Exercises and Extra-Curriculum
  • Урок 267. 00:09:36
    What is Machine Learning Model Deployment and Why Deploy a Machine Learning Model
  • Урок 268. 00:07:14
    Three Questions to Ask for Machine Learning Model Deployment
  • Урок 269. 00:13:35
    Where Is My Model Going to Go?
  • Урок 270. 00:08:00
    How Is My Model Going to Function?
  • Урок 271. 00:05:50
    Some Tools and Places to Deploy Machine Learning Models
  • Урок 272. 00:04:02
    What We Are Going to Cover
  • Урок 273. 00:06:16
    Getting Setup to Code
  • Урок 274. 00:03:24
    Downloading a Dataset for Food Vision Mini
  • Урок 275. 00:08:00
    Outlining Our Food Vision Mini Deployment Goals and Modelling Experiments
  • Урок 276. 00:09:46
    Creating an EffNetB2 Feature Extractor Model
  • Урок 277. 00:06:30
    Create a Function to Make an EffNetB2 Feature Extractor Model and Transforms
  • Урок 278. 00:03:32
    Creating DataLoaders for EffNetB2
  • Урок 279. 00:09:16
    Training Our EffNetB2 Feature Extractor and Inspecting the Loss Curves
  • Урок 280. 00:03:25
    Saving Our EffNetB2 Model to File
  • Урок 281. 00:05:52
    Getting the Size of Our EffNetB2 Model in Megabytes
  • Урок 282. 00:06:35
    Collecting Important Statistics and Performance Metrics for Our EffNetB2 Model
  • Урок 283. 00:07:52
    Creating a Vision Transformer Feature Extractor Model
  • Урок 284. 00:02:31
    Creating DataLoaders for Our ViT Feature Extractor Model
  • Урок 285. 00:06:20
    Training Our ViT Feature Extractor Model and Inspecting Its Loss Curves
  • Урок 286. 00:05:09
    Saving Our ViT Feature Extractor and Inspecting Its Size
  • Урок 287. 00:05:52
    Collecting Stats About Our ViT Feature Extractor
  • Урок 288. 00:11:16
    Outlining the Steps for Making and Timing Predictions for Our Models
  • Урок 289. 00:16:21
    Creating a Function to Make and Time Predictions with Our Models
  • Урок 290. 00:10:44
    Making and Timing Predictions with EffNetB2
  • Урок 291. 00:07:35
    Making and Timing Predictions with ViT
  • Урок 292. 00:11:32
    Comparing EffNetB2 and ViT Model Statistics
  • Урок 293. 00:15:55
    Visualizing the Performance vs Speed Trade-off
  • Урок 294. 00:08:40
    Gradio Overview and Installation
  • Урок 295. 00:08:50
    Gradio Function Outline
  • Урок 296. 00:09:52
    Creating a Predict Function to Map Our Food Vision Mini Inputs to Outputs
  • Урок 297. 00:05:27
    Creating a List of Examples to Pass to Our Gradio Demo
  • Урок 298. 00:12:13
    Bringing Food Vision Mini to Life in a Live Web Application
  • Урок 299. 00:06:27
    Getting Ready to Deploy Our App Hugging Face Spaces Overview
  • Урок 300. 00:08:12
    Outlining the File Structure of Our Deployed App
  • Урок 301. 00:04:12
    Creating a Food Vision Mini Demo Directory to House Our App Files
  • Урок 302. 00:09:14
    Creating an Examples Directory with Example Food Vision Mini Images
  • Урок 303. 00:07:43
    Writing Code to Move Our Saved EffNetB2 Model File
  • Урок 304. 00:04:02
    Turning Our EffNetB2 Model Creation Function Into a Python Script
  • Урок 305. 00:13:28
    Turning Our Food Vision Mini Demo App Into a Python Script
  • Урок 306. 00:04:12
    Creating a Requirements File for Our Food Vision Mini App
  • Урок 307. 00:11:31
    Downloading Our Food Vision Mini App Files from Google Colab
  • Урок 308. 00:13:37
    Uploading Our Food Vision Mini App to Hugging Face Spaces Programmatically
  • Урок 309. 00:07:45
    Running Food Vision Mini on Hugging Face Spaces and Trying it Out
  • Урок 310. 00:04:18
    Food Vision Big Project Outline
  • Урок 311. 00:09:39
    Preparing an EffNetB2 Feature Extractor Model for Food Vision Big
  • Урок 312. 00:07:46
    Downloading the Food 101 Dataset
  • Урок 313. 00:13:37
    Creating a Function to Split Our Food 101 Dataset into Smaller Portions
  • Урок 314. 00:07:24
    Turning Our Food 101 Datasets into DataLoaders
  • Урок 315. 00:20:16
    Training Food Vision Big: Our Biggest Model Yet!
  • Урок 316. 00:05:49
    Outlining the File Structure for Our Food Vision Big
  • Урок 317. 00:03:34
    Downloading an Example Image and Moving Our Food Vision Big Model File
  • Урок 318. 00:06:57
    Saving Food 101 Class Names to a Text File and Reading them Back In
  • Урок 319. 00:02:21
    Turning Our EffNetB2 Feature Extractor Creation Function into a Python Script
  • Урок 320. 00:10:42
    Creating an App Script for Our Food Vision Big Model Gradio Demo
  • Урок 321. 00:03:46
    Zipping and Downloading Our Food Vision Big App Files
  • Урок 322. 00:13:35
    Deploying Food Vision Big to Hugging Face Spaces
  • Урок 323. 00:06:14
    PyTorch Mode Deployment: Main Takeaways, Extra-Curriculum and Exercises
  • Урок 324. 00:06:02
    Introduction to PyTorch 2.0
  • Урок 325. 00:01:22
    What We Are Going to Cover and PyTorch 2 Reference Materials
  • Урок 326. 00:04:20
    Getting Started with PyTorch 2.0 in Google Colab
  • Урок 327. 00:03:21
    PyTorch 2.0 - 30 Second Intro
  • Урок 328. 00:02:23
    Getting Setup for PyTorch 2.0
  • Урок 329. 00:06:50
    Getting Info from Our GPUs and Seeing if They're Capable of Using PyTorch 2.0
  • Урок 330. 00:09:41
    Setting the Default Device in PyTorch 2.0
  • Урок 331. 00:06:43
    Discussing the Experiments We Are Going to Run for PyTorch 2.0
  • Урок 332. 00:10:18
    Creating a Function to Setup Our Model and Transforms
  • Урок 333. 00:08:24
    Discussing How to Get Better Relative Speedups for Training Models
  • Урок 334. 00:07:16
    Setting the Batch Size and Data Size Programmatically
  • Урок 335. 00:09:54
    Getting More Speedups with TensorFloat-32
  • Урок 336. 00:07:01
    Downloading the CIFAR10 Dataset
  • Урок 337. 00:07:39
    Creating Training and Test DataLoaders
  • Урок 338. 00:04:59
    Preparing Training and Testing Loops with Timing Steps
  • Урок 339. 00:08:23
    Experiment 1 - Single Run without Torch Compile
  • Урок 340. 00:10:39
    Experiment 2 - Single Run with Torch Compile
  • Урок 341. 00:11:20
    Comparing the Results of Experiments 1 and 2
  • Урок 342. 00:04:40
    Saving the Results of Experiments 1 and 2
  • Урок 343. 00:12:42
    Preparing Functions for Experiments 3 and 4
  • Урок 344. 00:12:45
    Experiment 3 - Training a Non-Compiled Model for Multiple Runs
  • Урок 345. 00:09:58
    Experiment 4 - Training a Compiled Model for Multiple Runs
  • Урок 346. 00:05:24
    Comparing the Results of Experiments 3 and 4
  • Урок 347. 00:05:51
    Potential Extensions and Resources to Learn More
  • Урок 348. 00:01:18
    Thank You!