![ritvikmath](/img/default-banner.jpg)
- Видео 489
- Просмотров 11 974 828
ritvikmath
Добавлен 14 июн 2012
Data science for all.
Interrupted Time Series : Data Science Concepts
How do we run an experiment without running an experiment?
Просмотров: 2 597
Видео
Gaussian Processes : Data Science Concepts
Просмотров 7 тыс.День назад
All about Gaussian Processes and how we can use them for regression. RBF Kernel : ruclips.net/video/Q0ExqOphnW0/видео.html 0:00 The Motivation 4:21 The Math 20:39 Importance of the Kernel 21:21 Extensions 22:18 Bayesian Stats
The Most Important Integral in Data Science
Просмотров 5 тыс.Месяц назад
Calculus and Data Science. Best Friends. ROC Curve Video : ruclips.net/video/SHM_GgNI4fY/видео.html 0:00 Intro 3:46 AUC and Tradeoffs 7:56 Integral Form of AUC 11:36 Comparing AUCs
Why You Shouldn't Trust Your ML Models (...too much)
Просмотров 4,8 тыс.2 месяца назад
Whether you call it feedback loops, selection bias, etc, this pesky problem rears its head in almost every problem out there. 0:00 The Problem 9:35 The Solution
I Traded $1000 with Every Tree-Based Machine Learning Model
Просмотров 3,8 тыс.2 месяца назад
Trading $1000 using every tree-based machine learning model! Decision Trees : ruclips.net/video/kakLu2is3ds/видео.html Random Forest : ruclips.net/video/w-eWTxbRQcU/видео.html Gradient Boosted Decision Trees : ruclips.net/video/en2bmeB4QUo/видео.html 0:00 Why Trees? 2:06 The Method 11:37 The Results
Cohen's Kappa : Data Science Basics
Просмотров 2,6 тыс.2 месяца назад
All about Cohen's Kappa in Data Science!
When Higher Volatility is Better in the Stock Market : Call Options
Просмотров 2,3 тыс.2 месяца назад
What affects the premium of call options and how higher volatility can actually be valuable in the stock market. Visuals Created with Excalidraw : excalidraw.com/ 0:00 Strike Price & Expiration Date 7:12 What about Volatility? 11:26 Next Up
Call Options : The Intuition and Math You Need
Просмотров 3,1 тыс.3 месяца назад
All the intuition and math you need to know about call options! 0:00 Intro to Call Options 3:10 Visual Profit Analysis 9:16 Call Options vs Stocks 14:49 Pricing Call Options
I Day Traded $1000 with the Hidden Markov Model
Просмотров 12 тыс.3 месяца назад
Method and results of day trading $1K using the Hidden Markov Model in Data Science 0:00 Method 6:57 Results
The Best Data Visualization of All Time
Просмотров 6 тыс.3 месяца назад
The power of Sankey Diagrams! Visuals Created with Excalidraw excalidraw.com/ Code : github.com/ritvikfood/RUclipsVideoCode/blob/main/Sankey Diagram.ipynb
I Day Traded $1000 : Autoregressive (AR) vs. Recurrent Neural Network (RNN)
Просмотров 29 тыс.3 месяца назад
Comparing the Autoregressive (AR) model vs. the Recurrent Neural Network (RNN) model for stock return prediction! 0:00 Intro 1:30 AR Model 5:17 RNN Model 10:46 Results Visuals Created with Excalidraw excalidraw.com/
I Used Data Science to Buy the Dip
Просмотров 7 тыс.4 месяца назад
Training a machine learning model to predict the bottom of the market and buy the dip!
How the Heck do Bending Genetics work in Avatar the Last Airbender?
Просмотров 1,5 тыс.4 месяца назад
Your dad's an airbender, your mom's a waterbender. What kind of bender are you? Inspiration thread : www.reddit.com/r/FanTheories/comments/7yt1zm/genetics_of_bending_avatar_the_last_airbender/
The Dirichlet Distribution : Data Science Basics
Просмотров 4,6 тыс.4 месяца назад
Beta Distribution Video : ruclips.net/video/1k8lF3BriXM/видео.html 0:00 Recap of Beta Distribution 2:43 Intro to Dirichlet Distribution 6:01 PDF of Dirichlet Distribution 15:07 Statistics and Convergence
Super Bowl Prediction by a Data Scientist
Просмотров 3,1 тыс.4 месяца назад
Link to Data : www.kaggle.com/datasets/tobycrabtree/nfl-scores-and-betting-data?resource=download&select=spreadspoke_scores.csv Logistic Regression Video : ruclips.net/video/9zw76PT3tzs/видео.html RNN Video : ruclips.net/video/DFZ1UA7-fxY/видео.html
Kernel Density Estimation : Data Science Concepts
Просмотров 16 тыс.5 месяцев назад
Kernel Density Estimation : Data Science Concepts
A Data Scientist's Prediction for the 2024 Election
Просмотров 9 тыс.5 месяцев назад
A Data Scientist's Prediction for the 2024 Election
The S&P 500 Isn't As Diversified as You Think. Here's Why.
Просмотров 2,3 тыс.5 месяцев назад
The S&P 500 Isn't As Diversified as You Think. Here's Why.
The Planet Fitness Problem : Improved Markov Chains
Просмотров 2,7 тыс.6 месяцев назад
The Planet Fitness Problem : Improved Markov Chains
The Easy Trick to Understand any Data Science Formula
Просмотров 5 тыс.6 месяцев назад
The Easy Trick to Understand any Data Science Formula
3 Psychological Tips to Hack the Data Science Interview
Просмотров 3,4 тыс.7 месяцев назад
3 Psychological Tips to Hack the Data Science Interview
The Unexpected Pure Math You Have to Know as a Data Scientist : Pythagorean Means
Просмотров 5 тыс.7 месяцев назад
The Unexpected Pure Math You Have to Know as a Data Scientist : Pythagorean Means
The Title and Thumbnail Change if You Watch this Video | Reinforcement Learning
Просмотров 2,4 тыс.7 месяцев назад
The Title and Thumbnail Change if You Watch this Video | Reinforcement Learning
This is the Math You Need to Master Reinforcement Learning
Просмотров 9 тыс.8 месяцев назад
This is the Math You Need to Master Reinforcement Learning
I Day Traded $1000 Using Reinforcement Learning and Bayesian Statistics
Просмотров 8 тыс.8 месяцев назад
I Day Traded $1000 Using Reinforcement Learning and Bayesian Statistics
Can You Solve the Two Radio Problem?
Просмотров 2 тыс.8 месяцев назад
Can You Solve the Two Radio Problem?
Detecting Phrases with Data Science : Natural Language Processing
Просмотров 2 тыс.9 месяцев назад
Detecting Phrases with Data Science : Natural Language Processing
BM25 : The Most Important Text Metric in Data Science
Просмотров 7 тыс.9 месяцев назад
BM25 : The Most Important Text Metric in Data Science
Spearman Correlation - Simply Explained
Просмотров 9 тыс.10 месяцев назад
Spearman Correlation - Simply Explained
I really appreciate your lucid explanations in all your presentations on this subject. In this video you did touch upon the of mean and standard deviation values for the sampled conditional distributions , asking the viewer to see earlier videos for explanation. Perhaps it would improve the presentation, if you at least mentioned that it is due to the bivariate dustribution's correlation parameter (rho = 0.5).
3 years later and still the goat
Really helpful~ Just got stuck at line 2 of the derivation. thought L(1) == C(t-1) ?
watching your 2nd video. Great explanation! The best thing is intuitive understanding. Thank you for help in learning)
I love this guy😭
Did you do training-test split on the data?
I'm a simple man. When Ritvik posts, I watch.
Knocking it out of the park, as usual.
Great video . Can you do some Multivariate analysis data series with different business case scenarios
What an excellent teacher you are! Thank you for all your videos.
Great Explanation
How do you decide whether to accept a sample proposed by g(x)? For example, say f(s) / (M*g(s)) = 0.4, do you accept it or not ?
Amazing
I would add that it's not experimental but quasi experimental because it's not really granting that the effect is due to our treatment. It might be due to a common cause or be causally biased by it so we are also assuming causal sufficiency and a sufficient knowledge of the underlying process. Still we have no way to guarantee that some unobserved cause happened in that time window. Good and clear video, as usual btw, keep going bro °u°
gender pay gap is fake
Great content! Thanks for posting it
Stationarity conditions - mean = c - var = c - no seasonality or periodic repetiotions
ACF - takes into account all possible t-k which can affect t. Calculated using Pearson coefficient. Direct and Indirect effects. PACF - takes into account only t-k, and how it affects t. A regression line is fitted for this. The coefficients obtained explains the affect of t-k on t. Direct effects only.
Dear ritvik why does the determinant value of the first explanation results comes out to be zero(I did try to change some conditions like a single path to every next point but then also that value also comes out to be zero ) what does it indicates?
Notes for future - TS is an extrapolation problem and error keeps on increasing as we move away from known data - Reg is an interpolation problem and error is more or less same, since prediction is usually made in the range of available data.
saved my life again!
I am still confused about how you developed the kernels in the first place. I know what they do but don't know how to obtain them without using the transformed space.
hi R, one ques. in what sequence should we watch these lectures ? the playlist seems to be jumbled , or it serves a purpose ?
Hi Ritvik! The way you've set up the linear model(s), the intercept parameters B0 and B2 will represent the intercept from Tt=0, rather than at the particular time that the interruption happens. So, if you got a large positive change in the B3 slope parameter, you'll probably get a negative change in the B2 parameter (since a steeper line at positive Tt will intercept the yt axis *below* where the original line did! Wouldn't it make more sense to do a shift/translation of the Tt parameter so that it works as if the interruption happens at the yt axis? For example, use a translated variable Ut = (Tt - t). So, if the interruption happens at t = 75, then when Tt = 75, Ut = (Tt - t) = (75 - 75) = 0. And your model can then be: yt = B0 + B1⋅Ut + Dt⋅(B2 + B3⋅Ut) At least this way, the B0 and B2 params will have a much more interpretable meaning. B0 will be the value of yt at Tt = t of the main linear model, and B2 will be the *initial change* in vt at Tt = t. In other words, how much 'immediate effect' did the interruption have; how much of a vertical 'jump' up or down. Granted, to calculate vt at any particular Tt, you'll have to first convert to Ut, but that's not so bad, just a simple shift. And if you really need to find the linear params in terms of Tt, it's fairly easy to just plug in Ut = (Tt - t) and expand out to find the transformed linear params for Tt.
Thank you so much
How can we take advantage of the GPU?
Thanks for this tutorial on ITS. I have to admit, I am a bit confused by this example. I wasn't sure at first why. But basically, it is because the "data" is "too simple". At first, I look at this and I say "Any one" can see that Pre-advertising is not as good as "post" advertising treatment. So, Why would I need to use this technique. Why not just count sales per "month" and say, yup ... we boosted total sales per month. There's nothing inherent in this data sample that confounds the analysis or makes it difficult or non-obvious. On the other hand, if you could show something where the sales data points are more difficult to track, so that it is non-obvious by immediately looking at the data (in graph form) that the trend after treatment is looking seemingly more positive, but using ITS we can see that while visually it seems to look better, in fact, it is not as good after treatment (per this method.) And perhaps a second example, where it seemingly looks more negative after treatment, but by using ITS we can prove that it is better after treatment Then, I think those would give more credence to this method. Because as far as I can tell, the graph makes it far to easy too see the trend... so why bother with ITS unless it can show something that isn't obvious. And, again, thanks for this explanation on Interrupted Time Series.
i still dont get the part of getting the effect without the experiment.... if i dont have the time series data of my ice cream shop where i decided to advertise as right now im in middle of taking this decision then how can i fit a model which can measure the effects of it?
I confess, I didn't understand any single word of explanations from my MIT professor about perceptron. How ever, after I saw this video and understood clearly what the idea is. Thanks.
thanks, may I ask what do you do for a living?
the equation of the hyperplane is w·x+b=0, isn't it? The video says - b=0 instead of +b=0
Great vid. Please explain how to compute the confidence intervals.
Absolute legend
Man your channel would blow up spectacularly if you invested the time into learning how to make really nice visuals.. the whole poorly hand drawn example thing is really 2005 && screams laziness and/or amateur..
Excellent video and content, thank you. I am not a trader nor a financial person but the training window size vs. the holding window size may be part of the problem unless the choosen entry time would be an outlier by chance. The holding size may play a huge role in certain market dinamics with periods of revovery much greater than the chose holding window time. When market goes down correlations tend to one and in many cases do revert to positive from negative in calm situation. Best
I thought U is mxm, V is nxn and SIGMA is mxn
Can you show us the code for this, especially for VAR
there's some pretty basic code here scikit-learn.org/stable/auto_examples/gaussian_process/plot_gpr_noisy_targets.html#sphx-glr-auto-examples-gaussian-process-plot-gpr-noisy-targets-py
This means that apples are far better than bananas right? So when I've been told that you cant compare apples and bananas ... thats a lie ...
Thank you for the video. it was nicely explained. There are a lot of simplifications. Could you also talk about how best select sigma and l - is it all done empirically? also do you have any example of implementation?
Pls recommend a book for data science n machine learning
Very clear and helpful. Thank you so much!
You're very welcome!
So simple… but so powerful!
glad you think so!
Go on, tell us how to test if and when a change occurred :)
Any data science text book recommendations?
Pattern Recognition and Machine Learning from Bishop
Thank you for this
I would argue that 2 has seasonality...
Hmmm i noticed that if two categories are strongly correlated, the plot will look close to a straight line. Going to multidimensional space, that "line" looks like the vector u1 in the video, on which the data are projected. Does that mean PCA will perform better the more correlated two (or more) categories are?
Can you do this again? Polls mean nothing. There are too many confounding variables.
amazingly taught. thank you so much
nice