ArtificialIntelligenceArticles
2.97K subscribers
1.64K photos
9 videos
5 files
3.86K links
for who have a passion for -
1. #ArtificialIntelligence
2. Machine Learning
3. Deep Learning
4. #DataScience
5. #Neuroscience

6. #ResearchPapers

7. Related Courses and Ebooks
Download Telegram
Sequential Convolutional Recurrent Neural Networks for Fast Automatic Modulation Classifi... https://arxiv.org/abs/1909.03050
Intel AI Developer Program (Free)

https://software.intel.com/en-us/ai/courses

Find The Most Updated and Free Artificial Intelligence, Machine Learning, Data Science, Deep Learning, Mathematics, Python, R Programming Resources.
https://www.marktechpost.com/free-resources/
Deep weakly-supervised learning methods for classification and localization in histology images: a survey
Rony et al.: https://arxiv.org/abs/1909.03354
#ArtificialIntelligence #DeepLearning #MachineLearning
SpeechBrain
A PyTorch-based Speech Toolkit : https://speechbrain.github.io
Project Leader: Mirco Ravanelli
#Speech #PyTorch #SpeechBrain
Modern Perspectives on Reinforcement Learning in Finance
Petter Kolm and Gordon Ritter : https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3449401
#ReinforcementLearning #Finance #Hedging
In fact, your deep learning paper is the all-time most important Nature paper when ranked with eigencentrality, surpassing the second place “small world” paper by Duncan Watts et al published in late 90’s. Atari is the third. We use eigencentrality because citation count can be gamed easily. https://academic.microsoft.com/search?q=nature&qe=%40%40%40Composite(J.JN%3D%3D%27nature%27)&f=&orderBy=0&skip=0&take=10
Interested in becoming a data scientist?⠀

These are the 10 most important machine learning algorithms that you need to master to break into the field:⠀

• Linear regression ⠀
• Logistic regression⠀
• SVM⠀
• Random forest⠀
• Gradient boosting⠀
• PCA⠀
• K-means clustering⠀
• Collaborative filtering⠀
• kNN⠀
• ARIMA⠀

Bonus: Neural networks⠀


And here are the course notes and book that I first used to learn machine learning:⠀

https://l2r.cs.uiuc.edu/~danr/Teaching/CS446-17/lectures.html

https://web.stanford.edu/~hastie/ElemStatLearn/printings/ESLII_print12.pdf


👆 These notes do an amazing job teaching the algorithms with deeper mathematical rigor while still being easy to follow.⠀


Master the above algorithms and you'll be well on your way to becoming a data scientist.⠀

#aspiring #datascientist #datascience #machinelearning #coding

For more detailed info, make sure to join my mailing list - you'll love the tips I share to help you break into the field -> https://www.datasciencedreamjob.com/free-tips
Batch Normalization is a Cause of Adversarial Vulnerability

Abstract - Batch normalization (batch norm) is often used in an attempt to stabilize and accelerate training in deep neural networks. In many cases it indeed decreases the number of parameter updates required to achieve low training error. However, it also reduces robustness to small adversarial input perturbations and noise by double-digit percentages, as we show on five standard data-sets. Furthermore, substituting weight decay for batch norm is sufficient to nullify the relationship between adversarial vulnerability and the input dimension. Our work is consistent with a mean-field analysis that found that batch norm causes exploding gradients.

Page - https://arxiv.org/abs/1905.02161

PDF - https://arxiv.org/pdf/1905.02161.pdf
Machine Learning for Physics and the Physics of Learning Tutorials"
Videos and slides, by IPAM (an NSF Math Institute at UCLA dedicated to promoting the interaction of math with other disciplines):
https://www.ipam.ucla.edu/programs/workshops/machine-learning-for-physics-and-the-physics-of-learning-tutorials/?tab=schedule
https://t.iss.one/ArtificialIntelligenceArticles

#MLP2019 #MachineLearning #Physics