Activation Functions in Deep Learning Sigmoid Tanh and Relu Activation Function











############################# Video Source: www.youtube.com/watch?v=7LcUkgzx3AY

In artificial neural networks, each neuron forms a weighted sum of its inputs and passes the resulting scalar value through a function referred to as an activation function or transfer function. In this video, we explain the basics of Sigmoid, Tanh, and Relu—important parts of how computers learn. • Digital Notes for Deep Learning: https://shorturl.at/NGtXg • 👍If you find this video helpful, consider giving it a thumbs up and subscribing for more educational videos on data science! • 💭Share your thoughts, experiences, or questions in the comments below. I love hearing from you! • ============================ • Do you want to learn from me? • Check my affordable mentorship program at : https://learnwith.campusx.in • ============================ • 📱 Grow with us: • CampusX' LinkedIn:   / campusx-official   • CampusX on Instagram for daily tips:   / campusx.official   • My LinkedIn:   / nitish-singh-03412789   • Discord:   / discord   • ✨ Hashtags✨ • #SimpleLearning #ActivationFunctionsExplained #EasyTech • ⌚Time Stamps⌚ • 00:00 - Intro • 00:47 - What are activation functions? • 03:28 - Importance of AF • 04:58 - Code Demo • 06:38 - Why activation functions are needed? • 11:05 - Ideal Activation function • 18:41 - Sigmoid Activation Function • 20:37 - Advantages • 22:56 - Disadvantages • 36:15 - Tan h Activation Function • 38:00 - Advantages • 39:02 - Disadvantages • 40:17 - Relu Activation Function • 40:50 - Advantages • 42:43 - Disadvantages • 44:24 - Outro

#############################









Content Report
Youtor.org / Youtor.org Torrents YT video Downloader © 2024

created by www.mixer.tube