Most Popular
1500 questions
6
votes
2 answers
What are the advantages of the Kullback-Leibler over the MSE/RMSE?
I've recently encountered different articles that are recommending to use the KL divergence instead of the MSE/RMSE (as the loss function), when trying to learn a probability distribution, but none of the articles are giving a clear reasoning why…
razvanc92
- 1,128
- 1
- 9
- 18
6
votes
1 answer
How do I know when to use which Monte Carlo method?
I'm a bit confused with extensive number of different Monte Carlo methods such as:
Hamiltonian/Hybrid Monte Carlo (HMC),
Dynamic Monte Carlo (DMC),
Markov chain Monte Carlo (MCMC),
Kinetic Monte Carlo (KMC),
Dynamic Monte Carlo (DMC)
Quasi-Monte…
kenorb
- 10,483
- 3
- 44
- 94
6
votes
3 answers
Why did a Tesla car mistake a truck with a bright sky?
Do we know why Tesla's autopilot mistaken empty sky with a high-sided lorry which resulted in fatal crash involving a car in self-drive mode? Was it AI fault or something else? Is there any technical explanation behind this why this happened?
The…
kenorb
- 10,483
- 3
- 44
- 94
6
votes
1 answer
How is the gradient of the loss function in DQN derived?
In the original DQN paper, page 1, the loss function of the DQN is
$$
L_{i}(\theta_{i}) = \mathbb{E}_{(s,a,r,s') \sim U(D)} [(r+\gamma \max_{a'} Q(s',a',\theta_{i}^{-}) - Q(s,a;\theta_{i}))^2]
$$
whose gradient is presented (on page…
Dimitris Monroe
- 171
- 8
6
votes
1 answer
Benchmarks for reinforcement learning in discrete MDPs
To compare the performance of various algorithms for perfect information games, reasonable benchmarks include reversi and m,n,k-games (generalized tic-tac-toe). For imperfect information games, something like simplified poker is a reasonable…
user76284
- 347
- 1
- 14
6
votes
2 answers
What is the current state-of-the-art in Reinforcement Learning regarding data efficiency?
In other words, which existing reinforcement method learns with fewest episodes? R-Max comes to mind, but it's very old and I'd like to know if there is something better now.
rcpinto
- 2,119
- 1
- 16
- 31
6
votes
3 answers
What are the state-of-the-art approaches for continual learning with neural networks?
There seems to be a lot of literature and research on the problems of stochastic gradient descent and catastrophic forgetting, but I can't find much on solutions to perform continual learning with neural network architectures.
By continual learning,…
gcorso
- 366
- 1
- 7
6
votes
1 answer
Is there any use of using 3D convolutions for traditional images (like cifar10, imagenet)?
I am curious if there is any advantage of using 3D convolutions on images like CIFAR-10/100 or ImageNet. I know that they are not usually used on this data set, though they could because the channel could be used as the "depth" channel.
I know that…
Charlie Parker
- 211
- 2
- 6
6
votes
1 answer
How to show temporal difference methods converge to MLE?
In chapter 6 of Sutton and Barto (p. 128), they claim temporal difference converges to the maximum likelihood estimate (MLE). How can this be shown formally?
fool
- 203
- 1
- 7
6
votes
1 answer
Video summarization similar to Summe's TextRank
We have the popular TextRank API which given a text, ranks keywords and can apply summarization given a predefined text length.
I am wondering if there is a similar tool for video summarization. Maybe a library, a deep model or ML-based tool that…
Mary
- 973
- 6
- 13
6
votes
1 answer
Are simple animal snares and traps a form of automation? Of computation?
I'm trying to understand the relationship of humans and automation, historically and culturally.
I ask because the waterclock is generally considered the earliest form of automation, but snares and deadfall traps constitute simple switch…
DukeZhou
- 6,227
- 5
- 25
- 53
6
votes
3 answers
Are fully connected layers necessary in a CNN?
I have implemented a CNN for image classification. I have not used fully connected layers, but only a softmax. Still, I am getting results.
Must I use fully-connected layers in a CNN?
SARIKA
- 61
- 1
- 2
6
votes
1 answer
What is the difference between asymmetric and depthwise separable convolution?
I have recently discovered asymmetric convolution layers in deep learning architectures, a concept which seems very similar to depthwise separable convolutions.
Are they really the same concept with different names? If not, where is the difference?…
Pierre Gramme
- 163
- 4
6
votes
3 answers
Is it ok to struggle with mathematics while learning AI as a beginner?
I have a decent background in Mathematics and Computer Science .I started learning AI from Andrew Ng's course from one month back. I understand logic and intuition behind everything taught but if someone asks me to write or derive mathematical…
user27556
6
votes
2 answers
What is the difference between learning without forgetting and transfer learning?
I would like to incrementally train my model with my current dataset and I asked this question on Github, which is what I'm using SSD MobileNet v1.
Someone there told me about learning without forgetting. I'm now confused between learning without…
abhimanyuaryan
- 191
- 1
- 4