I am now taking Andrew Ng's Machine Learning course on Coursera. This is my first exposure to Machine Learning. In the discussion of logistic regression, the course contents describe a decision boundary defined by parameters $\theta$, and the classification of input $X$ being defined as $\theta^TX>0$. The model for logistic regression however, uses the sigmoid function to produce probabilistic prediction rather than an a concrete one.
My question is, is there a fundamental limitation that prevents the regression from learning the decision surface directly, and making absolute predictions for test data, if we do not wish to make probabilistic predictions? I realize that we may need different cost functions, and possibly different optimization algorithms.