Recent Updates Toggle Comment Threads | Keyboard Shortcuts

  • wangxinxi 02:36 on May 13, 2014 Permalink | Reply
    Tags: theory   

    Nothing is more practical than a good theory

  • wangxinxi 02:20 on May 9, 2014 Permalink | Reply  

    WTF, how could you let Hank die? W. W. you are completely a disgusting, greedy, and cowardly loser.

  • wangxinxi 00:38 on March 21, 2014 Permalink | Reply

    An interesting paper about determinism, realism, and locality.

    Click to access term_paper_yaron_hadad1.pdf

  • wangxinxi 16:48 on March 17, 2014 Permalink | Reply
    Tags: Python   

    pickle.dump(model, open(sys.argv[1], ‘wb’))
    SystemError: error return without exception set

    The error can be reproduced by the following code:
    python2.7 -c “import cPickle; cPickle.dumps(‘\x00’ * 2**31)”

    Changing cPickle to pickle may solve this problem if you do not care about the efficiency issue of pickle. To completely solve the issue, you need to switch to Python 3. Using maybe good choices too.

  • wangxinxi 13:55 on March 14, 2014 Permalink | Reply
    Tags: , statistics   

    An efficient estimator is the one that achieves the equality of the Cramer-Rao bound (CRB)
    The Cramer-Rao bound says that the variance of an unbiased estimator cannot be smaller than the inverse of the Fisher information. However, a bias estimator can have smaller variance than the Cramer-Rao bound, which is also known as the bias-variance trade-off.

  • wangxinxi 11:34 on March 14, 2014 Permalink | Reply

    A major attraction of the connectionist approach to language, apart from its natural relation to neural computation, is that the very same processing mechanisms apply across the full range of linguistic structure.

    Douglas L. T. Rohde, David C. Plaut, Connectionist Models of Language Processing

  • wangxinxi 20:52 on March 6, 2014 Permalink | Reply

  • wangxinxi 18:29 on February 23, 2014 Permalink | Reply

    A good thread about some philosophical questions on PAC-Bayes, Occam’s Razor, and Bayesian priors.

    An introduction to PAC-Bayesian learning theory.

    Click to access ICML_PAC-Bayes_Tutorial.pdf

    PAC-Bayes bounds are a generalization of the Occam’s razor bound for algorithms which output a distribution over classifiers rather than just a single classifier. This includes the possibility of a distribution over a single classifier, so it is a generalization. Most classifiers do not output a distribution over base classifiers. Instead, they output either a classifier, or an average over base classifiers. Nonetheless, PAC-Bayes bounds are interesting for several reasons:

    PAC-Bayes bounds are much tighter (in practice) than most common VC-related approaches on continuous classifier spaces. This can be shown by application to stochastic neural networks (see section 13) as well as other classifiers. It also can be seen by observation: when specializing the PAC-Bayes bounds on discrete hypothesis spaces, only O(lnm) sample complexity is lost.
    Due to the achievable tightness, the result motivates new learning algorithms which strongly limit the amount of overfitting that a learning algorithm will incur.
    The result found here will turn out to be useful for averaging hypotheses.

  • wangxinxi 03:54 on February 23, 2014 Permalink | Reply

    A lot of philosophical discussions about latent variable models in this text.

  • wangxinxi 01:05 on February 22, 2014 Permalink | Reply

    rational-weighted recurrent NNs having boolean activation functions (simple thresholds) are equivalent to finite state automata (Minsky, “Computation: finite and infinite machines”, 1967);

    rational-weighted recurrent NNs having linear sigmoid activation functions are equivalent to Turing Machines (Siegelmann and Sontag, “On the computational power of neural nets”, 1995);

    real-weighted recurrent NNs having linear sigmoid activation functions are more powerful than Turing Machines (Siegelmann and Sontag, “Analog computation via neural networks”, 1993);

    but …

    real-weighted recurrent NNs with Gaussian noise on the outputs cannot recognize arbitrary regular languages (Maass and Sontag, “Analog Neural Nets with Gaussian or Other Common NoiseDistributions Cannot Recognize Arbitrary Regular Languages” , 1995);

    Some interesting slides:

    Click to access

    Click to access hypercomputationUCL.pdf

    To conclude: there is probably no real Turing machine (TM) in reality. There is also unlikely any “accurate” analog neural nets. Proving that the analog neural nets have better computation capability than TM is completely OK, but they all are just not implementable.

    • Min 16:14 on June 20, 2014 Permalink | Reply

      interesting, what’s difference between rational weighted and real weighted.

      • wangxinxi 16:24 on June 20, 2014 Permalink | Reply

        Rational numbers can be represented with finite memory, but real numbers in general, eg. Pi, cannot.

Compose new post
Next post/Next comment
Previous post/Previous comment
Show/Hide comments
Go to top
Go to login
Show/Hide help
shift + esc