Using Machine Learning to Understand and Enhance Human Learning Capacity

Research Projects

The overall goal of the project is to develop computational learning models and theory, originally aimed at computers, to predict and influence human learning behaviors.

Capacity measure of the human mind

What is the VC-dimension of the human mind? In machine learning, the VC-dimension is a well-known capacity measure for a model family. What if the "model family" is the human mind, e.g., all the classifiers that one can come up with? Can we estimate such a capacity for humans? We propose a method to estimate the Rademacher complexity of the human mind in binary categorization tasks. It will tell us the intrinsic complexity of the human thinking process. It also has direct application in understanding overfitting in human learning.

Read more:

Optimal teaching

Given a task and a learner, can a teacher design an optimal teaching strategy so that the learner "gets" the true concept quickly? Recent work in the machine learning community on teaching dimension and curriculum learning starts to address this question. We are developing new computational theory and performing human behavioral experiments to advance our understanding of optimal teaching.

Read more:

Human semi-supervised learning

Human category learning is traditionally thought of as supervised learning. We demonstrated that it is in fact greatly influenced by unlabeled data, and should be modeled as semi-supervised learning. For example, after learning, just performing categorization on unlabeled test items can change the human's mind about the decision boundary.

Read more:

Human active learning

Under certain conditions, an active machine learner provably outperforms a passive learner. If we allow a human learner to submit queries and obtain oracle labels, can they do better than their peers who passively receive iid training samples? We showed that the answer is yes.

Read more:

Publications from this project


  1. Ji Liu and Xiaojin Zhu. The teaching dimension of linear learners. Journal of Machine Learning Research, 17(162):1-25, 2016.
    This is the journal version of the ICML'16 paper, with a discussion on teacher-learner collusion.
    [link]

  2. Xiaojin Zhu, Ji Liu, and Manuel Lopes. No learner left behind: On the complexity of teaching multiple learners simultaneously. In The 26th International Joint Conference on Artificial Intelligence (IJCAI), 2017.
    Minimax teaching dimension to make the worst learner in a class learn. Partitioning the class into sections improves teaching dimension.
    [pdf]

  3. Scott Alfeld, Xiaojin Zhu, and Paul Barford. Explicit defense actions against test-set attacks. In The Thirty-First AAAI Conference on Artificial Intelligence (AAAI), 2017.
    [pdf]

  4. Tzu-Kuo Huang, Lihong Li, Ara Vartanian, Saleema Amershi, and Xiaojin Zhu. Active learning with oracle epiphany. In Advances in Neural Information Processing Systems (NIPS), 2016.
    This paper brings active learning theory and practice closer. We analyze active learning query complexity where the oracle initially may not know how to answer queries from a certain region in the input space. After seeing multiple queries from the region, the oracle can have an "epiphany", i.e. realizing how to answer any queries from that region.
    [pdf]

  5. Ji Liu, Xiaojin Zhu, and H. Gorune Ohannessian. The Teaching Dimension of Linear Learners. In The 33rd International Conference on Machine Learning (ICML), 2016.
    We provide lower bounds on training set size to perfectly teach a linear learning. We also provide the corresponding upper bounds (and thus teaching dimension) by exhibiting teaching sets for SVM, logistic regression, and ridge regression.
    [pdf | supplementary | arXiv preprint]

  6. Xiaojin Zhu, Ara Vartanian, Manish Bansal, Duy Nguyen, and Luke Brandl. Stochastic multiresolution persistent homology kernel. In The 25th International Joint Conference on Artificial Intelligence (IJCAI), 2016.
    A kernel built upon persistent homology at multiple resolutions, and with Monte Carlo to speed up. Ready to use as topological features for machine learning.
    [pdf]

  7. Kwang-Sung Jun, Kevin Jamieson, Rob Nowak, and Xiaojin Zhu. Top arm identification in multi-armed bandits with batch arm pulls. In The 19th International Conference on Artificial Intelligence and Statistics (AISTATS), 2016.
    [pdf]

  8. Scott Alfeld, Xiaojin Zhu, and Paul Barford. Data Poisoning Attacks against Autoregressive Models. In The Thirtieth AAAI Conference on Artificial Intelligence (AAAI), 2016.
    Machine teaching for autoregression, applied to computer security.
    [pdf]

  9. Kwang-Sung Jun, Xiaojin Zhu, Timothy Rogers, Zhuoran Yang, and Ming Yuan. Human memory search as initial-visit emitting random walk. In Advances in Neural Information Processing Systems (NIPS), 2015.
    A random walk that only emits an output when it visits a state for the first time.
    [pdf | supplemental | poster]

  10. Gautam Dasarathy, Robert Nowak, and Xiaojin Zhu. S2: An efficient graph based active learning algorithm with application to nonparametric classification. In Conference on Learning Theory (COLT), 2015.
    [pdf ]

  11. Shike Mei and Xiaojin Zhu. The security of latent Dirichlet allocation. In The Eighteenth International Conference on Artificial Intelligence and Statistics (AISTATS), 2015.
    [pdf | slides ]

  12. Xiaojin Zhu. Machine Teaching: an Inverse Problem to Machine Learning and an Approach Toward Optimal Education. In The Twenty-Ninth AAAI Conference on Artificial Intelligence (Senior Member Track, AAAI), 2015.
    [pdf | talk slides | project link]

  13. Shike Mei and Xiaojin Zhu. Using Machine Teaching to Identify Optimal Training-Set Attacks on Machine Learners. In The Twenty-Ninth AAAI Conference on Artificial Intelligence (AAAI), 2015.
    [pdf | poster ad | poster | Mendota ice data | Tech Report 1813]

  14. Bryan Gibson, Timothy Rogers, Charles Kalish, and Xiaojin Zhu. What causes category-shifting in human semi-supervised learning? In The 32nd Annual Conference of the Cognitive Science Society (CogSci), 2015.
    [pdf]

  15. Kaustubh Patil, Xiaojin Zhu, Lukasz Kopec, and Bradley Love. Optimal Teaching for Limited-Capacity Human Learners. In Advances in Neural Information Processing Systems (NIPS), 2014.
    [pdf | poster | spotlight | data]

  16. Charles Kalish, Xiaojin Zhu, and Timothy Rogers. Drift in children's categories: When experienced distributions conflict with prior learning. In Developmental Science, 2014.

  17. Xiaojin Zhu. Machine teaching for Bayesian learners in the exponential family. In Advances in Neural Information Processing Systems (NIPS), 2013.
    [pdf | poster]

  18. Kwang-Sung Jun, Xiaojin Zhu, Burr Settles, and Timothy Rogers. Learning from Human-Generated Lists. In The 30th International Conference on Machine Learning (ICML), 2013.
    [pdf | slides | SWIRL v1.0 code | video]

  19. Bryan R. Gibson, Timothy T. Rogers, and Xiaojin Zhu. Human semi-supervised learning. Topics in Cognitive Science, 5(1):132-172, 2013.
    [link]

  20. Xiaojin Zhu. Persistent homology: An introduction and a new text representation for natural language processing. In The 23rd International Joint Conference on Artificial Intelligence (IJCAI), 2013.
    [pdf | slides | data and code ]

  21. Burr Settles and Xiaojin Zhu. Behavioral factors in interactive training of text classifiers. In North American Chapter of the Association for Computational Linguistics - Human Language Technologies (NAACL HLT). Short paper. 2012.
    [pdf]

  22. Faisal Khan, Xiaojin Zhu, and Bilge Mutlu. How do humans teach: On curriculum learning and teaching dimension. In Advances in Neural Information Processing Systems (NIPS) 25. 2011.
    [pdf | data | slides]

  23. Shilin Ding, Grace Wahba, and Xiaojin Zhu. Learning higher-order graph structure with features by structure penalty. In Advances in Neural Information Processing Systems (NIPS) 25. 2011. [pdf]

  24. Jun-Ming Xu, Xiaojin Zhu, and Timothy T. Rogers. Metric learning for estimating psychological similarities. ACM Transactions on Intelligent Systems and Technology (ACM TIST), 2011. [journal link | unofficial version | data | code]

  25. David Andrzejewski, Xiaojin Zhu, Mark Craven, and Ben Recht. A framework for incorporating general domain knowledge into Latent Dirichlet Allocation using First-Order Logic. The Twenty-Second International Joint Conference on Artificial Intelligence (IJCAI-11), 2011. [pdf | slides | poster | code]

  26. Xiaojin Zhu, Bryan Gibson, and Timothy Rogers. Co-training as a human collaboration policy. In The Twenty-Fifth Conference on Artificial Intelligence (AAAI-11), 2011. [pdf]

  27. Andrew Goldberg, Xiaojin Zhu, Alex Furger, and Jun-Ming Xu. OASIS: Online active semisupervised learning. In The Twenty-Fifth Conference on Artificial Intelligence (AAAI-11), 2011. [pdf]

  28. Chen Yu, Jun-Ming Xu, and Xiaojin Zhu. Word learning through sensorimotor child-parent interaction: A feature selection approach. The 33rd Annual Conference of the Cognitive Science Society (CogSci 2011), 2011.
    [pdf]

  29. Charles W. Kalish, Timothy T. Rogers, Jonathan Lang, and Xiaojin Zhu. Can semi-supervised learning explain incorrect beliefs about categories? Cognition, 2011. [link]

  30. Bryan Gibson, Xiaojin Zhu, Tim Rogers, Chuck Kalish, and Joseph Harrison. Humans learn using manifolds, reluctantly. In Advances in Neural Information Processing Systems (NIPS) 24, 2010. [pdf | NIPS talk slides]

  31. Andrew Goldberg, Xiaojin Zhu, Benjamin Recht, Jun-Ming Xu, and Robert Nowak. Transduction with matrix completion: Three birds with one stone. In Advances in Neural Information Processing Systems (NIPS) 24. 2010. [pdf]

  32. Xiaojin Zhu, Bryan R. Gibson, Kwang-Sung Jun, Timothy T. Rogers, Joseph Harrison, and Chuck Kalish. Cognitive models of test-item effects in human category learning. In The 27th International Conference on Machine Learning (ICML), 2010. [paper pdf]

  33. Bryan R Gibson, Kwang-Sung Jun, and Xiaojin Zhu. With a little help from the computer: Hybrid human-machine systems on bandit problems. In NIPS 2010 Workshop on Computational Social Science and the Wisdom of Crowds, 2010.
    [pdf]

Selected highlights from publications

In terms of understanding learning, we have made a number of discoveries: In terms of enhancing learning, we have made the follow progress:

Data sets for download

Code for download

Research Group

Faculty

Graduate Students

Undergraduate Students

Staff

Collaborators


Related NIPS 2008 Workshop on Machine Learning Meets Human Learning


Professor Xiaojin Zhu in Computer Sciences at the University of Wisconsin-Madison is the recipient of a 2010 Faculty Early Career Development Award (CAREER) from the National Science Foundation, a five-year grant designed to boost young faculty in establishing integrated research and educational activities while helping to address areas of important need. Zhu's CAREER project is titled "Using Machine Learning to Understand and Enhance Human Learning Capacity." His project aims to discover the common mathematical principles that govern learning in both humans and computers. Examples include rigorous generalization error bounds (how well can a student or a robot generalize what the teacher taught to new problems?), sparsity (how well can the student or robot identify a few salient features of a problem, out of a haystack of irrelevant features?), and active learning (can the student or robot ask good questions to speed up its own learning?). He expects the project will lead to novel computational approaches to enhance human learning in and out of classrooms, and advance machine learning by incorporating insights on tasks where humans excel. This project is based upon work supported by the National Science Foundation under Grant No. IIS-0953219. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation.