Data, Knowledge & Life

Weiwei Cheng's blog

Archive for the ‘学术’ Category

My talk at NIPS 2011 workshop

leave a comment »

My NIPS 2011 talk at Choice Model and Preference Learning workshop, “Label Ranking with Abstention: Predicting Partial Orders by Thresholding Probability Distributions” is now available at


Written by Weiwei

24/01/2012 at 14:07

发表在 学术

XRCE is looking for an intern on Bayesian preference learning

leave a comment »

Xerox Research Center Europe is searching for an intern on Bayesian preference learning. It is a great opportunity for anyone interested in doing fantastic research at a fantastic place.

Check the specific here. And have fun in France!

Thank Shengbo for sharing this news with me.

Written by Weiwei

24/11/2011 at 03:48

发表在 学术

My talk at ECMLPKDD 2011

leave a comment »

My ECMLPKDD 2011 talk on “Learning Monotone Nonlinear Models using the Choquet Integral” is now available at Check it out, and let me know what you think.

Written by Weiwei

14/11/2011 at 02:30

发表在 学术

Some interesting papers at ICML 2011

with 2 comments

Papers accepted at ICML 2011 are now online and available for download. There are a number of exciting papers. Krzysztof went there already on Monday. After he comes back, we will get to know more details of the conference. The talks will probably be online soon as well. I will just stay tuned. Here I list several papers (in no particular order) that I have found very interesting. BTW, I think putting the abstracts online is a good idea. ECML should do this as well, so people can better decide which papers / talks to follow.

Multi-Label Classification on Tree- and DAG-Structured Hierarchies by Wei Bi and James Kwok

Abstract: Many real-world applications involve multi-label classification, in which the labels are organized in the form of a tree or directed acyclic graph (DAG). However, current research efforts typically ignore the label dependencies or can only exploit the dependencies in tree-structured hierarchies. In this paper, we present a novel hierarchical multi-label classification algorithm which can be used on both tree- and DAG-structured hierarchies. The key idea is to formulate the search for the optimal consistent multi-label as the finding of the best subgraph in a tree/DAG. Using a simple greedy strategy, the proposed algorithm is computationally efficient, easy to implement, does not suffer from the problem of insufficient/skewed training data in classifier training, and can be readily used on large hierarchies. Theoretical results guarantee the optimality of the obtained solution. Experiments are performed on a large number of functional genomics data sets. The proposed method consistently outperforms the state-of-the-art method on both tree- and DAG-structured hierarchies.

Classification-Based Policy Iteration with a Critic by Victor Gabillon, Alessandro Lazaric, Mohammad Ghavamzadeh, and Bruno Scherrer

Abstract: In this paper, we study the effect of adding a value function approximation component (critic) to rollout classification-based policy iteration (RCPI) algorithms. The idea is to use a critic to approximate the return after we truncate the rollout trajectories. This allows us to control the bias and variance of the rollout estimates of the action-value function. Therefore, the introduction of a critic can improve the accuracy of the rollout estimates, and as a result, enhance the performance of the RCPI algorithm. We present a new RCPI algorithm, called direct policy iteration with critic (DPI-Critic), and provide its finite-sample analysis when the critic is based on the LSTD method. We empirically evaluate the performance of DPI-Critic and compare it with DPI and LSPI in two benchmark reinforcement learning problems.

Learning Scoring Functions with Order-Preserving Losses and Standardized Supervision by David Buffoni, Clément Calauzenes, Patrick Gallinari, and Nicolas Usunier

Abstract: We address the problem of designing surrogate losses for learning scoring functions in the context of label ranking. We extend to ranking problems a notion of order preserving losses previously introduced for multiclass classification, and show that these losses lead to consistent formulations with respect to a family of ranking evaluation metrics. An order-preserving loss can be tailored for a given evaluation metric by appropriately setting some weights depending on this metric and the observed supervision. These weights, called the standard form of the supervision, do not always exist, but we show that previous consistency results for ranking were proved in special cases where they do. We then evaluate a new pairwise loss consistent with the (Normalized) Discounted Cumulative Gain on benchmark datasets.

Learning Mallows Models with Pairwise Preferences by Tyler Lu and Craig Boutilier

Abstract: Learning preference distributions is a key problem in many areas (e.g., recommender systems, IR, social choice). However, existing methods require restrictive data models for evidence about user preferences. We relax these restrictions by considering as data arbitrary pairwise comparisons—the fundamental building blocks of ordinal rankings. We develop the first algorithms for learning Mallows models (and mixtures) with pairwise comparisons. At the heart is a new algorithm, the generalized repeated insertion model (GRIM), for sampling from arbitrary ranking distributions. We develop approximate samplers that are exact for many important special cases—and have provable bounds with pairwise evidence—and derive algorithms for evaluating log-likelihood, learning Mallows mixtures, and non-parametric estimation. Experiments on large, real-world datasets show the effectiveness of our approach.

Online AUC Maximization by Peilin Zhao, Steven Hoi, Rong Jin, and Tianbao Yang

Abstract: Most studies of online learning measure the performance of a learner by classification accuracy, which is inappropriate for applications where the data are unevenly distributed among different classes. We address this limitation by developing online learning algorithm for maximizing Area Under the ROC curve (AUC), a metric that is widely used for measuring the classification performance for imbalanced data distributions. The key challenge of online AUC maximization is that it needs to optimize the pairwise loss between two instances from different classes. This is in contrast to the classical setup of online learning where the overall loss is a sum of losses over individual training examples. We address this challenge by exploiting the reservoir sampling technique, and present two algorithms for online AUC maximization with theoretic performance guarantee. Extensive experimental studies confirm the effectiveness and the efficiency of the proposed algorithms for maximizing AUC.

And of course the one from our group, which is co-authored by Wojciech, Krzysztof and Eyke:

Bipartite Ranking through Minimization of Univariate Loss by Wojciech Kotlowski, Krzysztof Dembczynski and Eyke Hüllermeier

Abstract: Minimization of the rank loss or, equivalently, maximization of the AUC in bipartite ranking calls for minimizing the number of disagreements between pairs of instances. Since the complexity of this problem is inherently quadratic in the number of training examples, it is tempting to ask how much is actually lost by minimizing a simple univariate loss function, as done by standard classification methods, as a surrogate. In this paper, we first note that minimization of 0/1 loss is not an option, as it may yield an arbitrarily high rank loss. We show, however, that better results can be achieved by means of a weighted (cost-sensitive) version of 0/1 loss. Yet, the real gain is obtained through margin-based loss functions, for which we are able to derive proper bounds, not only for rank risk but, more importantly, also for rank regret. The paper is completed with an experimental study in which we address specific questions raised by our theoretical analysis.

If you are interested in surrogate losses used in the instance ranking problem, it will be a very worth read 🙂

Written by Weiwei

27/06/2011 at 18:20

发表在 学术

A repository of datasets for monotone learning

leave a comment »

We are announcing a monotone learning data repository. The datasets can be downloaded at our KEBI website:

Monotone learning is an important research task in machine learning community and has a wide range of applications. Often, monotone learners lead to more satisfactory predictive performance; and sometimes monotonicity is even required for applications in medical diagnosis, security system design, etc. For example, a system to predict the patient’s overall wellness should have an outcome that is monotonic with respect to the toxicity measure of that patient. No medical doctor will accept a learning model violating this monotonicity.

In our group, we are working on some new ideas for monotone learning. Surprisingly, during our ongoing research we’ve found, despite a tremendous interest on this topic, there is still no public available dataset collection for monotone learning. We’ve worked hard on collecting some benchmark datasets, and we believe it is beneficial to all researchers interested in this topic that we make our collection publicly available.

This datasets collection is still preliminary. There are a number of points on our todo list. In case you have any comment, or if yourself have a monotone learning dataset that you would like to add, please let us know.

Written by Weiwei

05/05/2011 at 00:23

发表在 学术

UNESCO Conference Grant at ECML PKDD

leave a comment »

I was told to receive the UNESCO Conference Grant at the European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD 2010). I’d like to thank the organizer for the great support.

This grant, sponsored by UNESCO Chair in Data Privacy, is aim to offset the cost of attending ECML PKDD 2010 for selected participants.

ECML PKDD 2010 builds upon a very successful series of 20 ECML and 13 PKDD conferences, which have been jointly organized for the past nine years. It has become the major European scientific event in these fields.

感谢ECML PKDD 2010的组织者以及UNESCO Chair in Data Privacy所提供的资助。

Written by Weiwei

10/09/2010 at 02:13

发表在 学术

The key word is learning, of course! Again.

leave a comment »

It is the word cloud extracted from the titles of accepted papers at ECMLPKDD 2010. Click the image for an enlarged picture. It is also available at This year ECMLPKDD has accepted 120 papers with an acceptance rate of 18%.

You can check a similar word cloud extracted from ICML 2010 at my previous post. That makes a good comparison.

Written by Weiwei

02/07/2010 at 14:46

发表在 学术