James T. Kwok

ORCID: 0000-0002-4828-8248
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Sparse and Compressive Sensing Techniques
  • Face and Expression Recognition
  • Domain Adaptation and Few-Shot Learning
  • Neural Networks and Applications
  • Machine Learning and Data Classification
  • Machine Learning and ELM
  • Image Retrieval and Classification Techniques
  • Blind Source Separation Techniques
  • Text and Document Classification Technologies
  • Stochastic Gradient Optimization Techniques
  • Advanced Image and Video Retrieval Techniques
  • Machine Learning and Algorithms
  • Image and Signal Denoising Methods
  • Medical Image Segmentation Techniques
  • Speech and Audio Processing
  • Multimodal Machine Learning Applications
  • Advanced Graph Neural Networks
  • Music and Audio Processing
  • Natural Language Processing Techniques
  • Tensor decomposition and applications
  • Advanced Neural Network Applications
  • Recommender Systems and Techniques
  • Machine Learning in Bioinformatics
  • Topic Modeling
  • Anomaly Detection Techniques and Applications

Hong Kong University of Science and Technology
2016-2025

University of Hong Kong
2016-2025

Hong Kong Baptist University
1999-2023

Dominion (United States)
2023

Old Dominion University
2023

Antea Group (France)
2023

IBM (United States)
2023

Xiamen University
2023

University of California, Merced
2023

Beijing University of Posts and Telecommunications
2023

Domain adaptation allows knowledge from a source domain to be transferred different but related target domain. Intuitively, discovering good feature representation across domains is crucial. In this paper, we first propose find such through new learning method, transfer component analysis (TCA), for adaptation. TCA tries learn some components in reproducing kernel Hilbert space using maximum mean miscrepancy. the subspace spanned by these components, data properties are preserved and...

10.1109/tnn.2010.2091281 article EN IEEE Transactions on Neural Networks 2010-11-23

Standard SVM training has O(m3) time and O(m2) space complexities, where m is the set size. It thus computationally infeasible on very large data sets. By observing that practical implementations only approximate optimal solution by an iterative strategy, we scale up kernel methods exploiting such approximateness in this paper. We first show many can be equivalently formulated as minimum enclosing ball (MEB) problems computational geometry. Then, adopting efficient MEB algorithm, obtain...

10.5555/1046920.1058114 article EN Journal of Machine Learning Research 2005-12-01

Machine learning has been highly successful in data-intensive applications but is often hampered when the data set small. Recently, Few-Shot Learning (FSL) proposed to tackle this problem. Using prior knowledge, FSL can rapidly generalize new tasks containing only a few samples with supervised information. In paper, we conduct thorough survey fully understand FSL. Starting from formal definition of FSL, distinguish several relevant machine problems. We then point out that core issue...

10.48550/arxiv.1904.05046 preprint EN other-oa arXiv (Cornell University) 2019-01-01

In this paper, we address the problem of finding pre-image a feature vector in space induced by kernel. This is central importance some kernel applications, such as on using principal component analysis (PCA) for image denoising. Unlike traditional method which relies nonlinear optimization, our proposed directly finds location based distance constraints space. It noniterative, involves only linear algebra and does not suffer from numerical instability or local minimum problems. Evaluations...

10.1109/tnn.2004.837781 article EN IEEE Transactions on Neural Networks 2004-11-01

It is well-known that exploiting label correlations important to multi-label learning. Existing approaches either assume the are global and shared by all instances; or local only a data subset. In fact, in real-world applications, both cases may occur some globally applicable group of instances. Moreover, it also usual case partial labels observed, which makes exploitation much more difficult. That is, hard estimate when many absent. this paper, we propose new approach GLOCAL dealing with...

10.1109/tkde.2017.2785795 article EN IEEE Transactions on Knowledge and Data Engineering 2017-12-21

Low-rank matrix approximation is an effective tool in alleviating the memory and computational burdens of kernel methods sampling, as mainstream such algorithms, has drawn considerable attention both theory practice. This paper presents detailed studies on Nyström sampling scheme particular, error analysis that directly relates quality with encoding powers landmark points summarizing data. The resultant bound suggests a simple efficient scheme, k-means clustering algorithm, for low-rank...

10.1145/1390156.1390311 article EN 2008-01-01

10.1016/s0167-8655(02)00029-6 article EN Pattern Recognition Letters 2002-06-01

Principal component analysis (PCA) has been proven to be an efficient method in pattern recognition and image analysis. Recently, PCA extensively employed for face-recognition algorithms, such as eigenface fisherface. The encouraging results have reported discussed the literature. Many PCA-based systems also developed last decade. However, existing are hard scale up because of computational cost memory-requirement burden. To overcome this limitation, incremental approach is usually adopted....

10.1109/tsmcb.2006.870645 article EN IEEE Transactions on Systems Man and Cybernetics Part B (Cybernetics) 2006-07-21

Kernel (or similarity) matrix plays a key role in many machine learning algorithms such as kernel methods, manifold learning, and dimension reduction. However, the cost of storing manipulating complete makes it infeasible for large problems. The Nyström method is popular sampling-based low-rank approximation scheme reducing computational burdens handling matrices. In this paper, we analyze how approximating quality depends on choice landmark points, particular encoding powers points...

10.1109/tnn.2010.2064786 article EN IEEE Transactions on Neural Networks 2010-09-01

The core vector machine (CVM) is a recent approach for scaling up kernel methods based on the notion of minimum enclosing ball (MEB). Though conceptually simple, an efficient implementation still requires sophisticated numerical solver. In this paper, we introduce (EB) problem where ball's radius fixed and thus does not have to be minimized. We develop (1 + e)-approximation algorithms that are simple implement do require any For Gaussian in particular, suitable choice (fixed) easy determine,...

10.1145/1273496.1273611 article EN 2007-06-20

Many real world learning problems can be recast as multi-task which utilize correlations among different tasks to obtain better generalization performance than each task individually. The feature selection problem in setting has many applications fields of computer vision, text classification and bio-informatics. Generally, it realized by solving a L-1-infinity regularized optimization problem. And the solution automatically yields joint sparsity tasks. However, due nonsmooth nature norm,...

10.1109/icdm.2009.128 article EN 2009-12-01

Kernel methods, such as the support vector machine (SVM), are often formulated quadratic programming (QP) problems. However, given <tex xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">$m$</tex> training patterns, a naive implementation of QP solver takes xmlns:xlink="http://www.w3.org/1999/xlink">$O(m^3)$</tex> time and at least xmlns:xlink="http://www.w3.org/1999/xlink">$O(m^2)$</tex> space. Hence, scaling up these QPs is major stumbling block in...

10.1109/tnn.2006.878123 article EN IEEE Transactions on Neural Networks 2006-09-01

Previous chapter Next Full AccessProceedings Proceedings of the 2009 SIAM International Conference on Data Mining (SDM)Multiple Kernel ClusteringBin Zhao, James T. Kwok, and Changshui ZhangBin Zhangpp.638 - 649Chapter DOI:https://doi.org/10.1137/1.9781611972795.55PDFBibTexSections ToolsAdd to favoritesExport CitationTrack CitationsEmail SectionsAboutAbstract Maximum margin clustering (MMC) has recently attracted considerable interests in both data mining machine learning communities. It...

10.1137/1.9781611972795.55 article EN 2009-04-30

Deep neural network models, though very powerful and highly successful, are computationally expensive in terms of space time. Recently, there have been a number attempts on binarizing the weights activations. This greatly reduces size, replaces underlying multiplications to additions or even XNOR bit operations. However, existing binarization schemes based simple matrix approximation ignore effect loss. In this paper, we propose proximal Newton algorithm with diagonal Hessian that directly...

10.48550/arxiv.1611.01600 preprint EN other-oa arXiv (Cornell University) 2016-01-01
Coming Soon ...