News

MCL Research on Large-Scale Indoor Image Segmentation

Given a point cloud set, the goal of semantic segmentation is to label every point as one of the semantic categories. Semantic segmentation of large-scale point clouds finds a wide range of real-world applications such as autonomous driving in an out-door environment and robotic navigation in an in- or out-door environment. As compared with the point cloud classification problem that often targets at small-scale objects, a high-performance point cloud semantic segmentation method demands a good understanding of the complex global structure as well as the local neighborhood of each point. Meanwhile, efficiency measured by computational complexity and memory complexity is important for practical real-time systems.

State-of-the-art point cloud classification and segmentation methods are based on deep learning. Raw point clouds captured by the LiDAR sensors are irregular and unordered. They cannot be directly processed by deep learning networks designed for 2D images. This problem was addressed by the pioneering work on the PointNet. PointNet and its follow-ups achieved impressive performance in small-scale point cloud classification and segmentation tasks, but they can’t be generalized to handle large-scale point cloud directly due to the memory and time constraints.

An efficient solution to semantic segmentation of large-scale indoor scene point clouds is proposed in this work. It is named GSIP (Green Segmentation of Indoor Point clouds) [1], and its performance is evaluated on a representative large-scale benchmark — the Stanford 3D Indoor Segmentation (S3DIS) dataset. GSIP has two novel components: 1) a room-style data pre-processing method that selects a proper subset of points for further processing, and 2) a new feature extractor which is extended from PointHop. For the former, sampled points of each room form an input unit. For the latter, the weaknesses of PointHop’s feature extraction when [...]

By |November 21st, 2021|News|Comments Off on MCL Research on Large-Scale Indoor Image Segmentation|

MCL Research on Depth Estimation from Images

The target of the depth estimation is to estimate a high quality dense depth map from a single RGB input image. The depth map is an image containing distance information of surface of scene objects from the camera. Depth estimation is crucial for scene understanding, since for accurate scene analysis, more information is helpful. By using the depth estimation, we will have not only the color information from RGB images, but also distance information.

Currently, most depth estimation methods use deep learning by encoder and decoder structure, which are time and computation resources consuming, for example AdaBins[1]. We aim to design a successive subspace learning based method with less computation resources and mathematically explainable, while keeping high performance.

We use the NYU-Depth-v2 dataset[2] for training. We have proposed the method shown in the second image and get some results. In the second image, P represent the RGB images after conversion and D represent the correspondent depth images. In the future, we aim to improve the results by refine the model.

— By Ganning Zhao

Reference:

[1] Bhat, S. F., Alhashim, I., & Wonka, P. (2021). Adabins: Depth estimation using adaptive bins. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 4009-4018).

[2] Silberman, N., Hoiem, D., Kohli, P., & Fergus, R. (2012, October). Indoor segmentation and support inference from rgbd images. In European conference on computer vision (pp. 746-760). Springer, Berlin, Heidelberg.

 

Image credits:

Image showing the architecture of AdaBins is from [1].

Image showing the architecture of our current method.

By |November 14th, 2021|News|Comments Off on MCL Research on Depth Estimation from Images|

Welcome New MCL Member Mahtab Movahhedrad

We are so happy to welcome a new graduate member of MCL, Mahtab Movahhedrad. Here is an interview with Mahtab:

Could you briefly introduce yourself and your research interests?

My name is Mahtab Movahhedrad. I’m from Tabriz, Iran. I started my Ph.D. at USC in spring 2021. I received my bachelor’s and my master’s degree from the University of Tabriz and, Tehran polytechnic respectively. I have been engaged in many different fields of study such as integrated electronics, photonics, metamaterials, and Fourier optics. Now, I plan to gear my Ph.D. Studies more towards signal processing and computer vision and do research in the same area.

What is your impression about MCL and USC?

I have been exploring my passion for computer vision and multimedia for some time now and I think MCL is the best research group in USC when it comes to this field. My motivation to join as a Ph.D. student to MCL is its diverse and outstanding research group in Multimedia. There are about 30 PhD students and several post-doctoral research fellows/visiting scholars from all over the world working in the lab. Doing my PhD at USC, among erudite professors and interested students would create a challenging and at the same time amicable atmosphere for me to broaden my knowledge and collaborate with other scientists.

What is your future expectation and plan in MCL?

My future expectation from the MCL is to learn about green AI and the chance to do research in this field. This method has the potential to accelerate global efforts to protect the environment and conserve resources. I find the concept quite interesting and can’t wait to explore more. I also hope to make lasting connections and effective collaborations with individuals [...]

By |November 7th, 2021|News|Comments Off on Welcome New MCL Member Mahtab Movahhedrad|

MCL Research on Unsupervised Nuclei Image Segmentation

Nuclei segmentation has been extensively used in biological image analysis for reading histology images from microscopes. The population of nuclei, their shape and density play a significant role in clinical practice for cancer diagnosis and its aggressiveness assessment. The reading and annotation of those images is a fairly laborious and time consuming task, being carried out only from expertised pathologists. As such, the computer aided automation of this process is of high significance, since it reduces the physicians’ work load and provides a more objective segmentation output. Yet, different challenges such as color and intensity variations that result from images of different organs and acquisition settings, hinder the performance improvement of the algorithms.

In past years, both supervised and unsupervised solutions have been proposed to tackle those challenges. Most of the recent approaches [1] adopt deep learning (DL) networks, to directly learn from pathologists’ annotated segmentation masks. However, most of the available datasets have very few samples, relatively to the DL requirements, and also their annotations have reportedly a low inter-annotator rate of agreement. Therefore, it is quite challenging for DL models to generalize well to unseen images from multiple organs. At this point, the motivation behind using an unsupervised method should be obvious. We propose a data driven and parameter free methodology [2], named CBM, for the nuclei segmentation task that requires no training labels. The pipeline begins with a data-driven Color (C) transform, to highlight the nuclei cell regions over the background. Then, a data-driven Binarization (B) process is built on the bi-modal assumption that each local region has two histogram peaks, one corresponding to the background and one to the nuclei areas. We use a locally adaptive threshold to binarize each histology [...]

By |October 31st, 2021|News|Comments Off on MCL Research on Unsupervised Nuclei Image Segmentation|
  • Permalink Gallery

    Congratulations to MCL Alum, Professor Kyoung Mu Lee, to be Appointed as EiC of IEEE PAMI

Congratulations to MCL Alum, Professor Kyoung Mu Lee, to be Appointed as EiC of IEEE PAMI

The 2nd Ph.D. alumnus of MCL, Professor Kyoung Mu Lee of Seoul National University (SNU), has been appointed as the Editor-in-Chief (EiC) of the prestigious IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI) from 2022 to 2024. TPAMI publishes articles on areas of computer vision and image understanding, all traditional areas of pattern analysis and recognition, and selected areas of machine intelligence. It is one of the premier journals in all of computer science. Its excellence, combined with its focus on computer vision and machine learning, positions it as one of IEEE’s flagship journals. Its impact factor is 16.39 in 2020.

Kyoung Mu Lee received his Ph. D. degree from the University of Southern California in 1993. He is currently a professor in the department of ECE and the director of the Interdisciplinary Graduate Program in Artificial Intelligence at Seoul National University (SNU). His primary research areas are Computer Vision and Machine Learning. He has served as an Editorial board member of many journals, including an Associate Editor in Chief (AEiC) of the IEEE TPAMI, Area Editor of the Computer Vision and Image Understanding (CVIU), and Associate Editor (AE) of the IEEE TPAMI, the Machine Vision Application (MVA), the IPSJ Transactions on Computer Vision and Applications (CVA), and the IEEE Signal Processing Letter. He is an Advisory Board Member of the Computer Vision Foundation (CVF) and an Editorial Advisory Board Member for Academic Press/Elsevier. He has served as a General co-Chair of the prestigious ICCV2019, ACM MM2018, ACCV2018, and an Area Chair of CVPR, ICCV, and ECCV many times. He was a Distinguished Lecturer of the Asia-Pacific Signal and Information Processing Association (APSIPA) for 2012-2013. He is currently serving as the president of the [...]

By |October 24th, 2021|News|Comments Off on Congratulations to MCL Alum, Professor Kyoung Mu Lee, to be Appointed as EiC of IEEE PAMI|
  • Permalink Gallery

    MCL Research on Negative Sampling for Knowledge Graph Learning

MCL Research on Negative Sampling for Knowledge Graph Learning

A knowledge graph is a collection of factual triples (h, r, t) consisting of two entities and one relation. Most knowledge graphs suffer from the incompleteness that there are many missing relations between entities. To predict missing links, each relation is modeled by a binary classifier to predict whether the links between two entities exist or not. Negative sampling is a task to draw negative samples efficiently and effectively from the unobserved triples to train the classifiers. The quality and quantity of the negative samples will highly affect the performance on link prediction.
Naive negative sampling [1] suggests generating negative samples by corrupting one of the entities in the observed triples, e.g. (h’, r, t) or (h, r, t’). Despite the simplicity of naive negative samples, the generated negative samples carry little semantics. For example, given a positive triple (Hulk, movie_genre, Science Fiction), a negative sample (Hulk, movie_genre, New York City) might be generated by naive negative sampling, which will never be a valid triple in the real-world scenario. Instead, we are looking for negative examples, such as (Hulk, movie_genre, Romance), that provide more information to the classifiers. Based on the observation, we only draw the corrupted entities within the set of observed entities that have been linked by the given relation, also known as the ‘range’ for the relation. However, a drawback is that the chance of drawing false negatives is high. Therefore, we further filter the drawn corrupted entities based on the entity-entity co-occurrence. For example, it’s not likely for us to generate a negative sample (Hulk, movie_genre, Adventure) because we know from the dataset that movie genres ‘Science Fiction’ and ‘Adventure’ are highly co-occurred. The first figure shows how the positive [...]

By |October 17th, 2021|News|Comments Off on MCL Research on Negative Sampling for Knowledge Graph Learning|
  • Permalink Gallery

    MCL Research on Type Prediction for Knowledge Graph Learning

MCL Research on Type Prediction for Knowledge Graph Learning

Entity type is a very important piece of information in Knowledge Graphs. Researchers have leveraged entity type information to get better results in many Knowledge Graph related tasks such as link prediction. Besides, entity type is also important for Information Extraction tasks including entity linking and relation extraction. However, Knowledge Graph entity type information is often incomplete and noisy. Therefore, there is a need to develop effective algorithms for predicting missing types for entities.

Knowledge Graph (KG) Embeddings in complex vector space have demonstrated superior performance in relation prediction and triple classification. Representing entities and relations in complex space has several advantages than traditional models such as better expressive power, and better capabilities of modeling one-to-many and asymmetric relations. We leverage these characteristics of complex KG Embeddings and formulate the type prediction problem as a complex space regression problem. Experimental results confirm our hypothesis that the expressiveness of embedding models correlates with the performance on type prediction. Our newly proposed method achieves state-of-the-art results in type prediction for many benchmarking datasets.

[1] Sun, Zhiqing, et al. “Rotate: Knowledge graph embedding by relational rotation in complex space.” arXiv preprint arXiv:1902.10197 (2019).

[2] Zhao, Yu, et al. “Connecting embeddings for knowledge graph entity typing.” arXiv preprint arXiv:2007.10873 (2020).

— by Xiou Ge

By |October 11th, 2021|News|Comments Off on MCL Research on Type Prediction for Knowledge Graph Learning|

Introduction to MCL New Visitor – Rafael Luiz Testa

In Fall 2021, we have a new MCL member, Rafael Luiz Testa, joining our big family. Here is a short interview with Rafael with our great welcome.

1. Could you briefly introduce yourself and your research interests?

My name is Rafael Luiz Testa. I received Bachelor’s (2014) and Master’s (2018) degrees in Information Systems from the University of São Paulo, Brazil. I am currently pursuing my PhD degree at the University of São Paulo with a nine-month stay at MCL/USC. I have been working on computer graphics since 2012. In 2013, I joined a project to help people with psychiatric disorders to recognize emotions in facial expressions. My main research interests are image/video analysis and synthesis, as well as facial expression.

2. What is your impression about MCL and USC?

The USC and MCL are both internationally recognized for the quality of their work. I am so happy that I found here at MCL a place where everyone has plenty of opportunities to achieve their best. The MCL Director, Dr. C.-C. Jay Kuo, instigates innovation and collaboration between students. Furthermore, the MCL has such a friendly and kind environment.

3. What is your future expectation and plan in MCL?

I believe my stay at MCL will give me an opportunity to see my research from an entirely novel perspective. I hope I can enjoy to the fullest such an innovative environment. In the future, I would like to be a professor and pursue an academic research career. Thus, I am confident that the work developed at MCL will significantly impact my future projects and help me achieve my career goals.

By |October 3rd, 2021|News|Comments Off on Introduction to MCL New Visitor – Rafael Luiz Testa|

MCL Research on GAN-generated Fake Images Detection

In recent years, there has been a rapid development of image synthesis techniques based on convolutional neural networks (CNNs), such as the variational auto-encoder (VAE) and generative adversarial networks (GANs). They have shown their ability to generate realistic images that are hard for people to tell which is fake and which is real. Most state-of-the-art CNN generated image detection methods are formulated on deep neural networks. However, their performance can be easily restrained on specific fake image datasets and fail to generalize well to other datasets.

We propose a new CNN-generated-image detector, named Attentive PixelHop (or A-PixelHop). A-PixelHop is designed under the assumption that it is difficult to synthesize high-quality high-frequency components in local regions. Specifically, we first select edge/texture blocks that contain significant high frequency components, then apply multiple filter banks to them to obtain rich sets of spatial-spectral responses as features. Different filter bank features may have different importance on the deciding fake and real, therefore, we feed features to multiple binary classifiers to obtain a set of soft decisions, and we only select the ones with highest discrimination ability. Finally, we develop an effective ensemble scheme to fuse the soft decisions from more discriminant channels into the final decision. System design is shown in Figure 1 below. Compared with CNN-based fake image detection methods, our method has low computational complexity and a small model size, high detection performance against a wide range of generative models, and mathematical transparency since. Experimental results show that A-PixelHop outperforms all state-of-the-art benchmarking methods for CycleGAN-generated images, see Table 1. Furthermore, it can generalize well to unseen generative models and datasets, see Table 2.

By |October 3rd, 2021|News|Comments Off on MCL Research on GAN-generated Fake Images Detection|

MCL Research on Geographic Fake Images Detection

Misinformation on the Internet and social media, ranging from fake news to fake multimedia such as images and videos, is a significant threat to our society. Effective misinformation detection has become a research focus, driven by commercial and government funding. With the fast-growing deep learning techniques, real-looking fake images can be easily generated using generative adversarial networks (GANs). The problem of fake satellite images detection was recently introduced. Fake satellite images could be generated with the intention of hiding important infrastructure and/or creating fake buildings to deceive others. Although it may be feasible to check whether these images are real or fake using another satellite, the cost is high. Furthermore, the general public and media do not have the proper resource to verify the authenticity of fake satellite images. Consequently, fake satellite images pose serious challenges to our society, as recognized by government organizations concerned about the political and military implications of such technology. Handcrafted features were used for fake satellite image detection, and its best detection performance measured by the F-1 score is 87%. 

A new method, called PSL-DefakeHop, is proposed to detect fake satellite images based on the parallel subspace learning (PSL) framework in this work. The DefakeHop method was developed previously for the detection of Deepfake generated faces under the successive subspace learning (SSL) framework. PSL is proposed to extract features from responses of multiple single-stage filter banks (or called PixelHops), which operate in parallel, and it improves SSL that extracts features from multi-stage cascaded filter banks. PSL has two advantages. First, PSL preserves discriminant features often lie in high-frequency channels, which are however ignored by SSL. Second, decisions from multiple filter banks can be ensembled to further improve detection accuracy. To [...]

By |September 20th, 2021|News|Comments Off on MCL Research on Geographic Fake Images Detection|