Gustav Grund Pihlgren

ORCID: 0000-0003-0100-4030
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Generative Adversarial Networks and Image Synthesis
  • Domain Adaptation and Few-Shot Learning
  • Advanced Image Processing Techniques
  • Visual Attention and Saliency Detection
  • Cell Image Analysis Techniques
  • Advanced Neural Network Applications
  • Digital Imaging for Blood Diseases
  • AI in cancer detection
  • Video Analysis and Summarization
  • Anomaly Detection Techniques and Applications
  • Sparse and Compressive Sensing Techniques
  • Radiomics and Machine Learning in Medical Imaging
  • Image Enhancement Techniques
  • Image Retrieval and Classification Techniques
  • Digital Media Forensic Detection
  • Human Pose and Action Recognition
  • Multimodal Machine Learning Applications
  • Explainable Artificial Intelligence (XAI)
  • Topic Modeling
  • Advanced Image and Video Retrieval Techniques
  • Image Processing Techniques and Applications
  • Artificial Intelligence in Games
  • Neural Networks and Applications

Luleå University of Technology
2019-2023

This work presents a novel self-supervised pre-training method to learn efficient representations without labels on histopathology medical images utilizing magnification factors. Other state-of-the-art works mainly focus fully supervised learning approaches that rely heavily human annotations. However, the scarcity of labeled and unlabeled data is long-standing challenge in histopathology. Currently, representation remains unexplored domain. The proposed method, Magnification Prior...

10.1109/wacv56688.2023.00274 article EN 2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) 2023-01-01

Autoencoders are commonly trained using element-wise loss. However, loss disregards high-level structures in the image which can lead to embeddings that disregard them as well. A recent improvement autoencoders helps alleviate this problem is use of perceptual This work investigates from perspective encoder themselves. embed images three different computer vision datasets based on a pretrained model well pixel-wise host predictors perform object positioning and classification given embedded...

10.1109/ijcnn48605.2020.9207431 article EN 2022 International Joint Conference on Neural Networks (IJCNN) 2020-07-01

In this paper, we introduce the use of Semantic Hashing as embedding for task Intent Classification and achieve state-of-the-art performance on three frequently used benchmarks. a small dataset is challenging data-hungry Deep Learning based systems. an attempt to overcome such challenge learn robust text classification. Current word are dependent vocabularies. One major drawbacks methods out-of-vocabulary terms, especially when having training datasets using wider vocabulary. This case in...

10.1109/ijcnn.2019.8852420 article EN 2022 International Joint Conference on Neural Networks (IJCNN) 2019-07-01

This work presents a novel self-supervised pre-training method to learn efficient representations without labels on histopathology medical images utilizing magnification factors. Other state-of-theart works mainly focus fully supervised learning approaches that rely heavily human annotations. However, the scarcity of labeled and unlabeled data is long-standing challenge in histopathology. Currently, representation remains unexplored for domain. The proposed method, Magnification Prior...

10.48550/arxiv.2203.07707 preprint EN cc-by arXiv (Cornell University) 2022-01-01

Perturbation-based post-hoc image explanation methods are commonly used to explain prediction models by perturbing parts of the input measure how those affect output. Due intractability each pixel individually, images typically attributed larger segments. The Randomized Input Sampling for Explanations (RISE) method solved this issue using smooth perturbation masks. While has proven effective and popular, it not been investigated which responsible its success. This work tests many...

10.48550/arxiv.2409.04116 preprint EN arXiv (Cornell University) 2024-09-06

Measuring the similarity of images is a fundamental problem to computer vision for which no universal solution exists. While simple metrics such as pixel-wise L2-norm have been shown significant flaws, they remain popular. One group recent state-of-the-art that mitigates some those flaws are Deep Perceptual Similarity (DPS) metrics, where evaluated distance in deep features neural networks.However, DPS themselves less thoroughly examined their benefits and, especially, flaws. This work...

10.7557/18.6795 article EN Proceedings of the Northern Lights Deep Learning Workshop 2023-01-23

Deep perceptual loss is a type of function in computer vision that aims to mimic human perception by using the deep features extracted from neural networks. In recent years, method has been applied great effect on host interesting tasks, especially for tasks with image or image-like outputs, such as synthesis, segmentation, depth prediction, and more. Many applications use pretrained networks, often convolutional calculation. Despite increased interest broader use, more effort needed toward...

10.48550/arxiv.2302.04032 preprint EN other-oa arXiv (Cornell University) 2023-01-01

This work investigates three methods for calculating loss autoencoder-based pretraining of image encoders: The commonly used reconstruction loss, the more recently introduced deep perceptual similarity and a feature prediction proposed here; latter turning out to be most efficient choice. Standard auto-encoder learning tasks is done by comparing input reconstructed image. Recent shows that predictions based on embeddings generated autoencoders can improved training with i.e., adding network...

10.1109/icpr48806.2021.9412239 article EN 2022 26th International Conference on Pattern Recognition (ICPR) 2021-01-10

Autoencoders are commonly trained using element-wise loss. However, loss disregards high-level structures in the image which can lead to embeddings that disregard them as well. A recent improvement autoencoders helps alleviate this problem is use of perceptual This work investigates from perspective encoder themselves. embed images three different computer vision datasets based on a pretrained model well pixel-wise host predictors perform object positioning and classification given embedded...

10.48550/arxiv.2001.03444 preprint EN other-oa arXiv (Cornell University) 2020-01-01

The concept of image similarity is ambiguous, and images can be similar in one context not another. This ambiguity motivates the creation metrics for specific contexts. work explores ability deep perceptual (DPS) to adapt a given context. DPS use features neural networks comparing images. These have been successful on datasets that leverage average human perception limited settings. But question remains if they could adapted No single metric suit all contexts, previous rule-based are...

10.48550/arxiv.2304.02265 preprint EN other-oa arXiv (Cornell University) 2023-01-01

Measuring the similarity of images is a fundamental problem to computer vision for which no universal solution exists. While simple metrics such as pixel-wise L2-norm have been shown significant flaws, they remain popular. One group recent state-of-the-art that mitigates some those flaws are Deep Perceptual Similarity (DPS) metrics, where evaluated distance in deep features neural networks. However, DPS themselves less thoroughly examined their benefits and, especially, flaws. This work...

10.48550/arxiv.2207.02512 preprint EN other-oa arXiv (Cornell University) 2022-01-01

Automatically identifying harmful content in video is an important task with a wide range of applications. However, there lack professionally labeled open datasets available. In this work VidHarm, dataset 3589 clips from film trailers annotated by professionals, presented. An analysis the performed, revealing among other things relation between clip and trailer level annotations. Audiovisual models are trained on in-depth study modeling choices conducted. The results show that performance...

10.1109/icpr56361.2022.9956148 article EN 2022 26th International Conference on Pattern Recognition (ICPR) 2022-08-21

Automatically identifying harmful content in video is an important task with a wide range of applications. However, there lack professionally labeled open datasets available. In this work VidHarm, dataset 3589 clips from film trailers annotated by professionals, presented. An analysis the performed, revealing among other things relation between clip and trailer level annotations. Audiovisual models are trained on in-depth study modeling choices conducted. The results show that performance...

10.48550/arxiv.2106.08323 preprint EN cc-by-nc-sa arXiv (Cornell University) 2021-01-01
Coming Soon ...