Maciej Wołczyk

ORCID: 0000-0002-3933-9971
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Advanced Neural Network Applications
  • Generative Adversarial Networks and Image Synthesis
  • Domain Adaptation and Few-Shot Learning
  • Adversarial Robustness in Machine Learning
  • Reinforcement Learning in Robotics
  • Neural Networks and Applications
  • Autonomous Vehicle Technology and Safety
  • Multimodal Machine Learning Applications
  • Cell Image Analysis Techniques
  • Machine Learning and Data Classification
  • Advanced Vision and Imaging
  • Machine Learning and Algorithms
  • AI in cancer detection
  • Computer Graphics and Visualization Techniques
  • Advanced Image Processing Techniques
  • Advanced Image and Video Retrieval Techniques
  • Machine Learning in Materials Science
  • Traffic Prediction and Management Techniques
  • Digital Media Forensic Detection
  • Robotic Path Planning Algorithms
  • Multi-Agent Systems and Negotiation
  • Human Pose and Action Recognition
  • Explainable Artificial Intelligence (XAI)
  • Intelligent Tutoring Systems and Adaptive Learning
  • Language, Metaphor, and Cognition

Jagiellonian University
2018-2024

In this paper we present the first safe system for full control of self-driving vehicles trained from human demonstrations and deployed in challenging, real-world, urban environments.Current industry-standard solutions use rulebased systems planning.Although they perform reasonably well common scenarios, engineering complexity renders approach incompatible with human-level performance.On other hand, performance machine-learned (ML) planning can be improved by simply adding more exemplar...

10.1109/icra46639.2022.9811576 article EN 2022 International Conference on Robotics and Automation (ICRA) 2022-05-23

Continual reinforcement learning (CRL) is the study of optimal strategies for maximizing rewards in sequential environments that change over time. This particularly crucial domains such as robotics, where operational environment inherently dynamic and subject to continual change. Nevertheless, research this area has thus far concentrated on off-policy algorithms with replay buffers are capable amortizing impact distribution shifts. Such an approach not feasible on-policy learn solely from...

10.1609/aaai.v39i28.35251 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2025-04-11

The problem of reducing processing time large deep learning models is a fundamental challenge in many real-world applications. Early exit methods strive towards this goal by attaching additional Internal Classifiers (ICs) to intermediate layers neural network. ICs can quickly return predictions for easy examples and, as result, reduce the average inference whole model. However, if particular IC does not decide an answer early, its are discarded, with computations effectively being wasted. To...

10.48550/arxiv.2106.05409 preprint EN cc-by arXiv (Cornell University) 2021-01-01

In this work we are the first to present an offline policy gradient method for learning imitative policies complex urban driving from a large corpus of real-world demonstrations. This is achieved by building differentiable data-driven simulator on top perception outputs and high-fidelity HD maps area. It allows us synthesize new experiences existing demonstrations using mid-level representations. Using then train network in closed-loop employing gradients. We our proposed 100 hours expert...

10.48550/arxiv.2109.13333 preprint EN cc-by arXiv (Cornell University) 2021-01-01

Continual learning (CL) -- the ability to continuously learn, building on previously acquired knowledge is a natural requirement for long-lived autonomous reinforcement (RL) agents. While such agents, one needs balance opposing desiderata, as constraints capacity and compute, not catastrophically forget, exhibit positive transfer new tasks. Understanding right trade-off conceptually computationally challenging, which we argue has led community overly focus catastrophic forgetting. In...

10.48550/arxiv.2105.10919 preprint EN other-oa arXiv (Cornell University) 2021-01-01

We propose a semi-supervised generative model, SeGMA, which learns joint probability distribution of data and their classes is implemented in typical Wasserstein autoencoder framework. choose mixture Gaussians as target latent space, provides natural splitting into clusters. To connect Gaussian components with correct classes, we use small amount labeled classifier induced by the distribution. SeGMA optimized efficiently due to Cramer-Wold distance maximum mean discrepancy penalty, yields...

10.1109/tnnls.2020.3016221 article EN IEEE Transactions on Neural Networks and Learning Systems 2020-08-26

Fine-tuning is a widespread technique that allows practitioners to transfer pre-trained capabilities, as recently showcased by the successful applications of foundation models. However, fine-tuning reinforcement learning (RL) models remains challenge. This work conceptualizes one specific cause poor transfer, accentuated in RL setting interplay between actions and observations: forgetting capabilities. Namely, model deteriorates on state subspace downstream task not visited initial phase...

10.48550/arxiv.2402.02868 preprint EN arXiv (Cornell University) 2024-02-05

Although modern generative models achieve excellent quality in a variety of tasks, they often lack the essential ability to generate examples with requested properties, such as age person photo or weight generated molecule. To overcome these limitations we propose PluGeN (Plugin Generative Network), simple yet effective technique that can be used plugin for pre-trained models. The idea behind our approach is transform entangled latent representation using flow-based module into...

10.1109/tpami.2024.3382008 article EN IEEE Transactions on Pattern Analysis and Machine Intelligence 2024-03-26

Large Language Models (LLMs) and Vision (VLMs) possess extensive knowledge exhibit promising reasoning abilities; however, they still struggle to perform well in complex, dynamic environments. Real-world tasks require handling intricate interactions, advanced spatial reasoning, long-term planning, continuous exploration of new strategies-areas which we lack effective methodologies for comprehensively evaluating these capabilities. To address this gap, introduce BALROG, a novel benchmark...

10.48550/arxiv.2411.13543 preprint EN arXiv (Cornell University) 2024-11-20

The ability of continual learning systems to transfer knowledge from previously seen tasks in order maximize performance on new is a significant challenge for the field, limiting applicability solutions realistic scenarios. Consequently, this study aims broaden our understanding and its driving forces specific case reinforcement learning. We adopt SAC as underlying RL algorithm Continual World suite continuous control tasks. systematically how different components (the actor critic,...

10.48550/arxiv.2209.13900 preprint EN other-oa arXiv (Cornell University) 2022-01-01

The problem of reducing processing time large deep learning models is a fundamental challenge in many real-world applications. Early exit methods strive towards this goal by attaching additional Internal Classifiers (ICs) to intermediate layers neural network. ICs can quickly return predictions for easy examples and, as result, reduce the average inference whole model. However, if particular IC does not decide an answer early, its are discarded, with computations effectively being wasted. To...

10.1016/j.neunet.2023.10.003 article EN cc-by Neural Networks 2023-10-09

Modern generative models achieve excellent quality in a variety of tasks including image or text generation and chemical molecule modeling. However, existing methods often lack the essential ability to generate examples with requested properties, such as age person photo weight generated molecule. Incorporating additional conditioning factors would require rebuilding entire architecture optimizing parameters from scratch. Moreover, it is difficult disentangle selected attributes so that...

10.1609/aaai.v36i8.20843 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2022-06-28

Active Visual Exploration (AVE) is a task that involves dynamically selecting observations (glimpses), which critical to facilitate comprehension and navigation within an environment. While modern AVE methods have demonstrated impressive performance, they are constrained fixed-scale glimpses from rigid grids. In contrast, existing mobile platforms equipped with optical zoom capabilities can capture of arbitrary positions scales. To address this gap between software hardware capabilities, we...

10.48550/arxiv.2404.03482 preprint EN arXiv (Cornell University) 2024-04-04

A new breed of gated-linear recurrent neural networks has reached state-of-the-art performance on a range sequence modeling problems. Such models naturally handle long sequences efficiently, as the cost processing input is independent length. Here, we explore another advantage these stateful models, inspired by success model merging through parameter interpolation. Building parallels between fine-tuning and in-context learning, investigate whether can treat internal states task vectors that...

10.48550/arxiv.2406.08423 preprint EN arXiv (Cornell University) 2024-06-12

Visual perspective-taking (VPT), the ability to understand viewpoint of another person, enables individuals anticipate actions other people. For instance, a driver can avoid accidents by assessing what pedestrians see. Humans typically develop this skill in early childhood, but it remains unclear whether recently emerging Vision Language Models (VLMs) possess such capability. Furthermore, as these models are increasingly deployed real world, understanding how they perform nuanced tasks like...

10.48550/arxiv.2409.12969 preprint EN arXiv (Cornell University) 2024-09-02

We introduce bio-inspired artificial neural networks consisting of neurons that are additionally characterized by spatial positions. To simulate properties biological systems we add the costs penalizing long connections and proximity in a two-dimensional space. Our experiments show case where network performs two different tasks, naturally split into clusters, each cluster is responsible for processing task. This behavior not only corresponds to systems, but also allows further insight...

10.48550/arxiv.1910.02776 preprint EN other-oa arXiv (Cornell University) 2019-01-01

Experience replay is a simple and well-performing strategy for continual learning problems, often used as basis more advanced methods. However, the dynamics of experience are not yet well understood. To showcase this, we focus on single component this problem, namely choosing batch size buffer samples. We find that small batches perform much better at stopping forgetting than larger batches, contrary to intuitive assumption it recall samples from past avoid forgetting. show phenomenon does...

10.1609/aaai.v35i18.17958 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2021-05-18

Schedae Informaticae » 2018 Volume 27 Deep learning-based initialization for object packing A

10.4467/20838476si.18.001.10406 article EN Schedae Informaticae 2018-01-01

We introduce a new training paradigm that enforces interval constraints on neural network parameter space to control forgetting. Contemporary Continual Learning (CL) methods focus networks efficiently from stream of data, while reducing the negative impact catastrophic forgetting, yet they do not provide any firm guarantees performance will deteriorate uncontrollably over time. In this work, we show how put bounds forgetting by reformulating continual learning model as contraction its space....

10.48550/arxiv.2206.07996 preprint EN cc-by arXiv (Cornell University) 2022-01-01
Coming Soon ...