arxiv_id
stringlengths 7
11
| title
stringlengths 7
243
| abstract
stringlengths 3
2.79k
| link
stringlengths 21
49
| authors
sequencelengths 1
451
| updated
stringlengths 20
20
| published
stringlengths 20
20
|
---|---|---|---|---|---|---|
2403.14608 | Parameter-Efficient Fine-Tuning for Large Models: A Comprehensive Survey | Large models represent a groundbreaking advancement in multiple application fields, enabling remarkable achievements across various tasks. However, their unprecedented scale comes with significant computational costs. These models, often consisting of billions of parameters, require vast amounts of computational resources for execution. Especially, the expansive scale and computational demands pose considerable challenges when customizing them for particular downstream tasks, particularly over the hardware platforms constrained by computational capabilities. Parameter Efficient Fine-Tuning (PEFT) provides a practical solution by efficiently adjusting the large models over the various downstream tasks. In particular, PEFT refers to the process of adjusting the parameters of a pre-trained large models to adapt it to a specific task or domain while minimizing the number of additional parameters introduced or computational resources required. This approach is particularly important when dealing with large-scale language models with high parameter counts, as fine-tuning these models from scratch can be computationally expensive and resource-intensive, posing considerable challenges in the supporting system platform design. In this survey, we present comprehensive studies of various PEFT algorithms, examining their performance and computational overhead. Moreover, we provide an overview of applications developed using different PEFT algorithms and discuss common techniques employed to mitigate computation costs for PEFT. In addition to providing an extensive survey from an algorithmic standpoint, we also examine various real-world system designs to investigate the implementation costs associated with different PEFT approaches. This survey serves as an indispensable resource for researchers aiming to understand both the PEFT algorithm and its system implementation, offering detailed ...... | http://arxiv.org/pdf/2403.14608v6 | [
"Zeyu Han",
"Chao Gao",
"Jinyang Liu",
"Jeff Zhang",
"Sai Qian Zhang"
] | 2024-07-12T09:58:10Z | 2024-03-21T17:55:50Z |
2407.09127 | Robustness of Explainable Artificial Intelligence in Industrial Process
Modelling | eXplainable Artificial Intelligence (XAI) aims at providing understandable explanations of black box models. In this paper, we evaluate current XAI methods by scoring them based on ground truth simulations and sensitivity analysis. To this end, we used an Electric Arc Furnace (EAF) model to better understand the limits and robustness characteristics of XAI methods such as SHapley Additive exPlanations (SHAP), Local Interpretable Model-agnostic Explanations (LIME), as well as Averaged Local Effects (ALE) or Smooth Gradients (SG) in a highly topical setting. These XAI methods were applied to various types of black-box models and then scored based on their correctness compared to the ground-truth sensitivity of the data-generating processes using a novel scoring evaluation methodology over a range of simulated additive noise. The resulting evaluation shows that the capability of the Machine Learning (ML) models to capture the process accurately is, indeed, coupled with the correctness of the explainability of the underlying data-generating process. We furthermore show the differences between XAI methods in their ability to correctly predict the true sensitivity of the modeled industrial process. | http://arxiv.org/pdf/2407.09127v1 | [
"Benedikt Kantz",
"Clemens Staudinger",
"Christoph Feilmayr",
"Johannes Wachlmayr",
"Alexander Haberl",
"Stefan Schuster",
"Franz Pernkopf"
] | 2024-07-12T09:46:26Z | 2024-07-12T09:46:26Z |
2302.01029 | On Suppressing Range of Adaptive Stepsizes of Adam to Improve
Generalisation Performance | A number of recent adaptive optimizers improve the generalisation performance of Adam by essentially reducing the variance of adaptive stepsizes to get closer to SGD with momentum. Following the above motivation, we suppress the range of the adaptive stepsizes of Adam by exploiting the layerwise gradient statistics. In particular, at each iteration, we propose to perform three consecutive operations on the second momentum v_t before using it to update a DNN model: (1): down-scaling, (2): epsilon-embedding, and (3): down-translating. The resulting algorithm is referred to as SET-Adam, where SET is a brief notation of the three operations. The down-scaling operation on v_t is performed layerwise by making use of the angles between the layerwise subvectors of v_t and the corresponding all-one subvectors. Extensive experimental results show that SET-Adam outperforms eight adaptive optimizers when training transformers and LSTMs for NLP, and VGG and ResNet for image classification over CIAF10 and CIFAR100 while matching the best performance of the eight adaptive methods when training WGAN-GP models for image generation tasks. Furthermore, SET-Adam produces higher validation accuracies than Adam and AdaBelief for training ResNet18 over ImageNet. | http://arxiv.org/pdf/2302.01029v3 | [
"Guoqiang Zhang"
] | 2024-07-12T09:46:14Z | 2023-02-02T11:46:23Z |
2407.09124 | Decentralized multi-agent reinforcement learning algorithm using a
cluster-synchronized laser network | Multi-agent reinforcement learning (MARL) studies crucial principles that are applicable to a variety of fields, including wireless networking and autonomous driving. We propose a photonic-based decision-making algorithm to address one of the most fundamental problems in MARL, called the competitive multi-armed bandit (CMAB) problem. Our numerical simulations demonstrate that chaotic oscillations and cluster synchronization of optically coupled lasers, along with our proposed decentralized coupling adjustment, efficiently balance exploration and exploitation while facilitating cooperative decision-making without explicitly sharing information among agents. Our study demonstrates how decentralized reinforcement learning can be achieved by exploiting complex physical processes controlled by simple algorithms. | http://arxiv.org/pdf/2407.09124v1 | [
"Shun Kotoku",
"Takatomo Mihana",
"André Röhm",
"Ryoichi Horisaki"
] | 2024-07-12T09:38:47Z | 2024-07-12T09:38:47Z |
2407.09120 | URRL-IMVC: Unified and Robust Representation Learning for Incomplete
Multi-View Clustering | Incomplete multi-view clustering (IMVC) aims to cluster multi-view data that are only partially available. This poses two main challenges: effectively leveraging multi-view information and mitigating the impact of missing views. Prevailing solutions employ cross-view contrastive learning and missing view recovery techniques. However, they either neglect valuable complementary information by focusing only on consensus between views or provide unreliable recovered views due to the absence of supervision. To address these limitations, we propose a novel Unified and Robust Representation Learning for Incomplete Multi-View Clustering (URRL-IMVC). URRL-IMVC directly learns a unified embedding that is robust to view missing conditions by integrating information from multiple views and neighboring samples. Firstly, to overcome the limitations of cross-view contrastive learning, URRL-IMVC incorporates an attention-based auto-encoder framework to fuse multi-view information and generate unified embeddings. Secondly, URRL-IMVC directly enhances the robustness of the unified embedding against view-missing conditions through KNN imputation and data augmentation techniques, eliminating the need for explicit missing view recovery. Finally, incremental improvements are introduced to further enhance the overall performance, such as the Clustering Module and the customization of the Encoder. We extensively evaluate the proposed URRL-IMVC framework on various benchmark datasets, demonstrating its state-of-the-art performance. Furthermore, comprehensive ablation studies are performed to validate the effectiveness of our design. | http://arxiv.org/abs/2407.09120v1 | [
"Ge Teng",
"Ting Mao",
"Chen Shen",
"Xiang Tian",
"Xuesong Liu",
"Yaowu Chen",
"Jieping Ye"
] | 2024-07-12T09:35:25Z | 2024-07-12T09:35:25Z |
2401.10158 | DISTINQT: A Distributed Privacy Aware Learning Framework for QoS
Prediction for Future Mobile and Wireless Networks | Beyond 5G and 6G networks are expected to support new and challenging use cases and applications that depend on a certain level of Quality of Service (QoS) to operate smoothly. Predicting the QoS in a timely manner is of high importance, especially for safety-critical applications as in the case of vehicular communications. Although until recent years the QoS prediction has been carried out by centralized Artificial Intelligence (AI) solutions, a number of privacy, computational, and operational concerns have emerged. Alternative solutions have surfaced (e.g. Split Learning, Federated Learning), distributing AI tasks of reduced complexity across nodes, while preserving the privacy of the data. However, new challenges rise when it comes to scalable distributed learning approaches, taking into account the heterogeneous nature of future wireless networks. The current work proposes DISTINQT, a novel multi-headed input privacy-aware distributed learning framework for QoS prediction. Our framework supports multiple heterogeneous nodes, in terms of data types and model architectures, by sharing computations across them. This enables the incorporation of diverse knowledge into a sole learning process that will enhance the robustness and generalization capabilities of the final QoS prediction model. DISTINQT also contributes to data privacy preservation by encoding any raw input data into highly complex, compressed, and irreversible latent representations before any transmission. Evaluation results showcase that DISTINQT achieves a statistically identical performance compared to its centralized version, while also proving the validity of the privacy preserving claims. DISTINQT manages to achieve a reduction in prediction error of up to 65% on average against six state-of-the-art centralized baseline solutions presented in the Tele-Operated Driving use case. | http://arxiv.org/pdf/2401.10158v2 | [
"Nikolaos Koursioumpas",
"Lina Magoula",
"Ioannis Stavrakakis",
"Nancy Alonistioti",
"M. A. Gutierrez-Estevez",
"Ramin Khalili"
] | 2024-07-12T09:27:57Z | 2024-01-15T13:00:48Z |
2407.09111 | Inference Optimization of Foundation Models on AI Accelerators | Powerful foundation models, including large language models (LLMs), with Transformer architectures have ushered in a new era of Generative AI across various industries. Industry and research community have witnessed a large number of new applications, based on those foundation models. Such applications include question and answer, customer services, image and video generation, and code completions, among others. However, as the number of model parameters reaches to hundreds of billions, their deployment incurs prohibitive inference costs and high latency in real-world scenarios. As a result, the demand for cost-effective and fast inference using AI accelerators is ever more higher. To this end, our tutorial offers a comprehensive discussion on complementary inference optimization techniques using AI accelerators. Beginning with an overview of basic Transformer architectures and deep learning system frameworks, we deep dive into system optimization techniques for fast and memory-efficient attention computations and discuss how they can be implemented efficiently on AI accelerators. Next, we describe architectural elements that are key for fast transformer inference. Finally, we examine various model compression and fast decoding strategies in the same context. | http://arxiv.org/pdf/2407.09111v1 | [
"Youngsuk Park",
"Kailash Budhathoki",
"Liangfu Chen",
"Jonas Kübler",
"Jiaji Huang",
"Matthäus Kleindessner",
"Jun Huan",
"Volkan Cevher",
"Yida Wang",
"George Karypis"
] | 2024-07-12T09:24:34Z | 2024-07-12T09:24:34Z |
2407.09105 | Enhancing Training Efficiency Using Packing with Flash Attention | Padding is often used in tuning LLM models by adding special tokens to shorter training examples to match the length of the longest sequence in each batch. While this ensures uniformity for batch processing, it introduces inefficiencies by including irrelevant padding tokens in the computation and wastes GPU resources. On the other hand, the Hugging Face SFT trainer offers the option to use packing to combine multiple training examples up to the maximum sequence length. This allows for maximal utilization of GPU resources. However, without proper masking of each packed training example, attention will not be computed correctly when using SFT trainer. We enable and then analyse packing and Flash Attention with proper attention masking of each example and show the benefits of this training paradigm. | http://arxiv.org/pdf/2407.09105v1 | [
"Achintya Kundu",
"Rhui Dih Lee",
"Laura Wynter",
"Raghu Kiran Ganti"
] | 2024-07-12T09:10:37Z | 2024-07-12T09:10:37Z |
2407.09104 | UserBoost: Generating User-specific Synthetic Data for Faster Enrolment
into Behavioural Biometric Systems | Behavioural biometric authentication systems entail an enrolment period that is burdensome for the user. In this work, we explore generating synthetic gestures from a few real user gestures with generative deep learning, with the application of training a simple (i.e. non-deep-learned) authentication model. Specifically, we show that utilising synthetic data alongside real data can reduce the number of real datapoints a user must provide to enrol into a biometric system. To validate our methods, we use the publicly available dataset of WatchAuth, a system proposed in 2022 for authenticating smartwatch payments using the physical gesture of reaching towards a payment terminal. We develop a regularised autoencoder model for generating synthetic user-specific wrist motion data representing these physical gestures, and demonstrate the diversity and fidelity of our synthetic gestures. We show that using synthetic gestures in training can improve classification ability for a real-world system. Through this technique we can reduce the number of gestures required to enrol a user into a WatchAuth-like system by more than 40% without negatively impacting its error rates. | http://arxiv.org/pdf/2407.09104v1 | [
"George Webber",
"Jack Sturgess",
"Ivan Martinovic"
] | 2024-07-12T09:10:07Z | 2024-07-12T09:10:07Z |
2407.09096 | STD-LLM: Understanding Both Spatial and Temporal Properties of
Spatial-Temporal Data with LLMs | Spatial-temporal forecasting and imputation are important for real-world dynamic systems such as intelligent transportation, urban planning, and public health. Most existing methods are tailored for individual forecasting or imputation tasks but are not designed for both. Additionally, they are less effective for zero-shot and few-shot learning. While large language models (LLMs) have exhibited strong pattern recognition and reasoning abilities across various tasks, including few-shot and zero-shot learning, their development in understanding spatial-temporal data has been constrained by insufficient modeling of complex correlations such as the temporal correlations, spatial connectivity, non-pairwise and high-order spatial-temporal correlations within data. In this paper, we propose STD-LLM for understanding both spatial and temporal properties of underline{S}patial-underline{T}emporal underline{D}ata with underline{LLM}s, which is capable of implementing both spatial-temporal forecasting and imputation tasks. STD-LLM understands spatial-temporal correlations via explicitly designed spatial and temporal tokenizers as well as virtual nodes. Topology-aware node embeddings are designed for LLMs to comprehend and exploit the topology structure of data. Additionally, to capture the non-pairwise and higher-order correlations, we design a hypergraph learning module for LLMs, which can enhance the overall performance and improve efficiency. Extensive experiments demonstrate that STD-LLM exhibits strong performance and generalization capabilities across the forecasting and imputation tasks on various datasets. Moreover, STD-LLM achieves promising results on both few-shot and zero-shot learning tasks. | http://arxiv.org/pdf/2407.09096v1 | [
"Yiheng Huang",
"Xiaowei Mao",
"Shengnan Guo",
"Yubin Chen",
"Youfang Lin",
"Huaiyu Wan"
] | 2024-07-12T08:48:16Z | 2024-07-12T08:48:16Z |
2407.09093 | On Exact Bit-level Reversible Transformers Without Changing
Architectures | In the literature, various reversible deep neural networks (DNN) models have been proposed to reduce memory consumption or improve data-throughput in the training process. However, almost all existing reversible DNNs either are constrained to have special structures or are constructed by modifying the original DNN architectures considerably to enable reversibility. In this work, we propose exact bit-level reversible transformers without changing the architectures in the inference procedure. The basic idea is to first treat each transformer block as the Euler integration approximation for solving an ordinary differential equation (ODE) and then incorporate the technique of bidirectional integration approximation (BDIA) (see [26]) for BDIA-based diffusion inversion) into the neural architecture together with activation quantization to make it exactly bit-level reversible, referred to as BDIA-transformer. In the training process, we let a hyper-parameter $gamma$ in BDIA-transformer randomly take one of the two values ${0.5, -0.5}$ per transformer block for averaging two consecutive integration approximations, which regularizes the models for improving the validation accuracy. Light-weight side information per transformer block is required to be stored in the forward process to account for binary quantization loss to enable exact bit-level reversibility. In the inference procedure, the expectation $mathbb{E}(gamma)=0$ is taken to make the resulting architectures of BDIA-transformer be identical to transformers up to activation quantization. Empirical study indicates that BDIA-transformers outperform their original counterparts notably due to the regularization effect of the $gamma$ parameter. | http://arxiv.org/pdf/2407.09093v1 | [
"Guoqiang Zhang",
"J. P. Lewis",
"W. B. Kleijn"
] | 2024-07-12T08:42:58Z | 2024-07-12T08:42:58Z |
2402.06165 | Learning Contrastive Feature Representations for Facial Action Unit
Detection | Facial action unit (AU) detection has long encountered the challenge of detecting subtle feature differences when AUs activate. Existing methods often rely on encoding pixel-level information of AUs, which not only encodes additional redundant information but also leads to increased model complexity and limited generalizability. Additionally, the accuracy of AU detection is negatively impacted by the class imbalance issue of each AU type, and the presence of noisy and false AU labels. In this paper, we introduce a novel contrastive learning framework aimed for AU detection that incorporates both self-supervised and supervised signals, thereby enhancing the learning of discriminative features for accurate AU detection. To tackle the class imbalance issue, we employ a negative sample re-weighting strategy that adjusts the step size of updating parameters for minority and majority class samples. Moreover, to address the challenges posed by noisy and false AU labels, we employ a sampling technique that encompasses three distinct types of positive sample pairs. This enables us to inject self-supervised signals into the supervised signal, effectively mitigating the adverse effects of noisy labels. Our experimental assessments, conducted on four widely-utilized benchmark datasets (BP4D, DISFA, GFT and Aff-Wild2), underscore the superior performance of our approach compared to state-of-the-art methods of AU detection. Our code is available at url{https://github.com/Ziqiao-Shang/AUNCE}. | http://arxiv.org/pdf/2402.06165v2 | [
"Ziqiao Shang",
"Bin Liu",
"Fengmao Lv",
"Fei Teng",
"Tianrui Li"
] | 2024-07-12T08:41:21Z | 2024-02-09T03:48:20Z |
2407.04302 | Fair Federated Data Clustering through Personalization: Bridging the Gap
between Diverse Data Distributions | The rapid growth of data from edge devices has catalyzed the performance of machine learning algorithms. However, the data generated resides at client devices thus there are majorly two challenge faced by traditional machine learning paradigms - centralization of data for training and secondly for most the generated data the class labels are missing and there is very poor incentives to clients to manually label their data owing to high cost and lack of expertise. To overcome these issues, there have been initial attempts to handle unlabelled data in a privacy preserving distributed manner using unsupervised federated data clustering. The goal is partition the data available on clients into $k$ partitions (called clusters) without actual exchange of data. Most of the existing algorithms are highly dependent on data distribution patterns across clients or are computationally expensive. Furthermore, due to presence of skewed nature of data across clients in most of practical scenarios existing models might result in clients suffering high clustering cost making them reluctant to participate in federated process. To this, we are first to introduce the idea of personalization in federated clustering. The goal is achieve balance between achieving lower clustering cost and at same time achieving uniform cost across clients. We propose p-FClus that addresses these goal in a single round of communication between server and clients. We validate the efficacy of p-FClus against variety of federated datasets showcasing it's data independence nature, applicability to any finite $ell$-norm, while simultaneously achieving lower cost and variance. | http://arxiv.org/pdf/2407.04302v2 | [
"Shivam Gupta",
"Tarushi",
"Tsering Wangzes",
"Shweta Jain"
] | 2024-07-12T08:35:33Z | 2024-07-05T07:10:26Z |
2407.09087 | On the Role of Discrete Tokenization in Visual Representation Learning | In the realm of self-supervised learning (SSL), masked image modeling (MIM) has gained popularity alongside contrastive learning methods. MIM involves reconstructing masked regions of input images using their unmasked portions. A notable subset of MIM methodologies employs discrete tokens as the reconstruction target, but the theoretical underpinnings of this choice remain underexplored. In this paper, we explore the role of these discrete tokens, aiming to unravel their benefits and limitations. Building upon the connection between MIM and contrastive learning, we provide a comprehensive theoretical understanding on how discrete tokenization affects the model's generalization capabilities. Furthermore, we propose a novel metric named TCAS, which is specifically designed to assess the effectiveness of discrete tokens within the MIM framework. Inspired by this metric, we contribute an innovative tokenizer design and propose a corresponding MIM method named ClusterMIM. It demonstrates superior performance on a variety of benchmark datasets and ViT backbones. Code is available at https://github.com/PKU-ML/ClusterMIM. | http://arxiv.org/pdf/2407.09087v1 | [
"Tianqi Du",
"Yifei Wang",
"Yisen Wang"
] | 2024-07-12T08:25:31Z | 2024-07-12T08:25:31Z |
2406.14393 | Jailbreaking as a Reward Misspecification Problem | The widespread adoption of large language models (LLMs) has raised concerns about their safety and reliability, particularly regarding their vulnerability to adversarial attacks. In this paper, we propose a novel perspective that attributes this vulnerability to reward misspecification during the alignment process. We introduce a metric ReGap to quantify the extent of reward misspecification and demonstrate its effectiveness and robustness in detecting harmful backdoor prompts. Building upon these insights, we present ReMiss, a system for automated red teaming that generates adversarial prompts against various target aligned LLMs. ReMiss achieves state-of-the-art attack success rates on the AdvBench benchmark while preserving the human readability of the generated prompts. Detailed analysis highlights the unique advantages brought by the proposed reward misspecification objective compared to previous methods. | http://arxiv.org/pdf/2406.14393v2 | [
"Zhihui Xie",
"Jiahui Gao",
"Lei Li",
"Zhenguo Li",
"Qi Liu",
"Lingpeng Kong"
] | 2024-07-12T08:15:45Z | 2024-06-20T15:12:27Z |
2406.02609 | Less is More: Pseudo-Label Filtering for Continual Test-Time Adaptation | Continual Test-Time Adaptation (CTTA) aims to adapt a pre-trained model to a sequence of target domains during the test phase without accessing the source data. To adapt to unlabeled data from unknown domains, existing methods rely on constructing pseudo-labels for all samples and updating the model through self-training. However, these pseudo-labels often involve noise, leading to insufficient adaptation. To improve the quality of pseudo-labels, we propose a pseudo-label selection method for CTTA, called Pseudo Labeling Filter (PLF). The key idea of PLF is to keep selecting appropriate thresholds for pseudo-labels and identify reliable ones for self-training. Specifically, we present three principles for setting thresholds during continuous domain learning, including initialization, growth and diversity. Based on these principles, we design Self-Adaptive Thresholding to filter pseudo-labels. Additionally, we introduce a Class Prior Alignment (CPA) method to encourage the model to make diverse predictions for unknown domain samples. Through extensive experiments, PLF outperforms current state-of-the-art methods, proving its effectiveness in CTTA. | http://arxiv.org/pdf/2406.02609v2 | [
"Jiayao Tan",
"Fan Lyu",
"Chenggong Ni",
"Tingliang Feng",
"Fuyuan Hu",
"Zhang Zhang",
"Shaochuang Zhao",
"Liang Wang"
] | 2024-07-12T08:15:22Z | 2024-06-03T04:09:36Z |
2407.09064 | Multi-Modal Dataset Creation for Federated~Learning with DICOM
Structured Reports | Purpose: Federated training is often hindered by heterogeneous datasets due to divergent data storage options, inconsistent naming schemes, varied annotation procedures, and disparities in label quality. This is particularly evident in the emerging multi-modal learning paradigms, where dataset harmonization including a uniform data representation and filtering options are of paramount importance. Methods: DICOM structured reports enable the standardized linkage of arbitrary information beyond the imaging domain and can be used within Python deep learning pipelines with highdicom. Building on this, we developed an open platform for data integration and interactive filtering capabilities that simplifies the process of assembling multi-modal datasets. Results: In this study, we extend our prior work by showing its applicability to more and divergent data types, as well as streamlining datasets for federated training within an established consortium of eight university hospitals in Germany. We prove its concurrent filtering ability by creating harmonized multi-modal datasets across all locations for predicting the outcome after minimally invasive heart valve replacement. The data includes DICOM data (i.e. computed tomography images, electrocardiography scans) as well as annotations (i.e. calcification segmentations, pointsets and pacemaker dependency), and metadata (i.e. prosthesis and diagnoses). Conclusion: Structured reports bridge the traditional gap between imaging systems and information systems. Utilizing the inherent DICOM reference system arbitrary data types can be queried concurrently to create meaningful cohorts for clinical studies. The graphical interface as well as example structured report templates will be made publicly available. | http://arxiv.org/pdf/2407.09064v1 | [
"Malte Tölle",
"Lukas Burger",
"Halvar Kelm",
"Florian André",
"Peter Bannas",
"Gerhard Diller",
"Norbert Frey",
"Philipp Garthe",
"Stefan Groß",
"Anja Hennemuth",
"Lars Kaderali",
"Nina Krüger",
"Andreas Leha",
"Simon Martin",
"Alexander Meyer",
"Eike Nagel",
"Stefan Orwat",
"Clemens Scherer",
"Moritz Seiffert",
"Jan Moritz Seliger",
"Stefan Simm",
"Tim Friede",
"Tim Seidler",
"Sandy Engelhardt"
] | 2024-07-12T07:34:10Z | 2024-07-12T07:34:10Z |
2403.18241 | NeuSDFusion: A Spatial-Aware Generative Model for 3D Shape Completion,
Reconstruction, and Generation | 3D shape generation aims to produce innovative 3D content adhering to specific conditions and constraints. Existing methods often decompose 3D shapes into a sequence of localized components, treating each element in isolation without considering spatial consistency. As a result, these approaches exhibit limited versatility in 3D data representation and shape generation, hindering their ability to generate highly diverse 3D shapes that comply with the specified constraints. In this paper, we introduce a novel spatial-aware 3D shape generation framework that leverages 2D plane representations for enhanced 3D shape modeling. To ensure spatial coherence and reduce memory usage, we incorporate a hybrid shape representation technique that directly learns a continuous signed distance field representation of the 3D shape using orthogonal 2D planes. Additionally, we meticulously enforce spatial correspondences across distinct planes using a transformer-based autoencoder structure, promoting the preservation of spatial relationships in the generated 3D shapes. This yields an algorithm that consistently outperforms state-of-the-art 3D shape generation methods on various tasks, including unconditional shape generation, multi-modal shape completion, single-view reconstruction, and text-to-shape synthesis. Our project page is available at https://weizheliu.github.io/NeuSDFusion/ . | http://arxiv.org/pdf/2403.18241v2 | [
"Ruikai Cui",
"Weizhe Liu",
"Weixuan Sun",
"Senbo Wang",
"Taizhang Shang",
"Yang Li",
"Xibin Song",
"Han Yan",
"Zhennan Wu",
"Shenzhou Chen",
"Hongdong Li",
"Pan Ji"
] | 2024-07-12T07:30:00Z | 2024-03-27T04:09:34Z |
2407.09061 | Spectral Self-supervised Feature Selection | Choosing a meaningful subset of features from high-dimensional observations in unsupervised settings can greatly enhance the accuracy of downstream analysis, such as clustering or dimensionality reduction, and provide valuable insights into the sources of heterogeneity in a given dataset. In this paper, we propose a self-supervised graph-based approach for unsupervised feature selection. Our method's core involves computing robust pseudo-labels by applying simple processing steps to the graph Laplacian's eigenvectors. The subset of eigenvectors used for computing pseudo-labels is chosen based on a model stability criterion. We then measure the importance of each feature by training a surrogate model to predict the pseudo-labels from the observations. Our approach is shown to be robust to challenging scenarios, such as the presence of outliers and complex substructures. We demonstrate the effectiveness of our method through experiments on real-world datasets, showing its robustness across multiple domains, particularly its effectiveness on biological datasets. | http://arxiv.org/pdf/2407.09061v1 | [
"Daniel Segal",
"Ofir Lindenbaum",
"Ariel Jaffe"
] | 2024-07-12T07:29:08Z | 2024-07-12T07:29:08Z |
2406.14794 | ImageFlowNet: Forecasting Multiscale Trajectories of Disease Progression
with Irregularly-Sampled Longitudinal Medical Images | The forecasting of disease progression from images is a holy grail for clinical decision making. However, this task is complicated by the inherent high dimensionality, temporal sparsity and sampling irregularity in longitudinal image acquisitions. Existing methods often rely on extracting hand-crafted features and performing time-series analysis in this vector space, leading to a loss of rich spatial information within the images. To overcome these challenges, we introduce ImageFlowNet, a novel framework that learns latent-space flow fields that evolve multiscale representations in joint embedding spaces using neural ODEs and SDEs to model disease progression in the image domain. Notably, ImageFlowNet learns multiscale joint representation spaces by combining cohorts of patients together so that information can be transferred between the patient samples. The dynamics then provide plausible trajectories of progression, with the SDE providing alternative trajectories from the same starting point. We provide theoretical insights that support our formulation of ODEs, and motivate our regularizations involving high-level visual features, latent space organization, and trajectory smoothness. We then demonstrate ImageFlowNet's effectiveness through empirical evaluations on three longitudinal medical image datasets depicting progression in retinal geographic atrophy, multiple sclerosis, and glioblastoma. | http://arxiv.org/pdf/2406.14794v3 | [
"Chen Liu",
"Ke Xu",
"Liangbo L. Shen",
"Guillaume Huguet",
"Zilong Wang",
"Alexander Tong",
"Danilo Bzdok",
"Jay Stewart",
"Jay C. Wang",
"Lucian V. Del Priore",
"Smita Krishnaswamy"
] | 2024-07-12T07:28:55Z | 2024-06-20T23:51:32Z |
2407.09055 | Advanced Graph Clustering Methods: A Comprehensive and In-Depth Analysis | Graph clustering, which aims to divide a graph into several homogeneous groups, is a critical area of study with applications that span various fields such as social network analysis, bioinformatics, and image segmentation. This paper explores both traditional and more recent approaches to graph clustering. Firstly, key concepts and definitions in graph theory are introduced. The background section covers essential topics, including graph Laplacians and the integration of Deep Learning in graph analysis. The paper then delves into traditional clustering methods, including Spectral Clustering and the Leiden algorithm. Following this, state-of-the-art clustering techniques that leverage deep learning are examined. A comprehensive comparison of these methods is made through experiments. The paper concludes with a discussion of the practical applications of graph clustering and potential future research directions. | http://arxiv.org/pdf/2407.09055v1 | [
"Timothé Watteau",
"Aubin Bonnefoy",
"Simon Illouz-Laurent",
"Joaquim Jusseau",
"Serge Iovleff"
] | 2024-07-12T07:22:45Z | 2024-07-12T07:22:45Z |
2407.09050 | Refusing Safe Prompts for Multi-modal Large Language Models | Multimodal large language models (MLLMs) have become the cornerstone of today's generative AI ecosystem, sparking intense competition among tech giants and startups. In particular, an MLLM generates a text response given a prompt consisting of an image and a question. While state-of-the-art MLLMs use safety filters and alignment techniques to refuse unsafe prompts, in this work, we introduce MLLM-Refusal, the first method that induces refusals for safe prompts. In particular, our MLLM-Refusal optimizes a nearly-imperceptible refusal perturbation and adds it to an image, causing target MLLMs to likely refuse a safe prompt containing the perturbed image and a safe question. Specifically, we formulate MLLM-Refusal as a constrained optimization problem and propose an algorithm to solve it. Our method offers competitive advantages for MLLM model providers by potentially disrupting user experiences of competing MLLMs, since competing MLLM's users will receive unexpected refusals when they unwittingly use these perturbed images in their prompts. We evaluate MLLM-Refusal on four MLLMs across four datasets, demonstrating its effectiveness in causing competing MLLMs to refuse safe prompts while not affecting non-competing MLLMs. Furthermore, we explore three potential countermeasures -- adding Gaussian noise, DiffPure, and adversarial training. Our results show that they are insufficient: though they can mitigate MLLM-Refusal's effectiveness, they also sacrifice the accuracy and/or efficiency of the competing MLLM. The code is available at https://github.com/Sadcardation/MLLM-Refusal. | http://arxiv.org/pdf/2407.09050v1 | [
"Zedian Shao",
"Hongbin Liu",
"Yuepeng Hu",
"Neil Zhenqiang Gong"
] | 2024-07-12T07:18:05Z | 2024-07-12T07:18:05Z |
2406.12945 | Under the Hood of Tabular Data Generation Models: the Strong Impact of
Hyperparameter Tuning | We investigate the impact of dataset-specific hyperparameter, feature encoding, and architecture tuning on five recent model families for tabular data generation through an extensive benchmark on 16 datasets. This study addresses the practical need for a unified evaluation of models that fully considers hyperparameter optimization. Additionally, we propose a reduced search space for each model that allows for quick optimization, achieving nearly equivalent performance at a significantly lower cost.Our benchmark demonstrates that, for most models, large-scale dataset-specific tuning substantially improves performance compared to the original configurations. Furthermore, we confirm that diffusion-based models generally outperform other models on tabular data. However, this advantage is not significant when the entire tuning and training process is restricted to the same GPU budget for all models. | http://arxiv.org/pdf/2406.12945v2 | [
"G. Charbel N. Kindji",
"Lina Maria Rojas-Barahona",
"Elisa Fromont",
"Tanguy Urvoy"
] | 2024-07-12T07:16:33Z | 2024-06-18T07:27:38Z |
2407.09039 | Overcoming Catastrophic Forgetting in Tabular Data Classification: A
Pseudorehearsal-based approach | Continual learning (CL) poses the important challenge of adapting to evolving data distributions without forgetting previously acquired knowledge while consolidating new knowledge. In this paper, we introduce a new methodology, coined as Tabular-data Rehearsal-based Incremental Lifelong Learning framework (TRIL3), designed to address the phenomenon of catastrophic forgetting in tabular data classification problems. TRIL3 uses the prototype-based incremental generative model XuILVQ to generate synthetic data to preserve old knowledge and the DNDF algorithm, which was modified to run in an incremental way, to learn classification tasks for tabular data, without storing old samples. After different tests to obtain the adequate percentage of synthetic data and to compare TRIL3 with other CL available proposals, we can conclude that the performance of TRIL3 outstands other options in the literature using only 50% of synthetic data. | http://arxiv.org/pdf/2407.09039v1 | [
"Pablo García-Santaclara",
"Bruno Fernández-Castro",
"Rebeca P. Díaz-Redondo"
] | 2024-07-12T07:04:06Z | 2024-07-12T07:04:06Z |
2311.17833 | DiG-IN: Diffusion Guidance for Investigating Networks -- Uncovering
Classifier Differences Neuron Visualisations and Visual Counterfactual
Explanations | While deep learning has led to huge progress in complex image classification tasks like ImageNet, unexpected failure modes, e.g. via spurious features, call into question how reliably these classifiers work in the wild. Furthermore, for safety-critical tasks the black-box nature of their decisions is problematic, and explanations or at least methods which make decisions plausible are needed urgently. In this paper, we address these problems by generating images that optimize a classifier-derived objective using a framework for guided image generation. We analyze the decisions of image classifiers by visual counterfactual explanations (VCEs), detection of systematic mistakes by analyzing images where classifiers maximally disagree, and visualization of neurons and spurious features. In this way, we validate existing observations, e.g. the shape bias of adversarially robust models, as well as novel failure modes, e.g. systematic errors of zero-shot CLIP classifiers. Moreover, our VCEs outperform previous work while being more versatile. | http://arxiv.org/pdf/2311.17833v3 | [
"Maximilian Augustin",
"Yannic Neuhaus",
"Matthias Hein"
] | 2024-07-12T06:53:50Z | 2023-11-29T17:35:29Z |
2407.09032 | DRM Revisited: A Complete Error Analysis | In this work, we address a foundational question in the theoretical analysis of the Deep Ritz Method (DRM) under the over-parameteriztion regime: Given a target precision level, how can one determine the appropriate number of training samples, the key architectural parameters of the neural networks, the step size for the projected gradient descent optimization procedure, and the requisite number of iterations, such that the output of the gradient descent process closely approximates the true solution of the underlying partial differential equation to the specified precision? | http://arxiv.org/pdf/2407.09032v1 | [
"Yuling Jiao",
"Ruoxuan Li",
"Peiying Wu",
"Jerry Zhijian Yang",
"Pingwen Zhang"
] | 2024-07-12T06:48:00Z | 2024-07-12T06:48:00Z |
2407.09026 | HPC: Hierarchical Progressive Coding Framework for Volumetric Video | Volumetric video based on Neural Radiance Field (NeRF) holds vast potential for various 3D applications, but its substantial data volume poses significant challenges for compression and transmission. Current NeRF compression lacks the flexibility to adjust video quality and bitrate within a single model for various network and device capacities. To address these issues, we propose HPC, a novel hierarchical progressive volumetric video coding framework achieving variable bitrate using a single model. Specifically, HPC introduces a hierarchical representation with a multi-resolution residual radiance field to reduce temporal redundancy in long-duration sequences while simultaneously generating various levels of detail. Then, we propose an end-to-end progressive learning approach with a multi-rate-distortion loss function to jointly optimize both hierarchical representation and compression. Our HPC trained only once can realize multiple compression levels, while the current methods need to train multiple fixed-bitrate models for different rate-distortion (RD) tradeoffs. Extensive experiments demonstrate that HPC achieves flexible quality levels with variable bitrate by a single model and exhibits competitive RD performance, even outperforming fixed-bitrate models across various datasets. | http://arxiv.org/pdf/2407.09026v1 | [
"Zihan Zheng",
"Houqiang Zhong",
"Qiang Hu",
"Xiaoyun Zhang",
"Li Song",
"Ya Zhang",
"Yanfeng Wang"
] | 2024-07-12T06:34:24Z | 2024-07-12T06:34:24Z |
2407.09024 | Aligning Diffusion Behaviors with Q-functions for Efficient Continuous
Control | Drawing upon recent advances in language model alignment, we formulate offline Reinforcement Learning as a two-stage optimization problem: First pretraining expressive generative policies on reward-free behavior datasets, then fine-tuning these policies to align with task-specific annotations like Q-values. This strategy allows us to leverage abundant and diverse behavior data to enhance generalization and enable rapid adaptation to downstream tasks using minimal annotations. In particular, we introduce Efficient Diffusion Alignment (EDA) for solving continuous control problems. EDA utilizes diffusion models for behavior modeling. However, unlike previous approaches, we represent diffusion policies as the derivative of a scalar neural network with respect to action inputs. This representation is critical because it enables direct density calculation for diffusion models, making them compatible with existing LLM alignment theories. During policy fine-tuning, we extend preference-based alignment methods like Direct Preference Optimization (DPO) to align diffusion behaviors with continuous Q-functions. Our evaluation on the D4RL benchmark shows that EDA exceeds all baseline methods in overall performance. Notably, EDA maintains about 95% of performance and still outperforms several baselines given only 1% of Q-labelled data during fine-tuning. | http://arxiv.org/pdf/2407.09024v1 | [
"Huayu Chen",
"Kaiwen Zheng",
"Hang Su",
"Jun Zhu"
] | 2024-07-12T06:32:36Z | 2024-07-12T06:32:36Z |
2407.09017 | AI-Driven Guided Response for Security Operation Centers with Microsoft
Copilot for Security | Security operation centers contend with a constant stream of security incidents, ranging from straightforward to highly complex. To address this, we developed Copilot Guided Response (CGR), an industry-scale ML architecture that guides security analysts across three key tasks -- (1) investigation, providing essential historical context by identifying similar incidents; (2) triaging to ascertain the nature of the incident -- whether it is a true positive, false positive, or benign positive; and (3) remediation, recommending tailored containment actions. CGR is integrated into the Microsoft Defender XDR product and deployed worldwide, generating millions of recommendations across thousands of customers. Our extensive evaluation, incorporating internal evaluation, collaboration with security experts, and customer feedback, demonstrates that CGR delivers high-quality recommendations across all three tasks. We provide a comprehensive overview of the CGR architecture, setting a precedent as the first cybersecurity company to openly discuss these capabilities in such depth. Additionally, we GUIDE, the largest public collection of real-world security incidents, spanning 13M evidences across 1M annotated incidents. By enabling researchers and practitioners to conduct research on real-world data, GUIDE advances the state of cybersecurity and supports the development of next-generation machine learning systems. | http://arxiv.org/pdf/2407.09017v1 | [
"Scott Freitas",
"Jovan Kalajdjieski",
"Amir Gharib",
"Rob McCann"
] | 2024-07-12T06:10:01Z | 2024-07-12T06:10:01Z |
2406.06287 | VS-PINN: A fast and efficient training of physics-informed neural
networks using variable-scaling methods for solving PDEs with stiff behavior | Physics-informed neural networks (PINNs) have recently emerged as a promising way to compute the solutions of partial differential equations (PDEs) using deep neural networks. However, despite their significant success in various fields, it remains unclear in many aspects how to effectively train PINNs if the solutions of PDEs exhibit stiff behaviors or high frequencies. In this paper, we propose a new method for training PINNs using variable-scaling techniques. This method is simple and it can be applied to a wide range of problems including PDEs with rapidly-varying solutions. Throughout various numerical experiments, we will demonstrate the effectiveness of the proposed method for these problems and confirm that it can significantly improve the training efficiency and performance of PINNs. Furthermore, based on the analysis of the neural tangent kernel (NTK), we will provide theoretical evidence for this phenomenon and show that our methods can indeed improve the performance of PINNs. | http://arxiv.org/pdf/2406.06287v2 | [
"Seungchan Ko",
"Sang Hyeon Park"
] | 2024-07-12T06:08:09Z | 2024-06-10T14:11:15Z |
2401.17036 | Data organization limits the predictability of binary classification | The structure of data organization is widely recognized as having a substantial influence on the efficacy of machine learning algorithms, particularly in binary classification tasks. Our research provides a theoretical framework suggesting that the maximum potential of binary classifiers on a given dataset is primarily constrained by the inherent qualities of the data. Through both theoretical reasoning and empirical examination, we employed standard objective functions, evaluative metrics, and binary classifiers to arrive at two principal conclusions. Firstly, we show that the theoretical upper bound of binary classification performance on actual datasets can be theoretically attained. This upper boundary represents a calculable equilibrium between the learning loss and the metric of evaluation. Secondly, we have computed the precise upper bounds for three commonly used evaluation metrics, uncovering a fundamental uniformity with our overarching thesis: the upper bound is intricately linked to the dataset's characteristics, independent of the classifier in use. Additionally, our subsequent analysis uncovers a detailed relationship between the upper limit of performance and the level of class overlap within the binary classification data. This relationship is instrumental for pinpointing the most effective feature subsets for use in feature engineering. | http://arxiv.org/pdf/2401.17036v2 | [
"Fei Jing",
"Zi-Ke Zhang",
"Yi-Cheng Zhang",
"Qingpeng Zhang"
] | 2024-07-12T06:04:47Z | 2024-01-30T14:16:02Z |
2407.09013 | Procedural Content Generation via Generative Artificial Intelligence | The attempt to utilize machine learning in PCG has been made in the past. In this survey paper, we investigate how generative artificial intelligence (AI), which saw a significant increase in interest in the mid-2010s, is being used for PCG. We review applications of generative AI for the creation of various types of content, including terrains, items, and even storylines. While generative AI is effective for PCG, one significant issues it faces is that building high-performance generative AI requires vast amounts of training data. Because content generally highly customized, domain-specific training data is scarce, and straightforward approaches to generative AI models may not work well. For PCG research to advance further, issues related to limited training data must be overcome. Thus, we also give special consideration to research that addresses the challenges posed by limited training data. | http://arxiv.org/pdf/2407.09013v1 | [
"Xinyu Mao",
"Wanli Yu",
"Kazunori D Yamada",
"Michael R. Zielewski"
] | 2024-07-12T06:03:38Z | 2024-07-12T06:03:38Z |
2407.09011 | One Stone, Four Birds: A Comprehensive Solution for QA System Using
Supervised Contrastive Learning | This paper presents a novel and comprehensive solution to enhance both the robustness and efficiency of question answering (QA) systems through supervised contrastive learning (SCL). Training a high-performance QA system has become straightforward with pre-trained language models, requiring only a small amount of data and simple fine-tuning. However, despite recent advances, existing QA systems still exhibit significant deficiencies in functionality and training efficiency. We address the functionality issue by defining four key tasks: user input intent classification, out-of-domain input detection, new intent discovery, and continual learning. We then leverage a unified SCL-based representation learning method to efficiently build an intra-class compact and inter-class scattered feature space, facilitating both known intent classification and unknown intent detection and discovery. Consequently, with minimal additional tuning on downstream tasks, our approach significantly improves model efficiency and achieves new state-of-the-art performance across all tasks. | http://arxiv.org/pdf/2407.09011v1 | [
"Bo Wang",
"Tsunenori Mine"
] | 2024-07-12T06:01:51Z | 2024-07-12T06:01:51Z |
2309.11798 | A Comprehensive Review of Community Detection in Graphs | The study of complex networks has significantly advanced our understanding of community structures which serves as a crucial feature of real-world graphs. Detecting communities in graphs is a challenging problem with applications in sociology, biology, and computer science. Despite the efforts of an interdisciplinary community of scientists, a satisfactory solution to this problem has not yet been achieved. This review article delves into the topic of community detection in graphs, which serves as a thorough exposition of various community detection methods from perspectives of modularity-based method, spectral clustering, probabilistic modelling, and deep learning. Along with the methods, a new community detection method designed by us is also presented. Additionally, the performance of these methods on the datasets with and without ground truth is compared. In conclusion, this comprehensive review provides a deep understanding of community detection in graphs. | http://arxiv.org/pdf/2309.11798v5 | [
"Jiakang Li",
"Songning Lai",
"Zhihao Shuai",
"Yuan Tan",
"Yifan Jia",
"Mianyang Yu",
"Zichen Song",
"Xiaokang Peng",
"Ziyang Xu",
"Yongxin Ni",
"Haifeng Qiu",
"Jiayu Yang",
"Yutong Liu",
"Yonggang Lu"
] | 2024-07-12T05:55:47Z | 2023-09-21T06:04:06Z |
2402.19449 | Heavy-Tailed Class Imbalance and Why Adam Outperforms Gradient Descent
on Language Models | Adam has been shown to outperform gradient descent on large language models by a larger margin than on other tasks, but it is unclear why. We show that a key factor in this performance gap is the heavy-tailed class imbalance found in language tasks. When trained with gradient descent, the loss of infrequent words decreases more slowly than the loss of frequent ones. This leads to a slow decrease on the average loss as most samples come from infrequent words. On the other hand, Adam and sign-based methods are less sensitive to this problem. To establish that this behavior is caused by class imbalance, we show empirically that it can be reproduced across architectures and data types, on language transformers, vision CNNs, and linear models. On a linear model with cross-entropy loss, we show that class imbalance leads to imbalanced, correlated gradients and Hessians that have been hypothesized to benefit Adam. We also prove that, in continuous time, gradient descent converges slowly on low-frequency classes while sign descent does not. | http://arxiv.org/pdf/2402.19449v2 | [
"Frederik Kunstner",
"Robin Yadav",
"Alan Milligan",
"Mark Schmidt",
"Alberto Bietti"
] | 2024-07-12T05:10:32Z | 2024-02-29T18:47:52Z |
2402.13468 | STENCIL: Submodular Mutual Information Based Weak Supervision for
Cold-Start Active Learning | As supervised fine-tuning of pre-trained models within NLP applications increases in popularity, larger corpora of annotated data are required, especially with increasing parameter counts in large language models. Active learning, which attempts to mine and annotate unlabeled instances to improve model performance maximally fast, is a common choice for reducing the annotation cost; however, most methods typically ignore class imbalance and either assume access to initial annotated data or require multiple rounds of active learning selection before improving rare classes. We present STENCIL, which utilizes a set of text exemplars and the recently proposed submodular mutual information to select a set of weakly labeled rare-class instances that are then strongly labeled by an annotator. We show that STENCIL improves overall accuracy by $10%-18%$ and rare-class F-1 score by $17%-40%$ on multiple text classification datasets over common active learning methods within the class-imbalanced cold-start setting. | http://arxiv.org/pdf/2402.13468v2 | [
"Nathan Beck",
"Adithya Iyer",
"Rishabh Iyer"
] | 2024-07-12T04:44:39Z | 2024-02-21T01:54:58Z |
2407.08987 | Parameter inference from a non-stationary unknown process | Non-stationary systems are found throughout the world, from climate patterns under the influence of variation in carbon dioxide concentration, to brain dynamics driven by ascending neuromodulation. Accordingly, there is a need for methods to analyze non-stationary processes, and yet most time-series analysis methods that are used in practice, on important problems across science and industry, make the simplifying assumption of stationarity. One important problem in the analysis of non-stationary systems is the problem class that we refer to as Parameter Inference from a Non-stationary Unknown Process (PINUP). Given an observed time series, this involves inferring the parameters that drive non-stationarity of the time series, without requiring knowledge or inference of a mathematical model of the underlying system. Here we review and unify a diverse literature of algorithms for PINUP. We formulate the problem, and categorize the various algorithmic contributions. This synthesis will allow researchers to identify gaps in the literature and will enable systematic comparisons of different methods. We also demonstrate that the most common systems that existing methods are tested on - notably the non-stationary Lorenz process and logistic map - are surprisingly easy to perform well on using simple statistical features like windowed mean and variance, undermining the practice of using good performance on these systems as evidence of algorithmic performance. We then identify more challenging problems that many existing methods perform poorly on and which can be used to drive methodological advances in the field. Our results unify disjoint scientific contributions to analyzing non-stationary systems and suggest new directions for progress on the PINUP problem and the broader study of non-stationary phenomena. | http://arxiv.org/pdf/2407.08987v1 | [
"Kieran S. Owens",
"Ben D. Fulcher"
] | 2024-07-12T04:44:29Z | 2024-07-12T04:44:29Z |
2403.12986 | BaCon: Boosting Imbalanced Semi-supervised Learning via Balanced
Feature-Level Contrastive Learning | Semi-supervised Learning (SSL) reduces the need for extensive annotations in deep learning, but the more realistic challenge of imbalanced data distribution in SSL remains largely unexplored. In Class Imbalanced Semi-supervised Learning (CISSL), the bias introduced by unreliable pseudo-labels can be exacerbated by imbalanced data distributions. Most existing methods address this issue at instance-level through reweighting or resampling, but the performance is heavily limited by their reliance on biased backbone representation. Some other methods do perform feature-level adjustments like feature blending but might introduce unfavorable noise. In this paper, we discuss the bonus of a more balanced feature distribution for the CISSL problem, and further propose a Balanced Feature-Level Contrastive Learning method (BaCon). Our method directly regularizes the distribution of instances' representations in a well-designed contrastive manner. Specifically, class-wise feature centers are computed as the positive anchors, while negative anchors are selected by a straightforward yet effective mechanism. A distribution-related temperature adjustment is leveraged to control the class-wise contrastive degrees dynamically. Our method demonstrates its effectiveness through comprehensive experiments on the CIFAR10-LT, CIFAR100-LT, STL10-LT, and SVHN-LT datasets across various settings. For example, BaCon surpasses instance-level method FixMatch-based ABC on CIFAR10-LT with a 1.21% accuracy improvement, and outperforms state-of-the-art feature-level method CoSSL on CIFAR100-LT with a 0.63% accuracy improvement. When encountering more extreme imbalance degree, BaCon also shows better robustness than other methods. | http://arxiv.org/pdf/2403.12986v2 | [
"Qianhan Feng",
"Lujing Xie",
"Shijie Fang",
"Tong Lin"
] | 2024-07-12T04:43:48Z | 2024-03-04T06:43:16Z |
2407.08983 | Towards More Trustworthy and Interpretable LLMs for Code through
Syntax-Grounded Explanations | Trustworthiness and interpretability are inextricably linked concepts for LLMs. The more interpretable an LLM is, the more trustworthy it becomes. However, current techniques for interpreting LLMs when applied to code-related tasks largely focus on accuracy measurements, measures of how models react to change, or individual task performance instead of the fine-grained explanations needed at prediction time for greater interpretability, and hence trust. To improve upon this status quo, this paper introduces ASTrust, an interpretability method for LLMs of code that generates explanations grounded in the relationship between model confidence and syntactic structures of programming languages. ASTrust explains generated code in the context of syntax categories based on Abstract Syntax Trees and aids practitioners in understanding model predictions at both local (individual code snippets) and global (larger datasets of code) levels. By distributing and assigning model confidence scores to well-known syntactic structures that exist within ASTs, our approach moves beyond prior techniques that perform token-level confidence mapping by offering a view of model confidence that directly aligns with programming language concepts with which developers are familiar. To put ASTrust into practice, we developed an automated visualization that illustrates the aggregated model confidence scores superimposed on sequence, heat-map, and graph-based visuals of syntactic structures from ASTs. We examine both the practical benefit that ASTrust can provide through a data science study on 12 popular LLMs on a curated set of GitHub repos and the usefulness of ASTrust through a human study. | http://arxiv.org/pdf/2407.08983v1 | [
"David N. Palacio",
"Daniel Rodriguez-Cardenas",
"Alejandro Velasco",
"Dipin Khati",
"Kevin Moran",
"Denys Poshyvanyk"
] | 2024-07-12T04:38:28Z | 2024-07-12T04:38:28Z |
2407.08978 | Towards Chapter-to-Chapter Context-Aware Literary Translation via Large
Language Models | Discourse phenomena in existing document-level translation datasets are sparse, which has been a fundamental obstacle in the development of context-aware machine translation models. Moreover, most existing document-level corpora and context-aware machine translation methods rely on an unrealistic assumption on sentence-level alignments. To mitigate these issues, we first curate a novel dataset of Chinese-English literature, which consists of 160 books with intricate discourse structures. Then, we propose a more pragmatic and challenging setting for context-aware translation, termed chapter-to-chapter (Ch2Ch) translation, and investigate the performance of commonly-used machine translation models under this setting. Furthermore, we introduce a potential approach of finetuning large language models (LLMs) within the domain of Ch2Ch literary translation, yielding impressive improvements over baselines. Through our comprehensive analysis, we unveil that literary translation under the Ch2Ch setting is challenging in nature, with respect to both model learning methods and translation decoding algorithms. | http://arxiv.org/pdf/2407.08978v1 | [
"Linghao Jin",
"Li An",
"Xuezhe Ma"
] | 2024-07-12T04:18:22Z | 2024-07-12T04:18:22Z |
2407.08976 | Computational-Statistical Trade-off in Kernel Two-Sample Testing with
Random Fourier Features | Recent years have seen a surge in methods for two-sample testing, among which the Maximum Mean Discrepancy (MMD) test has emerged as an effective tool for handling complex and high-dimensional data. Despite its success and widespread adoption, the primary limitation of the MMD test has been its quadratic-time complexity, which poses challenges for large-scale analysis. While various approaches have been proposed to expedite the procedure, it has been unclear whether it is possible to attain the same power guarantee as the MMD test at sub-quadratic time cost. To fill this gap, we revisit the approximated MMD test using random Fourier features, and investigate its computational-statistical trade-off. We start by revealing that the approximated MMD test is pointwise consistent in power only when the number of random features approaches infinity. We then consider the uniform power of the test and study the time-power trade-off under the minimax testing framework. Our result shows that, by carefully choosing the number of random features, it is possible to attain the same minimax separation rates as the MMD test within sub-quadratic time. We demonstrate this point under different distributional assumptions such as densities in a Sobolev ball. Our theoretical findings are corroborated by simulation studies. | http://arxiv.org/pdf/2407.08976v1 | [
"Ikjun Choi",
"Ilmun Kim"
] | 2024-07-12T04:08:01Z | 2024-07-12T04:08:01Z |
2407.08974 | Topology-enhanced machine learning model (Top-ML) for anticancer peptide
prediction | Recently, therapeutic peptides have demonstrated great promise for cancer treatment. To explore powerful anticancer peptides, artificial intelligence (AI)-based approaches have been developed to systematically screen potential candidates. However, the lack of efficient featurization of peptides has become a bottleneck for these machine-learning models. In this paper, we propose a topology-enhanced machine learning model (Top-ML) for anticancer peptide prediction. Our Top-ML employs peptide topological features derived from its sequence "connection" information characterized by vector and spectral descriptors. Our Top-ML model has been validated on two widely used AntiCP 2.0 benchmark datasets and has achieved state-of-the-art performance. Our results highlight the potential of leveraging novel topology-based featurization to accelerate the identification of anticancer peptides. | http://arxiv.org/pdf/2407.08974v1 | [
"Joshua Zhi En Tan",
"JunJie Wee",
"Xue Gong",
"Kelin Xia"
] | 2024-07-12T04:04:54Z | 2024-07-12T04:04:54Z |
2407.08973 | Integrating White and Black Box Techniques for Interpretable Machine
Learning | In machine learning algorithm design, there exists a trade-off between the interpretability and performance of the algorithm. In general, algorithms which are simpler and easier for humans to comprehend tend to show worse performance than more complex, less transparent algorithms. For example, a random forest classifier is likely to be more accurate than a simple decision tree, but at the expense of interpretability. In this paper, we present an ensemble classifier design which classifies easier inputs using a highly-interpretable classifier (i.e., white box model), and more difficult inputs using a more powerful, but less interpretable classifier (i.e., black box model). | http://arxiv.org/pdf/2407.08973v1 | [
"Eric M. Vernon",
"Naoki Masuyama",
"Yusuke Nojima"
] | 2024-07-12T03:58:04Z | 2024-07-12T03:58:04Z |
2306.04647 | Compressed Sensing: A Discrete Optimization Approach | We study the Compressed Sensing (CS) problem, which is the problem of finding the most sparse vector that satisfies a set of linear measurements up to some numerical tolerance. We introduce an $ell_2$ regularized formulation of CS which we reformulate as a mixed integer second order cone program. We derive a second order cone relaxation of this problem and show that under mild conditions on the regularization parameter, the resulting relaxation is equivalent to the well studied basis pursuit denoising problem. We present a semidefinite relaxation that strengthens the second order cone relaxation and develop a custom branch-and-bound algorithm that leverages our second order cone relaxation to solve small-scale instances of CS to certifiable optimality. When compared against solutions produced by three state of the art benchmark methods on synthetic data, our numerical results show that our approach produces solutions that are on average $6.22%$ more sparse. When compared only against the experiment-wise best performing benchmark method on synthetic data, our approach produces solutions that are on average $3.10%$ more sparse. On real world ECG data, for a given $ell_2$ reconstruction error our approach produces solutions that are on average $9.95%$ more sparse than benchmark methods ($3.88%$ more sparse if only compared against the best performing benchmark), while for a given sparsity level our approach produces solutions that have on average $10.77%$ lower reconstruction error than benchmark methods ($1.42%$ lower error if only compared against the best performing benchmark). When used as a component of a multi-label classification algorithm, our approach achieves greater classification accuracy than benchmark compressed sensing methods. This improved accuracy comes at the cost of an increase in computation time by several orders of magnitude. | http://arxiv.org/pdf/2306.04647v3 | [
"Dimitris Bertsimas",
"Nicholas A. G. Johnson"
] | 2024-07-12T03:46:02Z | 2023-06-05T01:29:24Z |
2407.08970 | Soft Prompts Go Hard: Steering Visual Language Models with Hidden
Meta-Instructions | We introduce a new type of indirect injection vulnerabilities in language models that operate on images: hidden "meta-instructions" that influence how the model interprets the image and steer the model's outputs to express an adversary-chosen style, sentiment, or point of view. We explain how to create meta-instructions by generating images that act as soft prompts. Unlike jailbreaking attacks and adversarial examples, the outputs resulting from these images are plausible and based on the visual content of the image, yet follow the adversary's (meta-)instructions. We describe the risks of these attacks, including misinformation and spin, evaluate their efficacy for multiple visual language models and adversarial meta-objectives, and demonstrate how they can "unlock" the capabilities of the underlying language models that are unavailable via explicit text instructions. Finally, we discuss defenses against these attacks. | http://arxiv.org/pdf/2407.08970v1 | [
"Tingwei Zhang",
"Collin Zhang",
"John X. Morris",
"Eugene Bagdasaryan",
"Vitaly Shmatikov"
] | 2024-07-12T03:40:13Z | 2024-07-12T03:40:13Z |
2407.02813 | Data Overfitting for On-Device Super-Resolution with Dynamic Algorithm
and Compiler Co-Design | Deep neural networks (DNNs) are frequently employed in a variety of computer vision applications. Nowadays, an emerging trend in the current video distribution system is to take advantage of DNN's overfitting properties to perform video resolution upscaling. By splitting videos into chunks and applying a super-resolution (SR) model to overfit each chunk, this scheme of SR models plus video chunks is able to replace traditional video transmission to enhance video quality and transmission efficiency. However, many models and chunks are needed to guarantee high performance, which leads to tremendous overhead on model switching and memory footprints at the user end. To resolve such problems, we propose a Dynamic Deep neural network assisted by a Content-Aware data processing pipeline to reduce the model number down to one (Dy-DCA), which helps promote performance while conserving computational resources. Additionally, to achieve real acceleration on the user end, we designed a framework that optimizes dynamic features (e.g., dynamic shapes, sizes, and control flow) in Dy-DCA to enable a series of compilation optimizations, including fused code generation, static execution planning, etc. By employing such techniques, our method achieves better PSNR and real-time performance (33 FPS) on an off-the-shelf mobile phone. Meanwhile, assisted by our compilation optimization, we achieve a 1.7$times$ speedup while saving up to 1.61$times$ memory consumption. Code available in https://github.com/coulsonlee/Dy-DCA-ECCV2024. | http://arxiv.org/pdf/2407.02813v2 | [
"Gen Li",
"Zhihao Shu",
"Jie Ji",
"Minghai Qin",
"Fatemeh Afghah",
"Wei Niu",
"Xiaolong Ma"
] | 2024-07-12T03:39:05Z | 2024-07-03T05:17:26Z |
2407.08966 | LAPT: Label-driven Automated Prompt Tuning for OOD Detection with
Vision-Language Models | Out-of-distribution (OOD) detection is crucial for model reliability, as it identifies samples from unknown classes and reduces errors due to unexpected inputs. Vision-Language Models (VLMs) such as CLIP are emerging as powerful tools for OOD detection by integrating multi-modal information. However, the practical application of such systems is challenged by manual prompt engineering, which demands domain expertise and is sensitive to linguistic nuances. In this paper, we introduce Label-driven Automated Prompt Tuning (LAPT), a novel approach to OOD detection that reduces the need for manual prompt engineering. We develop distribution-aware prompts with in-distribution (ID) class names and negative labels mined automatically. Training samples linked to these class labels are collected autonomously via image synthesis and retrieval methods, allowing for prompt learning without manual effort. We utilize a simple cross-entropy loss for prompt optimization, with cross-modal and cross-distribution mixing strategies to reduce image noise and explore the intermediate space between distributions, respectively. The LAPT framework operates autonomously, requiring only ID class names as input and eliminating the need for manual intervention. With extensive experiments, LAPT consistently outperforms manually crafted prompts, setting a new standard for OOD detection. Moreover, LAPT not only enhances the distinction between ID and OOD samples, but also improves the ID classification accuracy and strengthens the generalization robustness to covariate shifts, resulting in outstanding performance in challenging full-spectrum OOD detection tasks. Codes are available at url{https://github.com/YBZh/LAPT}. | http://arxiv.org/pdf/2407.08966v1 | [
"Yabin Zhang",
"Wenjie Zhu",
"Chenhang He",
"Lei Zhang"
] | 2024-07-12T03:30:53Z | 2024-07-12T03:30:53Z |
2407.08965 | Lite-SAM Is Actually What You Need for Segment Everything | This paper introduces Lite-SAM, an efficient end-to-end solution for the SegEvery task designed to reduce computational costs and redundancy. Lite-SAM is composed of four main components: a streamlined CNN-Transformer hybrid encoder (LiteViT), an automated prompt proposal network (AutoPPN), a traditional prompt encoder, and a mask decoder. All these components are integrated within the SAM framework. Our LiteViT, a high-performance lightweight backbone network, has only 1.16M parameters, which is a 23% reduction compared to the lightest existing backbone network Shufflenet. We also introduce AutoPPN, an innovative end-to-end method for prompt boxes and points generation. This is an improvement over traditional grid search sampling methods, and its unique design allows for easy integration into any SAM series algorithm, extending its usability. we have thoroughly benchmarked Lite-SAM across a plethora of both public and private datasets. The evaluation encompassed a broad spectrum of universal metrics, including the number of parameters, SegEvery execution time, and accuracy. The findings reveal that Lite-SAM, operating with a lean 4.2M parameters, significantly outpaces its counterparts, demonstrating performance improvements of 43x, 31x, 20x, 21x, and 1.6x over SAM, MobileSAM, Edge-SAM, EfficientViT-SAM, and MobileSAM-v2 respectively, all the while maintaining competitive accuracy. This underscores Lite-SAM's prowess in achieving an optimal equilibrium between performance and precision, thereby setting a new state-of-the-art(SOTA) benchmark in the domain. | http://arxiv.org/pdf/2407.08965v1 | [
"Jianhai Fu",
"Yuanjie Yu",
"Ningchuan Li",
"Yi Zhang",
"Qichao Chen",
"Jianping Xiong",
"Jun Yin",
"Zhiyu Xiang"
] | 2024-07-12T03:28:46Z | 2024-07-12T03:28:46Z |
2407.08964 | Communication-Aware Reinforcement Learning for Cooperative Adaptive
Cruise Control | Cooperative Adaptive Cruise Control (CACC) plays a pivotal role in enhancing traffic efficiency and safety in Connected and Autonomous Vehicles (CAVs). Reinforcement Learning (RL) has proven effective in optimizing complex decision-making processes in CACC, leading to improved system performance and adaptability. Among RL approaches, Multi-Agent Reinforcement Learning (MARL) has shown remarkable potential by enabling coordinated actions among multiple CAVs through Centralized Training with Decentralized Execution (CTDE). However, MARL often faces scalability issues, particularly when CACC vehicles suddenly join or leave the platoon, resulting in performance degradation. To address these challenges, we propose Communication-Aware Reinforcement Learning (CA-RL). CA-RL includes a communication-aware module that extracts and compresses vehicle communication information through forward and backward information transmission modules. This enables efficient cyclic information propagation within the CACC traffic flow, ensuring policy consistency and mitigating the scalability problems of MARL in CACC. Experimental results demonstrate that CA-RL significantly outperforms baseline methods in various traffic scenarios, achieving superior scalability, robustness, and overall system performance while maintaining reliable performance despite changes in the number of participating vehicles. | http://arxiv.org/pdf/2407.08964v1 | [
"Sicong Jiang",
"Seongjin Choi",
"Lijun Sun"
] | 2024-07-12T03:28:24Z | 2024-07-12T03:28:24Z |
2407.08953 | Attribution Methods in Asset Pricing: Do They Account for Risk? | Over the past few decades, machine learning models have been extremely successful. As a result of axiomatic attribution methods, feature contributions have been explained more clearly and rigorously. There are, however, few studies that have examined domain knowledge in conjunction with the axioms. In this study, we examine asset pricing in finance, a field closely related to risk management. Consequently, when applying machine learning models, we must ensure that the attribution methods reflect the underlying risks accurately. In this work, we present and study several axioms derived from asset pricing domain knowledge. It is shown that while Shapley value and Integrated Gradients preserve most axioms, neither can satisfy all axioms. Using extensive analytical and empirical examples, we demonstrate how attribution methods can reflect risks and when they should not be used. | http://arxiv.org/pdf/2407.08953v1 | [
"Dangxing Chen",
"Yuan Gao"
] | 2024-07-12T03:16:54Z | 2024-07-12T03:16:54Z |
2407.08947 | Constructing Concept-based Models to Mitigate Spurious Correlations with
Minimal Human Effort | Enhancing model interpretability can address spurious correlations by revealing how models draw their predictions. Concept Bottleneck Models (CBMs) can provide a principled way of disclosing and guiding model behaviors through human-understandable concepts, albeit at a high cost of human efforts in data annotation. In this paper, we leverage a synergy of multiple foundation models to construct CBMs with nearly no human effort. We discover undesirable biases in CBMs built on pre-trained models and propose a novel framework designed to exploit pre-trained models while being immune to these biases, thereby reducing vulnerability to spurious correlations. Specifically, our method offers a seamless pipeline that adopts foundation models for assessing potential spurious correlations in datasets, annotating concepts for images, and refining the annotations for improved robustness. We evaluate the proposed method on multiple datasets, and the results demonstrate its effectiveness in reducing model reliance on spurious correlations while preserving its interpretability. | http://arxiv.org/pdf/2407.08947v1 | [
"Jeeyung Kim",
"Ze Wang",
"Qiang Qiu"
] | 2024-07-12T03:07:28Z | 2024-07-12T03:07:28Z |
2407.08946 | Your Diffusion Model is Secretly a Noise Classifier and Benefits from
Contrastive Training | Diffusion models learn to denoise data and the trained denoiser is then used to generate new samples from the data distribution. In this paper, we revisit the diffusion sampling process and identify a fundamental cause of sample quality degradation: the denoiser is poorly estimated in regions that are far Outside Of the training Distribution (OOD), and the sampling process inevitably evaluates in these OOD regions. This can become problematic for all sampling methods, especially when we move to parallel sampling which requires us to initialize and update the entire sample trajectory of dynamics in parallel, leading to many OOD evaluations. To address this problem, we introduce a new self-supervised training objective that differentiates the levels of noise added to a sample, leading to improved OOD denoising performance. The approach is based on our observation that diffusion models implicitly define a log-likelihood ratio that distinguishes distributions with different amounts of noise, and this expression depends on denoiser performance outside the standard training distribution. We show by diverse experiments that the proposed contrastive diffusion training is effective for both sequential and parallel settings, and it improves the performance and speed of parallel samplers significantly. | http://arxiv.org/pdf/2407.08946v1 | [
"Yunshu Wu",
"Yingtao Luo",
"Xianghao Kong",
"Evangelos E. Papalexakis",
"Greg Ver Steeg"
] | 2024-07-12T03:03:50Z | 2024-07-12T03:03:50Z |
2407.08418 | PredBench: Benchmarking Spatio-Temporal Prediction across Diverse
Disciplines | In this paper, we introduce PredBench, a benchmark tailored for the holistic evaluation of spatio-temporal prediction networks. Despite significant progress in this field, there remains a lack of a standardized framework for a detailed and comparative analysis of various prediction network architectures. PredBench addresses this gap by conducting large-scale experiments, upholding standardized and appropriate experimental settings, and implementing multi-dimensional evaluations. This benchmark integrates 12 widely adopted methods with 15 diverse datasets across multiple application domains, offering extensive evaluation of contemporary spatio-temporal prediction networks. Through meticulous calibration of prediction settings across various applications, PredBench ensures evaluations relevant to their intended use and enables fair comparisons. Moreover, its multi-dimensional evaluation framework broadens the analysis with a comprehensive set of metrics, providing deep insights into the capabilities of models. The findings from our research offer strategic directions for future developments in the field. Our codebase is available at https://github.com/OpenEarthLab/PredBench. | http://arxiv.org/pdf/2407.08418v2 | [
"ZiDong Wang",
"Zeyu Lu",
"Di Huang",
"Tong He",
"Xihui Liu",
"Wanli Ouyang",
"Lei Bai"
] | 2024-07-12T02:55:16Z | 2024-07-11T11:51:36Z |
2407.08934 | Compositional Structures in Neural Embedding and Interaction
Decompositions | We describe a basic correspondence between linear algebraic structures within vector embeddings in artificial neural networks and conditional independence constraints on the probability distributions modeled by these networks. Our framework aims to shed light on the emergence of structural patterns in data representations, a phenomenon widely acknowledged but arguably still lacking a solid formal grounding. Specifically, we introduce a characterization of compositional structures in terms of "interaction decompositions," and we establish necessary and sufficient conditions for the presence of such structures within the representations of a model. | http://arxiv.org/pdf/2407.08934v1 | [
"Matthew Trager",
"Alessandro Achille",
"Pramuditha Perera",
"Luca Zancato",
"Stefano Soatto"
] | 2024-07-12T02:39:50Z | 2024-07-12T02:39:50Z |
2407.08933 | Machine Learning in High Volume Media Manufacturing | Errors or failures in a high-volume manufacturing environment can have significant impact that can result in both the loss of time and money. Identifying such failures early has been a top priority for manufacturing industries and various rule-based algorithms have been developed over the years. However, catching these failures is time consuming and such algorithms cannot adapt well to changes in designs, and sometimes variations in everyday behavior. More importantly, the number of units to monitor in a high-volume manufacturing environment is too big for manual monitoring or for a simple program. Here we develop a novel program that combines both rule-based decisions and machine learning models that can not only learn and adapt to such day-to-day variations or long-term design changes, but also can be applied at scale to the high number of manufacturing units in use today. Using the current state-of-the-art technologies, we then deploy this program at-scale to handle the needs of ever-increasing demand from the manufacturing environment. | http://arxiv.org/pdf/2407.08933v1 | [
"Siddarth Reddy Karuka",
"Abhinav Sunderrajan",
"Zheng Zheng",
"Yong Woon Tiean",
"Ganesh Nagappan",
"Allan Luk"
] | 2024-07-12T02:34:54Z | 2024-07-12T02:34:54Z |
2405.10757 | Rethinking Graph Backdoor Attacks: A Distribution-Preserving Perspective | Graph Neural Networks (GNNs) have shown remarkable performance in various tasks. However, recent works reveal that GNNs are vulnerable to backdoor attacks. Generally, backdoor attack poisons the graph by attaching backdoor triggers and the target class label to a set of nodes in the training graph. A GNN trained on the poisoned graph will then be misled to predict test nodes attached with trigger to the target class. Despite their effectiveness, our empirical analysis shows that triggers generated by existing methods tend to be out-of-distribution (OOD), which significantly differ from the clean data. Hence, these injected triggers can be easily detected and pruned with widely used outlier detection methods in real-world applications. Therefore, in this paper, we study a novel problem of unnoticeable graph backdoor attacks with in-distribution (ID) triggers. To generate ID triggers, we introduce an OOD detector in conjunction with an adversarial learning strategy to generate the attributes of the triggers within distribution. To ensure a high attack success rate with ID triggers, we introduce novel modules designed to enhance trigger memorization by the victim model trained on poisoned graph. Extensive experiments on real-world datasets demonstrate the effectiveness of the proposed method in generating in distribution triggers that can by-pass various defense strategies while maintaining a high attack success rate. | http://arxiv.org/abs/2405.10757v3 | [
"Zhiwei Zhang",
"Minhua Lin",
"Enyan Dai",
"Suhang Wang"
] | 2024-07-12T02:21:54Z | 2024-05-17T13:09:39Z |
2407.08922 | Leveraging large language models for nano synthesis mechanism
explanation: solid foundations or mere conjectures? | With the rapid development of artificial intelligence (AI), large language models (LLMs) such as GPT-4 have garnered significant attention in the scientific community, demonstrating great potential in advancing scientific discovery. This progress raises a critical question: are these LLMs well-aligned with real-world physicochemical principles? Current evaluation strategies largely emphasize fact-based knowledge, such as material property prediction or name recognition, but they often lack an understanding of fundamental physicochemical mechanisms that require logical reasoning. To bridge this gap, our study developed a benchmark consisting of 775 multiple-choice questions focusing on the mechanisms of gold nanoparticle synthesis. By reflecting on existing evaluation metrics, we question whether a direct true-or-false assessment merely suggests conjecture. Hence, we propose a novel evaluation metric, the confidence-based score (c-score), which probes the output logits to derive the precise probability for the correct answer. Based on extensive experiments, our results show that in the context of gold nanoparticle synthesis, LLMs understand the underlying physicochemical mechanisms rather than relying on conjecture. This study underscores the potential of LLMs to grasp intrinsic scientific mechanisms and sets the stage for developing more reliable and effective AI tools across various scientific domains. | http://arxiv.org/pdf/2407.08922v1 | [
"Yingming Pu",
"Liping Huang",
"Tao Lin",
"Hongyu Chen"
] | 2024-07-12T02:05:59Z | 2024-07-12T02:05:59Z |
2401.11410 | Agricultural Recommendation System based on Deep Learning: A
Multivariate Weather Forecasting Approach | Agriculture plays a fundamental role in driving economic growth and ensuring food security for populations around the world. Although labor-intensive agriculture has led to steady increases in food grain production in many developing countries, it is frequently challenged by adverse weather conditions, including heavy rainfall, low temperatures, and drought. These factors substantially hinder food production, posing significant risks to global food security. In order to have a profitable, sustainable, and farmer-friendly agricultural practice, this paper proposes a context-based crop recommendation system powered by a weather forecast model. For implementation purposes, we have considered the whole territory of Bangladesh. With extensive evaluation, the multivariate Stacked Bi-LSTM (three Bi-LSTM layers with a time Distributed layer) Network is employed as the weather forecasting model. The proposed weather model can forecast Rainfall, Temperature, Humidity, and Sunshine for any given location in Bangladesh with an average R-Squared value of 0.9824, and the model outperforms other state-of-the-art LSTM models. These predictions guide our system in generating viable farming decisions. Additionally, our full-fledged system is capable of alerting the farmers about extreme weather conditions so that preventive measures can be undertaken to protect the crops. Finally, the system is also adept at making knowledge-based crop suggestions for flood and drought-prone regions. | http://arxiv.org/pdf/2401.11410v3 | [
"Md Zubair",
"Md. Shahidul Salim",
"Mehrab Mustafy Rahman",
"Mohammad Jahid Ibna Basher",
"Shahin Imran",
"Iqbal H. Sarker"
] | 2024-07-12T02:02:45Z | 2024-01-21T06:33:45Z |
2306.08388 | Skill-Critic: Refining Learned Skills for Hierarchical Reinforcement
Learning | Hierarchical reinforcement learning (RL) can accelerate long-horizon decision-making by temporally abstracting a policy into multiple levels. Promising results in sparse reward environments have been seen with skills, i.e. sequences of primitive actions. Typically, a skill latent space and policy are discovered from offline data. However, the resulting low-level policy can be unreliable due to low-coverage demonstrations or distribution shifts. As a solution, we propose the Skill-Critic algorithm to fine-tune the low-level policy in conjunction with high-level skill selection. Our Skill-Critic algorithm optimizes both the low-level and high-level policies; these policies are initialized and regularized by the latent space learned from offline demonstrations to guide the parallel policy optimization. We validate Skill-Critic in multiple sparse-reward RL environments, including a new sparse-reward autonomous racing task in Gran Turismo Sport. The experiments show that Skill-Critic's low-level policy fine-tuning and demonstration-guided regularization are essential for good performance. Code and videos are available at our website: https://sites.google.com/view/skill-critic. | http://arxiv.org/pdf/2306.08388v3 | [
"Ce Hao",
"Catherine Weaver",
"Chen Tang",
"Kenta Kawamoto",
"Masayoshi Tomizuka",
"Wei Zhan"
] | 2024-07-12T01:59:00Z | 2023-06-14T09:24:32Z |
2407.09578 | Unsupervised Anomaly Detection Using Diffusion Trend Analysis | Conventional anomaly detection techniques based on reconstruction via denoising diffusion model are widely used due to their ability to identify anomaly locations and shapes with high performance. However, there is a limitation in determining appropriate noise parameters that can degrade anomalies while preserving normal characteristics. Also, due to the volatility of the diffusion model, normal regions can fluctuate considerably during reconstruction, resulting in false detection. In this paper, we propose a method to detect anomalies by analysis of reconstruction trend depending on the degree of degradation, effectively solving the both problems of existing methods. The proposed method is validated on an open dataset for industrial anomaly detection, improving the performance of existing methods on a number of evaluation criteria. With the ease of combination with existing anomaly detection methods, it provides a tradeoff between computational cost and performance, allowing it high application potential in manufacturing industry. | http://arxiv.org/pdf/2407.09578v1 | [
"Eunwoo Kim",
"Un Yang",
"Cheol Lae Roh",
"Stefano Ermon"
] | 2024-07-12T01:50:07Z | 2024-07-12T01:50:07Z |
2407.06886 | Aligning Cyber Space with Physical World: A Comprehensive Survey on
Embodied AI | Embodied Artificial Intelligence (Embodied AI) is crucial for achieving Artificial General Intelligence (AGI) and serves as a foundation for various applications that bridge cyberspace and the physical world. Recently, the emergence of Multi-modal Large Models (MLMs) and World Models (WMs) have attracted significant attention due to their remarkable perception, interaction, and reasoning capabilities, making them a promising architecture for the brain of embodied agents. However, there is no comprehensive survey for Embodied AI in the era of MLMs. In this survey, we give a comprehensive exploration of the latest advancements in Embodied AI. Our analysis firstly navigates through the forefront of representative works of embodied robots and simulators, to fully understand the research focuses and their limitations. Then, we analyze four main research targets: 1) embodied perception, 2) embodied interaction, 3) embodied agent, and 4) sim-to-real adaptation, covering the state-of-the-art methods, essential paradigms, and comprehensive datasets. Additionally, we explore the complexities of MLMs in virtual and real embodied agents, highlighting their significance in facilitating interactions in dynamic digital and physical environments. Finally, we summarize the challenges and limitations of embodied AI and discuss their potential future directions. We hope this survey will serve as a foundational reference for the research community and inspire continued innovation. The associated project can be found at https://github.com/HCPLab-SYSU/Embodied_AI_Paper_List. | http://arxiv.org/pdf/2407.06886v2 | [
"Yang Liu",
"Weixing Chen",
"Yongjie Bai",
"Jingzhou Luo",
"Xinshuai Song",
"Kaixuan Jiang",
"Zhida Li",
"Ganlong Zhao",
"Junyi Lin",
"Guanbin Li",
"Wen Gao",
"Liang Lin"
] | 2024-07-12T01:48:00Z | 2024-07-09T14:14:47Z |
2406.18145 | Beyond Statistical Estimation: Differentially Private Individual
Computation via Shuffling | In data-driven applications, preserving user privacy while enabling valuable computations remains a critical challenge. Technologies like Differential Privacy (DP) have been pivotal in addressing these concerns. The shuffle model of DP requires no trusted curators and can achieve high utility by leveraging the privacy amplification effect yielded from shuffling. These benefits have led to significant interest in the shuffle model. However, the computation tasks in the shuffle model are limited to statistical estimation, making the shuffle model inapplicable to real-world scenarios in which each user requires a personalized output. This paper introduces a novel paradigm termed Private Individual Computation (PIC), expanding the shuffle model to support a broader range of permutation-equivariant computations. PIC enables personalized outputs while preserving privacy, and enjoys privacy amplification through shuffling. We propose a concrete protocol that realizes PIC. By using one-time public keys, our protocol enables users to receive their outputs without compromising anonymity, which is essential for privacy amplification. Additionally, we present an optimal randomizer, the Minkowski Response, designed for the PIC model to enhance utility. We formally prove the security and privacy properties of the PIC protocol. Theoretical analysis and empirical evaluations demonstrate PIC's capability in handling non-statistical computation tasks, and the efficacy of PIC and the Minkowski randomizer in achieving superior utility compared to existing solutions. | http://arxiv.org/pdf/2406.18145v2 | [
"Shaowei Wang",
"Changyu Dong",
"Xiangfu Song",
"Jin Li",
"Zhili Zhou",
"Di Wang",
"Han Wu"
] | 2024-07-12T01:36:06Z | 2024-06-26T07:53:48Z |
2407.08916 | Transforming Movie Recommendations with Advanced Machine Learning: A
Study of NMF, SVD,and K-Means Clustering | This study develops a robust movie recommendation system using various machine learning techniques, including Non- Negative Matrix Factorization (NMF), Truncated Singular Value Decomposition (SVD), and K-Means clustering. The primary objective is to enhance user experience by providing personalized movie recommendations. The research encompasses data preprocessing, model training, and evaluation, highlighting the efficacy of the employed methods. Results indicate that the proposed system achieves high accuracy and relevance in recommendations, making significant contributions to the field of recommendations systems. | http://arxiv.org/pdf/2407.08916v1 | [
"Yubing Yan",
"Camille Moreau",
"Zhuoyue Wang",
"Wenhan Fan",
"Chengqian Fu"
] | 2024-07-12T01:26:33Z | 2024-07-12T01:26:33Z |
2407.08910 | PAIL: Performance based Adversarial Imitation Learning Engine for Carbon
Neutral Optimization | Achieving carbon neutrality within industrial operations has become increasingly imperative for sustainable development. It is both a significant challenge and a key opportunity for operational optimization in industry 4.0. In recent years, Deep Reinforcement Learning (DRL) based methods offer promising enhancements for sequential optimization processes and can be used for reducing carbon emissions. However, existing DRL methods need a pre-defined reward function to assess the impact of each action on the final sustainable development goals (SDG). In many real applications, such a reward function cannot be given in advance. To address the problem, this study proposes a Performance based Adversarial Imitation Learning (PAIL) engine. It is a novel method to acquire optimal operational policies for carbon neutrality without any pre-defined action rewards. Specifically, PAIL employs a Transformer-based policy generator to encode historical information and predict following actions within a multi-dimensional space. The entire action sequence will be iteratively updated by an environmental simulator. Then PAIL uses a discriminator to minimize the discrepancy between generated sequences and real-world samples of high SDG. In parallel, a Q-learning framework based performance estimator is designed to estimate the impact of each action on SDG. Based on these estimations, PAIL refines generated policies with the rewards from both discriminator and performance estimator. PAIL is evaluated on multiple real-world application cases and datasets. The experiment results demonstrate the effectiveness of PAIL comparing to other state-of-the-art baselines. In addition, PAIL offers meaningful interpretability for the optimization in carbon neutrality. | http://arxiv.org/abs/2407.08910v1 | [
"Yuyang Ye",
"Lu-An Tang",
"Haoyu Wang",
"Runlong Yu",
"Wenchao Yu",
"Erhu He",
"Haifeng Chen",
"Hui Xiong"
] | 2024-07-12T01:06:01Z | 2024-07-12T01:06:01Z |
2407.09577 | Flash normalization: fast RMSNorm for LLMs | RMSNorm is used by many LLMs such as Llama, Mistral, and OpenELM. This paper details FlashNorm, which is an exact but faster implementation of RMSNorm followed by linear layers. See https://huggingface.co./open-machine/FlashNorm for code and more transformer tricks. | http://arxiv.org/pdf/2407.09577v1 | [
"Nils Graef",
"Matthew Clapp",
"Andrew Wasielewski"
] | 2024-07-12T00:37:55Z | 2024-07-12T00:37:55Z |
2402.02619 | Increasing Trust in Language Models through the Reuse of Verified
Circuits | Language Models (LMs) are increasingly used for a wide range of prediction tasks, but their training can often neglect rare edge cases, reducing their reliability. Here, we define a stringent standard of trustworthiness whereby the task algorithm and circuit implementation must be verified, accounting for edge cases, with no known failure modes. We show that a model can be trained to meet this standard if built using mathematically and logically specified frameworks. In this paper, we fully verify an auto-regressive transformer model for n-digit integer addition. To exhibit the reusability of verified modules, we insert the trained integer addition model into a larger untrained model and train the combined model to perform both addition and subtraction. We find extensive reuse of the addition circuits for both tasks, easing verification of the more complex subtractor model. We discuss how inserting verified task modules into LMs can leverage model reuse to improve verifiability and trustworthiness of language models built using them. The reuse of verified circuits reduces the effort to verify more complex composite models which we believe to be a significant step towards safety of language models. | http://arxiv.org/pdf/2402.02619v8 | [
"Philip Quirke",
"Clement Neo",
"Fazl Barez"
] | 2024-07-12T00:34:01Z | 2024-02-04T21:33:18Z |
2303.11789 | Random Inverse Problems Over Graphs: Decentralized Online Learning | We establish a framework of distributed random inverse problems over network graphs with online measurements, and propose a decentralized online learning algorithm. This unifies the distributed parameter estimation in Hilbert spaces and the least mean square problem in reproducing kernel Hilbert spaces (RKHS-LMS). We transform the convergence of the algorithm into the asymptotic stability of a class of inhomogeneous random difference equations in Hilbert spaces with L2-bounded martingale difference terms and develop the L2 -asymptotic stability theory in Hilbert spaces. It is shown that if the network graph is connected and the sequence of forward operators satisfies the infinite-dimensional spatio-temporal persistence of excitation condition, then the estimates of all nodes are mean square and almost surely strongly consistent. Moreover, we propose a decentralized online learning algorithm in RKHS based on non-stationary and non-independent online data streams, and prove that the algorithm is mean square and almost surely strongly consistent if the operators induced by the random input data satisfy the infinite-dimensional spatio-temporal persistence of excitation condition. | http://arxiv.org/pdf/2303.11789v6 | [
"Tao Li",
"Xiwei Zhang"
] | 2024-07-12T00:17:04Z | 2023-03-20T08:37:08Z |
2407.08898 | IDAT: A Multi-Modal Dataset and Toolkit for Building and Evaluating
Interactive Task-Solving Agents | Seamless interaction between AI agents and humans using natural language remains a key goal in AI research. This paper addresses the challenges of developing interactive agents capable of understanding and executing grounded natural language instructions through the IGLU competition at NeurIPS. Despite advancements, challenges such as a scarcity of appropriate datasets and the need for effective evaluation platforms persist. We introduce a scalable data collection tool for gathering interactive grounded language instructions within a Minecraft-like environment, resulting in a Multi-Modal dataset with around 9,000 utterances and over 1,000 clarification questions. Additionally, we present a Human-in-the-Loop interactive evaluation platform for qualitative analysis and comparison of agent performance through multi-turn communication with human annotators. We offer to the community these assets referred to as IDAT (IGLU Dataset And Toolkit) which aim to advance the development of intelligent, interactive AI agents and provide essential resources for further research. | http://arxiv.org/pdf/2407.08898v1 | [
"Shrestha Mohanty",
"Negar Arabzadeh",
"Andrea Tupini",
"Yuxuan Sun",
"Alexey Skrynnik",
"Artem Zholus",
"Marc-Alexandre Côté",
"Julia Kiseleva"
] | 2024-07-12T00:07:43Z | 2024-07-12T00:07:43Z |
2402.09373 | Loss Shaping Constraints for Long-Term Time Series Forecasting | Several applications in time series forecasting require predicting multiple steps ahead. Despite the vast amount of literature in the topic, both classical and recent deep learning based approaches have mostly focused on minimising performance averaged over the predicted window. We observe that this can lead to disparate distributions of errors across forecasting steps, especially for recent transformer architectures trained on popular forecasting benchmarks. That is, optimising performance on average can lead to undesirably large errors at specific time-steps. In this work, we present a Constrained Learning approach for long-term time series forecasting that aims to find the best model in terms of average performance that respects a user-defined upper bound on the loss at each time-step. We call our approach loss shaping constraints because it imposes constraints on the loss at each time step, and leverage recent duality results to show that despite its non-convexity, the resulting problem has a bounded duality gap. We propose a practical Primal-Dual algorithm to tackle it, and demonstrate that the proposed approach exhibits competitive average performance in time series forecasting benchmarks, while shaping the distribution of errors across the predicted window. | http://arxiv.org/pdf/2402.09373v2 | [
"Ignacio Hounie",
"Javier Porras-Valenzuela",
"Alejandro Ribeiro"
] | 2024-07-11T23:43:18Z | 2024-02-14T18:20:44Z |
2407.08892 | Characterizing Prompt Compression Methods for Long Context Inference | Long context inference presents challenges at the system level with increased compute and memory requirements, as well as from an accuracy perspective in being able to reason over long contexts. Recently, several methods have been proposed to compress the prompt to reduce the context length. However, there has been little work on comparing the different proposed methods across different tasks through a standardized analysis. This has led to conflicting results. To address this, here we perform a comprehensive characterization and evaluation of different prompt compression methods. In particular, we analyze extractive compression, summarization-based abstractive compression, and token pruning methods. Surprisingly, we find that extractive compression often outperforms all the other approaches, and enables up to 10x compression with minimal accuracy degradation. Interestingly, we also find that despite several recent claims, token pruning methods often lag behind extractive compression. We only found marginal improvements on summarization tasks. | http://arxiv.org/pdf/2407.08892v1 | [
"Siddharth Jha",
"Lutfi Eren Erdogan",
"Sehoon Kim",
"Kurt Keutzer",
"Amir Gholami"
] | 2024-07-11T23:34:32Z | 2024-07-11T23:34:32Z |
2407.08890 | DeepCodeProbe: Towards Understanding What Models Trained on Code Learn | Machine learning models trained on code and related artifacts offer valuable support for software maintenance but suffer from interpretability issues due to their complex internal variables. These concerns are particularly significant in safety-critical applications where the models' decision-making processes must be reliable. The specific features and representations learned by these models remain unclear, adding to the hesitancy in adopting them widely. To address these challenges, we introduce DeepCodeProbe, a probing approach that examines the syntax and representation learning abilities of ML models designed for software maintenance tasks. Our study applies DeepCodeProbe to state-of-the-art models for code clone detection, code summarization, and comment generation. Findings reveal that while small models capture abstract syntactic representations, their ability to fully grasp programming language syntax is limited. Increasing model capacity improves syntax learning but introduces trade-offs such as increased training time and overfitting. DeepCodeProbe also identifies specific code patterns the models learn from their training data. Additionally, we provide best practices for training models on code to enhance performance and interpretability, supported by an open-source replication package for broader application of DeepCodeProbe in interpreting other code-related models. | http://arxiv.org/pdf/2407.08890v1 | [
"Vahid Majdinasab",
"Amin Nikanjam",
"Foutse Khomh"
] | 2024-07-11T23:16:44Z | 2024-07-11T23:16:44Z |
2407.08888 | Uncovering Semantics and Topics Utilized by Threat Actors to Deliver
Malicious Attachments and URLs | Recent threat reports highlight that email remains the top vector for delivering malware to endpoints. Despite these statistics, detecting malicious email attachments and URLs often neglects semantic cues linguistic features and contextual clues. Our study employs BERTopic unsupervised topic modeling to identify common semantics and themes embedded in email to deliver malicious attachments and call-to-action URLs. We preprocess emails by extracting and sanitizing content and employ multilingual embedding models like BGE-M3 for dense representations, which clustering algorithms(HDBSCAN and OPTICS) use to group emails by semantic similarity. Phi3-Mini-4K-Instruct facilitates semantic and hLDA aid in thematic analysis to understand threat actor patterns. Our research will evaluate and compare different clustering algorithms on topic quantity, coherence, and diversity metrics, concluding with insights into the semantics and topics commonly used by threat actors to deliver malicious attachments and URLs, a significant contribution to the field of threat detection. | http://arxiv.org/pdf/2407.08888v1 | [
"Andrey Yakymovych",
"Abhishek Singh"
] | 2024-07-11T23:04:16Z | 2024-07-11T23:04:16Z |
2310.16228 | On the Foundations of Shortcut Learning | Deep-learning models can extract a rich assortment of features from data. Which features a model uses depends not only on emph{predictivity} -- how reliably a feature indicates training-set labels -- but also on emph{availability} -- how easily the feature can be extracted from inputs. The literature on shortcut learning has noted examples in which models privilege one feature over another, for example texture over shape and image backgrounds over foreground objects. Here, we test hypotheses about which input properties are more available to a model, and systematically study how predictivity and availability interact to shape models' feature use. We construct a minimal, explicit generative framework for synthesizing classification datasets with two latent features that vary in predictivity and in factors we hypothesize to relate to availability, and we quantify a model's shortcut bias -- its over-reliance on the shortcut (more available, less predictive) feature at the expense of the core (less available, more predictive) feature. We find that linear models are relatively unbiased, but introducing a single hidden layer with ReLU or Tanh units yields a bias. Our empirical findings are consistent with a theoretical account based on Neural Tangent Kernels. Finally, we study how models used in practice trade off predictivity and availability in naturalistic datasets, discovering availability manipulations which increase models' degree of shortcut bias. Taken together, these findings suggest that the propensity to learn shortcut features is a fundamental characteristic of deep nonlinear architectures warranting systematic study given its role in shaping how models solve tasks. | http://arxiv.org/pdf/2310.16228v2 | [
"Katherine L. Hermann",
"Hossein Mobahi",
"Thomas Fel",
"Michael C. Mozer"
] | 2024-07-11T23:03:09Z | 2023-10-24T22:54:05Z |
2407.08887 | Automatic Pruning of Fine-tuning Datasets for Transformer-based Language
Models | Transformer-based language models have shown state-of-the-art performance on a variety of natural language understanding tasks. To achieve this performance, these models are first pre-trained on general corpus and then fine-tuned on downstream tasks. Previous work studied the effect of pruning the training set of the downstream tasks on the performance of the model on its evaluation set. In this work, we propose an automatic dataset pruning method for the training set of fine-tuning tasks. Our method is based on the model's success rate in correctly classifying each training data point. Unlike previous work which relies on user feedback to determine subset size, our method automatically extracts training subsets that are adapted for each pair of model and fine-tuning task. Our method provides multiple subsets for use in dataset pruning that navigate the trade-off between subset size and evaluation accuracy. Our largest subset, which we also refer to as the winning ticket subset, is on average $3 times$ smaller than the original training set of the fine-tuning task. Our experiments on 5 downstream tasks and 2 language models show that, on average, fine-tuning on the winning ticket subsets results in a $0.1 %$ increase in the evaluation performance of the model. | http://arxiv.org/pdf/2407.08887v1 | [
"Mohammadreza Tayaranian",
"Seyyed Hasan Mozafari",
"Brett H. Meyer",
"James J. Clark",
"Warren J. Gross"
] | 2024-07-11T22:46:18Z | 2024-07-11T22:46:18Z |
2407.08886 | Semi-Supervised Multi-Task Learning Based Framework for Power System
Security Assessment | This paper develops a novel machine learning-based framework using Semi-Supervised Multi-Task Learning (SS-MTL) for power system dynamic security assessment that is accurate, reliable, and aware of topological changes. The learning algorithm underlying the proposed framework integrates conditional masked encoders and employs multi-task learning for classification-aware feature representation, which improves the accuracy and scalability to larger systems. Additionally, this framework incorporates a confidence measure for its predictions, enhancing its reliability and interpretability. A topological similarity index has also been incorporated to add topological awareness to the framework. Various experiments on the IEEE 68-bus system were conducted to validate the proposed method, employing two distinct database generation techniques to generate the required data to train the machine learning algorithm. The results demonstrate that our algorithm outperforms existing state-of-the-art machine learning based techniques for security assessment in terms of accuracy and robustness. Finally, our work underscores the value of employing auto-encoders for security assessment, highlighting improvements in accuracy, reliability, and robustness. All datasets and codes used have been made publicly available to ensure reproducibility and transparency. | http://arxiv.org/pdf/2407.08886v1 | [
"Muhy Eddin Za'ter",
"Amirhossein Sajadi",
"Bri-Mathias Hodge"
] | 2024-07-11T22:42:53Z | 2024-07-11T22:42:53Z |
2309.01610 | Fairness in Ranking under Disparate Uncertainty | Ranking is a ubiquitous method for focusing the attention of human evaluators on a manageable subset of options. Its use as part of human decision-making processes ranges from surfacing potentially relevant products on an e-commerce site to prioritizing college applications for human review. While ranking can make human evaluation more effective by focusing attention on the most promising options, we argue that it can introduce unfairness if the uncertainty of the underlying relevance model differs between groups of options. Unfortunately, such disparity in uncertainty appears widespread, often to the detriment of minority groups for which relevance estimates can have higher uncertainty due to a lack of data or appropriate features. To address this fairness issue, we propose Equal-Opportunity Ranking (EOR) as a new fairness criterion for ranking and show that it corresponds to a group-wise fair lottery among the relevant options even in the presence of disparate uncertainty. EOR optimizes for an even cost burden on all groups, unlike the conventional Probability Ranking Principle, and is fundamentally different from existing notions of fairness in rankings, such as demographic parity and proportional Rooney rule constraints that are motivated by proportional representation relative to group size. To make EOR ranking practical, we present an efficient algorithm for computing it in time $O(n log(n))$ and prove its close approximation guarantee to the globally optimal solution. In a comprehensive empirical evaluation on synthetic data, a US Census dataset, and a real-world audit of Amazon search queries, we find that the algorithm reliably guarantees EOR fairness while providing effective rankings. | http://arxiv.org/pdf/2309.01610v3 | [
"Richa Rastogi",
"Thorsten Joachims"
] | 2024-07-11T21:35:37Z | 2023-09-04T13:49:48Z |
2402.02625 | Enhancing Transformer RNNs with Multiple Temporal Perspectives | We introduce the concept of multiple temporal perspectives, a novel approach applicable to Recurrent Neural Network (RNN) architectures for enhancing their understanding of sequential data. This method involves maintaining diverse temporal views of previously encountered text, significantly enriching the language models' capacity to interpret context. To show the efficacy of this approach, we incorporate it into the Receptance Weighted Key Value (RWKV) architecture, addressing its inherent challenge of retaining all historical information within a single hidden state. Notably, this improvement is achieved with a minimal increase in the number of parameters --even as little as $0.04%$ of the original number of parameters. Further, the additional parameters necessary for the multiple temporal perspectives are fine-tuned with minimal computational overhead, avoiding the need for a full pre-training. The resulting model maintains linear computational complexity during prompt inference, ensuring consistent efficiency across various sequence lengths. The empirical results and ablation studies included in our research validate the effectiveness of our approach, showcasing improved performance across multiple benchmarks. The code, model weights and datasets are open-sourced at: https://github.com/RazvanDu/TemporalRNNs. | http://arxiv.org/pdf/2402.02625v2 | [
"Razvan-Gabriel Dumitru",
"Darius Peteleaza",
"Mihai Surdeanu"
] | 2024-07-11T20:43:59Z | 2024-02-04T22:12:29Z |
2401.03609 | Multi-Modal Federated Learning for Cancer Staging over Non-IID Datasets
with Unbalanced Modalities | The use of machine learning (ML) for cancer staging through medical image analysis has gained substantial interest across medical disciplines. When accompanied by the innovative federated learning (FL) framework, ML techniques can further overcome privacy concerns related to patient data exposure. Given the frequent presence of diverse data modalities within patient records, leveraging FL in a multi-modal learning framework holds considerable promise for cancer staging. However, existing works on multi-modal FL often presume that all data-collecting institutions have access to all data modalities. This oversimplified approach neglects institutions that have access to only a portion of data modalities within the system. In this work, we introduce a novel FL architecture designed to accommodate not only the heterogeneity of data samples, but also the inherent heterogeneity/non-uniformity of data modalities across institutions. We shed light on the challenges associated with varying convergence speeds observed across different data modalities within our FL system. Subsequently, we propose a solution to tackle these challenges by devising a distributed gradient blending and proximity-aware client weighting strategy tailored for multi-modal FL. To show the superiority of our method, we conduct experiments using The Cancer Genome Atlas program (TCGA) datalake considering different cancer types and three modalities of data: mRNA sequences, histopathological image data, and clinical information. Our results further unveil the impact and severity of class-based vs type-based heterogeneity across institutions on the model performance, which widens the perspective to the notion of data heterogeneity in multi-modal FL literature. | http://arxiv.org/pdf/2401.03609v2 | [
"Kasra Borazjani",
"Naji Khosravan",
"Leslie Ying",
"Seyyedali Hosseinalipour"
] | 2024-07-11T20:12:22Z | 2024-01-07T23:45:01Z |
2403.08845 | Bifurcated Attention: Accelerating Massively Parallel Decoding with
Shared Prefixes in LLMs | This study introduces bifurcated attention, a method designed to enhance language model inference in shared-context batch decoding scenarios. Our approach addresses the challenge of redundant memory IO costs, a critical factor contributing to latency in high batch sizes and extended context lengths. Bifurcated attention achieves this by strategically dividing the attention mechanism during incremental decoding into two separate GEMM operations: one focusing on the KV cache from prefill, and another on the decoding process itself. While maintaining the computational load (FLOPs) of standard attention mechanisms, bifurcated attention ensures precise computation with significantly reduced memory IO. Our empirical results show over 2.1$times$ speedup when sampling 16 output sequences and more than 6.2$times$ speedup when sampling 32 sequences at context lengths exceeding 8k tokens on a 7B model that uses multi-head attention. The efficiency gains from bifurcated attention translate into lower latency, making it particularly suitable for real-time applications. For instance, it enables massively parallel answer generation without substantially increasing latency, thus enhancing performance when integrated with post-processing techniques such as re-ranking. | http://arxiv.org/pdf/2403.08845v2 | [
"Ben Athiwaratkun",
"Sujan Kumar Gonugondla",
"Sanjay Krishna Gouda",
"Haifeng Qian",
"Hantian Ding",
"Qing Sun",
"Jun Wang",
"Jiacheng Guo",
"Liangfu Chen",
"Parminder Bhatia",
"Ramesh Nallapati",
"Sudipta Sengupta",
"Bing Xiang"
] | 2024-07-11T20:07:30Z | 2024-03-13T16:30:57Z |
2310.04585 | Interventions Against Machine-Assisted Statistical Discrimination | I study statistical discrimination driven by verifiable beliefs, such as those generated by machine learning, rather than by humans. When beliefs are verifiable, interventions against statistical discrimination can move beyond simple, belief-free designs like affirmative action, to more sophisticated ones, that constrain decision makers based on what they are thinking. Such mind reading interventions can perform well where affirmative action does not, even when the minds being read are biased. My theory of belief-contingent intervention design sheds light on influential methods of regulating machine learning, and yields novel interventions robust to covariate shift and incorrect, biased beliefs. | http://arxiv.org/pdf/2310.04585v3 | [
"John Y. Zhu"
] | 2024-07-11T20:01:41Z | 2023-10-06T20:57:34Z |
2407.08843 | Inflationary Flows: Calibrated Bayesian Inference with Diffusion-Based
Models | Beyond estimating parameters of interest from data, one of the key goals of statistical inference is to properly quantify uncertainty in these estimates. In Bayesian inference, this uncertainty is provided by the posterior distribution, the computation of which typically involves an intractable high-dimensional integral. Among available approximation methods, sampling-based approaches come with strong theoretical guarantees but scale poorly to large problems, while variational approaches scale well but offer few theoretical guarantees. In particular, variational methods are known to produce overconfident estimates of posterior uncertainty and are typically non-identifiable, with many latent variable configurations generating equivalent predictions. Here, we address these challenges by showing how diffusion-based models (DBMs), which have recently produced state-of-the-art performance in generative modeling tasks, can be repurposed for performing calibrated, identifiable Bayesian inference. By exploiting a previously established connection between the stochastic and probability flow ordinary differential equations (pfODEs) underlying DBMs, we derive a class of models, inflationary flows, that uniquely and deterministically map high-dimensional data to a lower-dimensional Gaussian distribution via ODE integration. This map is both invertible and neighborhood-preserving, with controllable numerical error, with the result that uncertainties in the data are correctly propagated to the latent space. We demonstrate how such maps can be learned via standard DBM training using a novel noise schedule and are effective at both preserving and reducing intrinsic data dimensionality. The result is a class of highly expressive generative models, uniquely defined on a low-dimensional latent space, that afford principled Bayesian inference. | http://arxiv.org/pdf/2407.08843v1 | [
"Daniela de Albuquerque",
"John Pearson"
] | 2024-07-11T19:58:19Z | 2024-07-11T19:58:19Z |
2407.08840 | Data-driven Model Reduction for Soft Robots via Lagrangian Operator
Inference | Data-driven model reduction methods provide a nonintrusive way of constructing computationally efficient surrogates of high-fidelity models for real-time control of soft robots. This work leverages the Lagrangian nature of the model equations to derive structure-preserving linear reduced-order models via Lagrangian Operator Inference and compares their performance with prominent linear model reduction techniques through an anguilliform swimming soft robot model example with 231,336 degrees of freedom. The case studies demonstrate that preserving the underlying Lagrangian structure leads to learned models with higher predictive accuracy and robustness to unseen inputs. | http://arxiv.org/pdf/2407.08840v1 | [
"Harsh Sharma",
"Iman Adibnazari",
"Jacobo Cervera-Torralba",
"Michael T. Tolley",
"Boris Kramer"
] | 2024-07-11T19:55:21Z | 2024-07-11T19:55:21Z |
2407.08839 | A Survey on the Application of Generative Adversarial Networks in
Cybersecurity: Prospective, Direction and Open Research Scopes | With the proliferation of Artificial Intelligence, there has been a massive increase in the amount of data required to be accumulated and disseminated digitally. As the data are available online in digital landscapes with complex and sophisticated infrastructures, it is crucial to implement various defense mechanisms based on cybersecurity. Generative Adversarial Networks (GANs), which are deep learning models, have emerged as powerful solutions for addressing the constantly changing security issues. This survey studies the significance of the deep learning model, precisely on GANs, in strengthening cybersecurity defenses. Our survey aims to explore the various works completed in GANs, such as Intrusion Detection Systems (IDS), Mobile and Network Trespass, BotNet Detection, and Malware Detection. The focus is to examine how GANs can be influential tools to strengthen cybersecurity defenses in these domains. Further, the paper discusses the challenges and constraints of using GANs in these areas and suggests future research directions. Overall, the paper highlights the potential of GANs in enhancing cybersecurity measures and addresses the need for further exploration in this field. | http://arxiv.org/pdf/2407.08839v1 | [
"Md Mashrur Arifin",
"Md Shoaib Ahmed",
"Tanmai Kumar Ghosh",
"Jun Zhuang",
"Jyh-haw Yeh"
] | 2024-07-11T19:51:48Z | 2024-07-11T19:51:48Z |
2407.08838 | Deep Learning for Network Anomaly Detection under Data Contamination:
Evaluating Robustness and Mitigating Performance Degradation | Deep learning (DL) has emerged as a crucial tool in network anomaly detection (NAD) for cybersecurity. While DL models for anomaly detection excel at extracting features and learning patterns from data, they are vulnerable to data contamination -- the inadvertent inclusion of attack-related data in training sets presumed benign. This study evaluates the robustness of six unsupervised DL algorithms against data contamination using our proposed evaluation protocol. Results demonstrate significant performance degradation in state-of-the-art anomaly detection algorithms when exposed to contaminated data, highlighting the critical need for self-protection mechanisms in DL-based NAD models. To mitigate this vulnerability, we propose an enhanced auto-encoder with a constrained latent representation, allowing normal data to cluster more densely around a learnable center in the latent space. Our evaluation reveals that this approach exhibits improved resistance to data contamination compared to existing methods, offering a promising direction for more robust NAD systems. | http://arxiv.org/pdf/2407.08838v1 | [
"D'Jeff K. Nkashama",
"Jordan Masakuna Félicien",
"Arian Soltani",
"Jean-Charles Verdier",
"Pierre-Martin Tardif",
"Marc Frappier",
"Froduald Kabanza"
] | 2024-07-11T19:47:37Z | 2024-07-11T19:47:37Z |
2405.14422 | Unraveling overoptimism and publication bias in ML-driven science | Machine Learning (ML) is increasingly used across many disciplines with impressive reported results. However, recent studies suggest published performance of ML models are often overoptimistic. Validity concerns are underscored by findings of an inverse relationship between sample size and reported accuracy in published ML models, contrasting with the theory of learning curves where accuracy should improve or remain stable with increasing sample size. This paper investigates factors contributing to overoptimism in ML-driven science, focusing on overfitting and publication bias. We introduce a novel stochastic model for observed accuracy, integrating parametric learning curves and the aforementioned biases. We construct an estimator that corrects for these biases in observed data. Theoretical and empirical results show that our framework can estimate the underlying learning curve, providing realistic performance assessments from published results. Applying the model to meta-analyses of classifications of neurological conditions, we estimate the inherent limits of ML-based prediction in each domain. | http://arxiv.org/pdf/2405.14422v3 | [
"Pouria Saidi",
"Gautam Dasarathy",
"Visar Berisha"
] | 2024-07-11T19:40:20Z | 2024-05-23T10:43:20Z |
2407.08824 | Proving that Cryptic Crossword Clue Answers are Correct | Cryptic crossword clues are challenging cognitive tasks, for which new test sets are released on a daily basis by multiple international newspapers. Each cryptic clue contains both the definition of the answer to be placed in the crossword grid (in common with regular crosswords), and `wordplay' that proves that the answer is correct (i.e. a human solver can be confident that an answer is correct without needing crossing words to confirm it). Using an existing cryptic wordplay proving framework (operating on Python proofs created by an LLM), we show that it is possible to distinguish between correct answers and almost-correct ones based upon whether the wordplay `works'. | http://arxiv.org/pdf/2407.08824v1 | [
"Martin Andrews",
"Sam Witteveen"
] | 2024-07-11T19:13:16Z | 2024-07-11T19:13:16Z |
2407.08806 | HO-FMN: Hyperparameter Optimization for Fast Minimum-Norm Attacks | Gradient-based attacks are a primary tool to evaluate robustness of machine-learning models. However, many attacks tend to provide overly-optimistic evaluations as they use fixed loss functions, optimizers, step-size schedulers, and default hyperparameters. In this work, we tackle these limitations by proposing a parametric variation of the well-known fast minimum-norm attack algorithm, whose loss, optimizer, step-size scheduler, and hyperparameters can be dynamically adjusted. We re-evaluate 12 robust models, showing that our attack finds smaller adversarial perturbations without requiring any additional tuning. This also enables reporting adversarial robustness as a function of the perturbation budget, providing a more complete evaluation than that offered by fixed-budget attacks, while remaining efficient. We release our open-source code at https://github.com/pralab/HO-FMN. | http://arxiv.org/pdf/2407.08806v1 | [
"Raffaele Mura",
"Giuseppe Floris",
"Luca Scionis",
"Giorgio Piras",
"Maura Pintor",
"Ambra Demontis",
"Giorgio Giacinto",
"Battista Biggio",
"Fabio Roli"
] | 2024-07-11T18:30:01Z | 2024-07-11T18:30:01Z |
2309.07085 | Mitigating Group Bias in Federated Learning for Heterogeneous Devices | Federated Learning is emerging as a privacy-preserving model training approach in distributed edge applications. As such, most edge deployments are heterogeneous in nature i.e., their sensing capabilities and environments vary across deployments. This edge heterogeneity violates the independence and identical distribution (IID) property of local data across clients and produces biased global models i.e. models that contribute to unfair decision-making and discrimination against a particular community or a group. Existing bias mitigation techniques only focus on bias generated from label heterogeneity in non-IID data without accounting for domain variations due to feature heterogeneity and do not address global group-fairness property. Our work proposes a group-fair FL framework that minimizes group-bias while preserving privacy and without resource utilization overhead. Our main idea is to leverage average conditional probabilities to compute a cross-domain group textit{importance weights} derived from heterogeneous training data to optimize the performance of the worst-performing group using a modified multiplicative weights update method. Additionally, we propose regularization techniques to minimize the difference between the worst and best-performing groups while making sure through our thresholding mechanism to strike a balance between bias reduction and group performance degradation. Our evaluation of human emotion recognition and image classification benchmarks assesses the fair decision-making of our framework in real-world heterogeneous settings. | http://arxiv.org/abs/2309.07085v2 | [
"Khotso Selialia",
"Yasra Chandio",
"Fatima M. Anwar"
] | 2024-07-11T18:25:51Z | 2023-09-13T16:53:48Z |
2407.08803 | PID Accelerated Temporal Difference Algorithms | Long-horizon tasks, which have a large discount factor, pose a challenge for most conventional reinforcement learning (RL) algorithms. Algorithms such as Value Iteration and Temporal Difference (TD) learning have a slow convergence rate and become inefficient in these tasks. When the transition distributions are given, PID VI was recently introduced to accelerate the convergence of Value Iteration using ideas from control theory. Inspired by this, we introduce PID TD Learning and PID Q-Learning algorithms for the RL setting in which only samples from the environment are available. We give theoretical analysis of their convergence and acceleration compared to their traditional counterparts. We also introduce a method for adapting PID gains in the presence of noise and empirically verify its effectiveness. | http://arxiv.org/pdf/2407.08803v1 | [
"Mark Bedaywi",
"Amin Rakhsha",
"Amir-massoud Farahmand"
] | 2024-07-11T18:23:46Z | 2024-07-11T18:23:46Z |
2402.03855 | Challenges in Mechanistically Interpreting Model Representations | Mechanistic interpretability (MI) aims to understand AI models by reverse-engineering the exact algorithms neural networks learn. Most works in MI so far have studied behaviors and capabilities that are trivial and token-aligned. However, most capabilities important for safety and trust are not that trivial, which advocates for the study of hidden representations inside these networks as the unit of analysis. We formalize representations for features and behaviors, highlight their importance and evaluation, and perform an exploratory study of dishonesty representations in `Mistral-7B-Instruct-v0.1'. We justify that studying representations is an important and under-studied field, and highlight several challenges that arise while attempting to do so through currently established methods in MI, showing their insufficiency and advocating work on new frameworks for the same. | http://arxiv.org/pdf/2402.03855v2 | [
"Satvik Golechha",
"James Dao"
] | 2024-07-11T18:21:59Z | 2024-02-06T10:06:13Z |
2407.08800 | Local Clustering for Lung Cancer Image Classification via Sparse
Solution Technique | In this work, we propose to use a local clustering approach based on the sparse solution technique to study the medical image, especially the lung cancer image classification task. We view images as the vertices in a weighted graph and the similarity between a pair of images as the edges in the graph. The vertices within the same cluster can be assumed to share similar features and properties, thus making the applications of graph clustering techniques very useful for image classification. Recently, the approach based on the sparse solutions of linear systems for graph clustering has been found to identify clusters more efficiently than traditional clustering methods such as spectral clustering. We propose to use the two newly developed local clustering methods based on sparse solution of linear system for image classification. In addition, we employ a box spline-based tight-wavelet-framelet method to clean these images and help build a better adjacency matrix before clustering. The performance of our methods is shown to be very effective in classifying images. Our approach is significantly more efficient and either favorable or equally effective compared with other state-of-the-art approaches. Finally, we shall make a remark by pointing out two image deformation methods to build up more artificial image data to increase the number of labeled images. | http://arxiv.org/pdf/2407.08800v1 | [
"Jackson Hamel",
"Ming-Jun Lai",
"Zhaiming Shen",
"Ye Tian"
] | 2024-07-11T18:18:32Z | 2024-07-11T18:18:32Z |
2407.08797 | Deep Inverse Design for High-Level Synthesis | High-level synthesis (HLS) has significantly advanced the automation of digital circuits design, yet the need for expertise and time in pragma tuning remains challenging. Existing solutions for the design space exploration (DSE) adopt either heuristic methods, lacking essential information for further optimization potential, or predictive models, missing sufficient generalization due to the time-consuming nature of HLS and the exponential growth of the design space. To address these challenges, we propose Deep Inverse Design for HLS (DID4HLS), a novel approach that integrates graph neural networks and generative models. DID4HLS iteratively optimizes hardware designs aimed at compute-intensive algorithms by learning conditional distributions of design features from post-HLS data. Compared to four state-of-the-art DSE baselines, our method achieved an average improvement of 42.5% on average distance to reference set (ADRS) compared to the best-performing baselines across six benchmarks, while demonstrating high robustness and efficiency. | http://arxiv.org/pdf/2407.08797v1 | [
"Ping Chang",
"Tosiron Adegbija",
"Yuchao Liao",
"Claudio Talarico",
"Ao Li",
"Janet Roveda"
] | 2024-07-11T18:13:38Z | 2024-07-11T18:13:38Z |
2403.14183 | OTSeg: Multi-prompt Sinkhorn Attention for Zero-Shot Semantic
Segmentation | The recent success of CLIP has demonstrated promising results in zero-shot semantic segmentation by transferring muiltimodal knowledge to pixel-level classification. However, leveraging pre-trained CLIP knowledge to closely align text embeddings with pixel embeddings still has limitations in existing approaches. To address this issue, we propose OTSeg, a novel multimodal attention mechanism aimed at enhancing the potential of multiple text prompts for matching associated pixel embeddings. We first propose Multi-Prompts Sinkhorn (MPS) based on the Optimal Transport (OT) algorithm, which leads multiple text prompts to selectively focus on various semantic features within image pixels. Moreover, inspired by the success of Sinkformers in unimodal settings, we introduce the extension of MPS, called Multi-Prompts Sinkhorn Attention (MPSA) , which effectively replaces cross-attention mechanisms within Transformer framework in multimodal settings. Through extensive experiments, we demonstrate that OTSeg achieves state-of-the-art (SOTA) performance with significant gains on Zero-Shot Semantic Segmentation (ZS3) tasks across three benchmark datasets. | http://arxiv.org/pdf/2403.14183v2 | [
"Kwanyoung Kim",
"Yujin Oh",
"Jong Chul Ye"
] | 2024-07-11T18:09:48Z | 2024-03-21T07:15:37Z |
2306.06094 | Leveraging Large Language Models for Scalable Vector Graphics-Driven
Image Understanding | Large language models (LLMs) have made significant advancements in natural language understanding. However, through that enormous semantic representation that the LLM has learnt, is it somehow possible for it to understand images as well? This work investigates this question. To enable the LLM to process images, we convert them into a representation given by Scalable Vector Graphics (SVG). To study what the LLM can do with this XML-based textual description of images, we test the LLM on three broad computer vision tasks: (i) visual reasoning and question answering, (ii) image classification under distribution shift, few-shot learning, and (iii) generating new images using visual prompting. Even though we do not naturally associate LLMs with any visual understanding capabilities, our results indicate that the LLM can often do a decent job in many of these tasks, potentially opening new avenues for research into LLMs' ability to understand image data. Our code, data, and models can be found here https://github.com/mu-cai/svg-llm. | http://arxiv.org/pdf/2306.06094v2 | [
"Mu Cai",
"Zeyi Huang",
"Yuheng Li",
"Utkarsh Ojha",
"Haohan Wang",
"Yong Jae Lee"
] | 2024-07-11T17:59:53Z | 2023-06-09T17:57:01Z |
2301.11329 | Anatomy-aware and acquisition-agnostic joint registration with
SynthMorph | Affine image registration is a cornerstone of medical image analysis. While classical algorithms can achieve excellent accuracy, they solve a time-consuming optimization for every image pair. Deep-learning (DL) methods learn a function that maps an image pair to an output transform. Evaluating the function is fast, but capturing large transforms can be challenging, and networks tend to struggle if a test-image characteristic shifts from the training domain, such as resolution. Most affine methods are agnostic to the anatomy the user wishes to align, meaning the registration will be inaccurate if algorithms consider all structures in the image. We address these shortcomings with SynthMorph, a fast, symmetric, diffeomorphic, and easy-to-use DL tool for joint affine-deformable registration of any brain image without preprocessing. First, we leverage a strategy that trains networks with widely varying images synthesized from label maps, yielding robust performance for image types unseen at training. Second, we optimize the spatial overlap of select anatomical labels. This enables networks to distinguish anatomy of interest from irrelevant structures, removing the need for preprocessing that excludes content that may impinge on anatomy-specific registration. Third, we combine the affine model with a deformable hypernetwork that lets users choose the optimal deformation-field regularity for their specific data, at registration time, in a fraction of the time required by classical methods. We analyze how competing architectures learn affine transforms and compare state-of-the-art registration tools across an extremely diverse set of neuroimaging data, aiming to truly capture the behavior of methods in the real world. SynthMorph demonstrates high accuracy and is available at https://w3id.org/synthmorph, as a single complete end-to-end solution for registration of brain MRI. | http://arxiv.org/abs/2301.11329v4 | [
"Malte Hoffmann",
"Andrew Hoopes",
"Douglas N. Greve",
"Bruce Fischl",
"Adrian V. Dalca"
] | 2024-07-11T17:59:50Z | 2023-01-26T18:59:33Z |
2407.08737 | Video Diffusion Alignment via Reward Gradients | We have made significant progress towards building foundational video diffusion models. As these models are trained using large-scale unsupervised data, it has become crucial to adapt these models to specific downstream tasks. Adapting these models via supervised fine-tuning requires collecting target datasets of videos, which is challenging and tedious. In this work, we utilize pre-trained reward models that are learned via preferences on top of powerful vision discriminative models to adapt video diffusion models. These models contain dense gradient information with respect to generated RGB pixels, which is critical to efficient learning in complex search spaces, such as videos. We show that backpropagating gradients from these reward models to a video diffusion model can allow for compute and sample efficient alignment of the video diffusion model. We show results across a variety of reward models and video diffusion models, demonstrating that our approach can learn much more efficiently in terms of reward queries and computation than prior gradient-free approaches. Our code, model weights,and more visualization are available at https://vader-vid.github.io. | http://arxiv.org/pdf/2407.08737v1 | [
"Mihir Prabhudesai",
"Russell Mendonca",
"Zheyang Qin",
"Katerina Fragkiadaki",
"Deepak Pathak"
] | 2024-07-11T17:59:45Z | 2024-07-11T17:59:45Z |
2407.08734 | Transformer Circuit Faithfulness Metrics are not Robust | Mechanistic interpretability work attempts to reverse engineer the learned algorithms present inside neural networks. One focus of this work has been to discover 'circuits' -- subgraphs of the full model that explain behaviour on specific tasks. But how do we measure the performance of such circuits? Prior work has attempted to measure circuit 'faithfulness' -- the degree to which the circuit replicates the performance of the full model. In this work, we survey many considerations for designing experiments that measure circuit faithfulness by ablating portions of the model's computation. Concerningly, we find existing methods are highly sensitive to seemingly insignificant changes in the ablation methodology. We conclude that existing circuit faithfulness scores reflect both the methodological choices of researchers as well as the actual components of the circuit - the task a circuit is required to perform depends on the ablation used to test it. The ultimate goal of mechanistic interpretability work is to understand neural networks, so we emphasize the need for more clarity in the precise claims being made about circuits. We open source a library at https://github.com/UFO-101/auto-circuit that includes highly efficient implementations of a wide range of ablation methodologies and circuit discovery algorithms. | http://arxiv.org/pdf/2407.08734v1 | [
"Joseph Miller",
"Bilal Chughtai",
"William Saunders"
] | 2024-07-11T17:59:00Z | 2024-07-11T17:59:00Z |
2407.08729 | BiEquiFormer: Bi-Equivariant Representations for Global Point Cloud
Registration | The goal of this paper is to address the problem of textit{global} point cloud registration (PCR) i.e., finding the optimal alignment between point clouds irrespective of the initial poses of the scans. This problem is notoriously challenging for classical optimization methods due to computational constraints. First, we show that state-of-the-art deep learning methods suffer from huge performance degradation when the point clouds are arbitrarily placed in space. We propose that textit{equivariant deep learning} should be utilized for solving this task and we characterize the specific type of bi-equivariance of PCR. Then, we design BiEquiformer a novel and scalable textit{bi-equivariant} pipeline i.e. equivariant to the independent transformations of the input point clouds. While a naive approach would process the point clouds independently we design expressive bi-equivariant layers that fuse the information from both point clouds. This allows us to extract high-quality superpoint correspondences and in turn, robust point-cloud registration. Extensive comparisons against state-of-the-art methods show that our method achieves comparable performance in the canonical setting and superior performance in the robust setting in both the 3DMatch and the challenging low-overlap 3DLoMatch dataset. | http://arxiv.org/pdf/2407.08729v1 | [
"Stefanos Pertigkiozoglou",
"Evangelos Chatzipantazis",
"Kostas Daniilidis"
] | 2024-07-11T17:58:10Z | 2024-07-11T17:58:10Z |
2407.08723 | Topological Generalization Bounds for Discrete-Time Stochastic
Optimization Algorithms | We present a novel set of rigorous and computationally efficient topology-based complexity notions that exhibit a strong correlation with the generalization gap in modern deep neural networks (DNNs). DNNs show remarkable generalization properties, yet the source of these capabilities remains elusive, defying the established statistical learning theory. Recent studies have revealed that properties of training trajectories can be indicative of generalization. Building on this insight, state-of-the-art methods have leveraged the topology of these trajectories, particularly their fractal dimension, to quantify generalization. Most existing works compute this quantity by assuming continuous- or infinite-time training dynamics, complicating the development of practical estimators capable of accurately predicting generalization without access to test data. In this paper, we respect the discrete-time nature of training trajectories and investigate the underlying topological quantities that can be amenable to topological data analysis tools. This leads to a new family of reliable topological complexity measures that provably bound the generalization error, eliminating the need for restrictive geometric assumptions. These measures are computationally friendly, enabling us to propose simple yet effective algorithms for computing generalization indices. Moreover, our flexible framework can be extended to different domains, tasks, and architectures. Our experimental results demonstrate that our new complexity measures correlate highly with generalization error in industry-standards architectures such as transformers and deep graph networks. Our approach consistently outperforms existing topological bounds across a wide range of datasets, models, and optimizers, highlighting the practical relevance and effectiveness of our complexity measures. | http://arxiv.org/pdf/2407.08723v1 | [
"Rayna Andreeva",
"Benjamin Dupuis",
"Rik Sarkar",
"Tolga Birdal",
"Umut Şimşekli"
] | 2024-07-11T17:56:03Z | 2024-07-11T17:56:03Z |
2407.08722 | Unifying 3D Representation and Control of Diverse Robots with a Single
Camera | Mirroring the complex structures and diverse functions of natural organisms is a long-standing challenge in robotics. Modern fabrication techniques have dramatically expanded feasible hardware, yet deploying these systems requires control software to translate desired motions into actuator commands. While conventional robots can easily be modeled as rigid links connected via joints, it remains an open challenge to model and control bio-inspired robots that are often multi-material or soft, lack sensing capabilities, and may change their material properties with use. Here, we introduce Neural Jacobian Fields, an architecture that autonomously learns to model and control robots from vision alone. Our approach makes no assumptions about the robot's materials, actuation, or sensing, requires only a single camera for control, and learns to control the robot without expert intervention by observing the execution of random commands. We demonstrate our method on a diverse set of robot manipulators, varying in actuation, materials, fabrication, and cost. Our approach achieves accurate closed-loop control and recovers the causal dynamic structure of each robot. By enabling robot control with a generic camera as the only sensor, we anticipate our work will dramatically broaden the design space of robotic systems and serve as a starting point for lowering the barrier to robotic automation. | http://arxiv.org/pdf/2407.08722v1 | [
"Sizhe Lester Li",
"Annan Zhang",
"Boyuan Chen",
"Hanna Matusik",
"Chao Liu",
"Daniela Rus",
"Vincent Sitzmann"
] | 2024-07-11T17:55:49Z | 2024-07-11T17:55:49Z |