Bowen Liu
In this paper, we obtain the existence of non-planar circular homographic solutions and non-circular homographic solutions of the $(2+N)$- and $(3+N)$-body problems of the Lennard-Jones system. These results show the essential difference between the Lennard-Jones potential and the Newton's potential of universal gravitation.
Bowen Liu, Changwoo Lee, Ang Cao, Hun-Seok Kim
We propose a unified signal compression framework that uses a generative adversarial network (GAN) to compress heterogeneous signals. The compressed signal is represented as a latent vector and fed into a generator network that is trained to produce high quality realistic signals that minimize a target objective function. To efficiently quantize the compressed signal, non-uniformly quantized optimal latent vectors are identified by iterative back-propagation with alternating direction method of multipliers (ADMM) optimization performed for each iteration. The performance of the proposed signal compression method is assessed using multiple metrics including PSNR and MS-SSIM for image compression and also PESR, Kaldi, LSTM, and MLP performance for speech compression. Test results show that the proposed work outperforms recent state-of-the-art hand-crafted and deep learning-based signal compression methods.
Bowen Liu, Kaizhi Wang, Dongmei Xiao, Zhan Yu
Various combinatorial optimization NP-hard problems can be reduced to finding the minimizer of an Ising model, which is a discrete mathematical model. It is an intellectual challenge to develop some mathematical tools or algorithms for solving the Ising model. Over the past decades, some continuous approaches or algorithms have been proposed from physical, mathematical or computational views for optimizing the Ising model such as quantum annealing, the coherent Ising machine, simulated annealing, adiabatic Hamiltonian systems, etc.. However, the mathematical principle of these algorithms is far from being understood. In this paper, we reveal the mathematical mechanism of dynamical system algorithms for the Ising model by Morse theory and variational methods. We prove that the dynamical system algorithms can be designed to minimize a continuous function whose local minimum points give all the candidates of the Ising model and the global minimum gives the minimizer of Ising problem. Using this mathematical mechanism, we can easily understand several dynamical system algorithms of the Ising model such as the coherent Ising machine, the Kerr-nonlinear parametric oscillators and the simulated bifurcation algorithm. Furthermore, motivated by the works of C. Conley, we study transit and capture properties of the simulated bifurcation algorithm to explain its convergence by the low energy transit and capture in celestial mechanics. A detailed discussion on $2$-spin and $3$-spin Ising models is presented as application.
Bowen Liu, Yu Chen, Rakesh Chowdary Machineni, Shiyu Liu, Hun-Seok Kim
Learning-based video compression has been extensively studied over the past years, but it still has limitations in adapting to various motion patterns and entropy models. In this paper, we propose multi-mode video compression (MMVC), a block wise mode ensemble deep video compression framework that selects the optimal mode for feature domain prediction adapting to different motion patterns. Proposed multi-modes include ConvLSTM-based feature domain prediction, optical flow conditioned feature domain prediction, and feature propagation to address a wide range of cases from static scenes without apparent motions to dynamic scenes with a moving camera. We partition the feature space into blocks for temporal prediction in spatial block-based representations. For entropy coding, we consider both dense and sparse post-quantization residual blocks, and apply optional run-length coding to sparse residuals to improve the compression rate. In this sense, our method uses a dual-mode entropy coding scheme guided by a binary density map, which offers significant rate reduction surpassing the extra cost of transmitting the binary selection map. We validate our scheme with some of the most popular benchmarking datasets. Compared with state-of-the-art video compression schemes and standard codecs, our method yields better or competitive results measured with PSNR and MS-SSIM.
Bowen Liu, Jiankun Li
DNA storage technology offers new possibilities for addressing massive data storage due to its high storage density, long-term preservation, low maintenance cost, and compact size. To improve the reliability of stored information, base errors and missing storage sequences are challenges that must be faced. Currently, clustering and comparison of sequenced sequences are employed to recover the original sequence information as much as possible. Nonetheless, extracting DNA sequences of different lengths as features leads to the curse of dimensionality, which needs to be overcome. To address this, techniques like PCA, UMAP, and t-SNE are commonly employed to project high-dimensional features into low-dimensional space. Considering that these methods exhibit varying effectiveness in dimensionality reduction when dealing with different datasets, this paper proposes training a multilayer perceptron model to classify input DNA sequence features and adaptively select the most suitable dimensionality reduction method to enhance subsequent clustering results. Through testing on open-source datasets and comparing our approach with various baseline methods, experimental results demonstrate that our model exhibits superior classification performance and significantly improves clustering outcomes. This displays that our approach effectively mitigates the impact of the curse of dimensionality on clustering models.
Bowen Liu, Takasumi Tanabe
THz wireless communications have garnered significant attention due to their unprecedented data rates enabled by the abundant untapped spectrum. However, advanced modulation formats beyond 64-QAM remain largely unexplored, as phase errors introduced during up/down-conversion severely limit system performance. Particularly, OFDM transmission is highly susceptible to aggravated ICI induced by phase noise, undermining the orthogonality of subcarriers. While PLLs and pilot-assisted compensation can mitigate phase errors, excessive pilot overhead compromises spectral efficiency and energy consumption, and white phase noise remains unrecoverable. Therefore, quantifying phase noise tolerance is essential for practical physical layer protocols. Here, we reveal the impact of phase noise in a 64-QAM, 2048-subcarrier OFDM THz transmission system. 3σ-error estimation is proposed to quantify phase noise tolerance, indicating an intuitive EVM threshold of approximately 5%. This threshold further delineates the trade-offs among phase noise levels, SNR requirements, and pilot overhead. Moreover, by benchmarking representative oscillators with distinct phase noise spectra, microring resonators (MRRs) are identified as indispensable enablers for low-pilot-overhead OFDM THz links operating beyond 64-QAM.
Bowen Liu, Zhi Wu, Runquan Xie, Zhanhui Kang, Jia Li
Reinforcement Learning from Verifiable Rewards (RLVR) is bottlenecked by data: existing synthesis pipelines rely on expert-written code or fixed templates, confining growth to instance-level perturbations. We shift the evolvable unit from problem instances to task-family specifications. SSLogic is an agentic meta-synthesis framework in which LLM agents iteratively author and refine executable Generator-Validator pairs inside a closed Generate-Validate-Refine loop, producing families with new rules and difficulty gradients rather than parameter variations of old ones. A Multi-Gate Validation Protocol -- multi-strategy consensus plus Adversarial Blind Review, where independent agents solve each instance by writing and executing code -- filters ill-posed tasks before they enter training. Starting from 400 seed families, two evolution rounds yield 953 families and 21,389 verifiable instances. Three converging comparisons (step-matched, token-matched, and size-controlled on external Enigmata data) consistently show higher training utility of evolved data, with gains of SynLogic +5.2, AIME25 +3.0, and BBH +5.5 on Enigmata. Fine-grained KORBench evaluation reveals selective improvements in logic (+13.2%) and operation (+9.6%), linking structural evolution to downstream gains. Code: https://github.com/AdAstraAbyssoque/Scaling-the-Scaling-Logic
Bowen Liu, Malwane M. A. Ananda
It was observed that the number of cases and deaths for infectious diseases were associated with heavy-tailed power law distributions such as the Pareto distribution. While Pareto distribution was widely used to model the cases and deaths of infectious diseases, a major limitation of Pareto distribution is that it can only fit a given data set beyond a certain threshold. Thus, it can only model part of the data set. Thus, we proposed some novel discrete composite distributions with Pareto tails to fit the real infectious disease data. To provide necessary statistical inference for the tail behavior of the data, we developed a hypothesis testing procedure to test the tail index parameter. COVID-19 reported cases in Singapore and monkeypox reported cases in France were analyzed to evaluate the performance of the newly created distributions. The results from the analysis suggested that the discrete composite distributions could demonstrate competitive performance compared to the commonly used discrete distributions. Furthermore, the analysis of the tail index parameter can provide great insights into preventing and controlling infectious diseases.
Bowen Liu, Wei Liu, Siang Chen, Pengwei Xie, Guijin Wang
The goal of object pose estimation is to visually determine the pose of a specific object in the RGB-D input. Unfortunately, when faced with new categories, both instance-based and category-based methods are unable to deal with unseen objects of unseen categories, which is a challenge for pose estimation. To address this issue, this paper proposes a method to introduce geometric features for pose estimation of point clouds without requiring category information. The method is based only on the patch feature of the point cloud, a geometric feature with rotation invariance. After training without category information, our method achieves as good results as other category-based methods. Our method successfully achieved pose annotation of no category information instances on the CAMERA25 dataset and ModelNet40 dataset.
Bowen Liu, Chunlei Meng, Wei Lin, Hongda Zhang, Ziqing Zhou, Zhongxue Gan, Chun Ouyang
Accurate vascular segmentation is essential for coronary visualization and the diagnosis of coronary heart disease. This task involves the extraction of sparse tree-like vascular branches from the volumetric space. However, existing methods have faced significant challenges due to discontinuous vascular segmentation and missing endpoints. To address this issue, a 3D vision graph neural network framework, named ViG3D-UNet, was introduced. This method integrates 3D graph representation and aggregation within a U-shaped architecture to facilitate continuous vascular segmentation. The ViG3D module captures volumetric vascular connectivity and topology, while the convolutional module extracts fine vascular details. These two branches are combined through channel attention to form the encoder feature. Subsequently, a paperclip-shaped offset decoder minimizes redundant computations in the sparse feature space and restores the feature map size to match the original input dimensions. To evaluate the effectiveness of the proposed approach for continuous vascular segmentation, evaluations were performed on two public datasets, ASOCA and ImageCAS. The segmentation results show that the ViG3D-UNet surpassed competing methods in maintaining vascular segmentation connectivity while achieving high segmentation accuracy. Our code will be available soon.
Bowen Liu, Haoyang Li, Shuning Wang, Shuo Nie, Shanghang Zhang
Out-of-distribution (OOD) generalization in Graph Neural Networks (GNNs) has gained significant attention due to its critical importance in graph-based predictions in real-world scenarios. Existing methods primarily focus on extracting a single causal subgraph from the input graph to achieve generalizable predictions. However, relying on a single subgraph can lead to susceptibility to spurious correlations and is insufficient for learning invariant patterns behind graph data. Moreover, in many real-world applications, such as molecular property prediction, multiple critical subgraphs may influence the target label property. To address these challenges, we propose a novel framework, SubGraph Aggregation (SuGAr), designed to learn a diverse set of subgraphs that are crucial for OOD generalization on graphs. Specifically, SuGAr employs a tailored subgraph sampler and diversity regularizer to extract a diverse set of invariant subgraphs. These invariant subgraphs are then aggregated by averaging their representations, which enriches the subgraph signals and enhances coverage of the underlying causal structures, thereby improving OOD generalization. Extensive experiments on both synthetic and real-world datasets demonstrate that \ours outperforms state-of-the-art methods, achieving up to a 24% improvement in OOD generalization on graphs. To the best of our knowledge, this is the first work to study graph OOD generalization by learning multiple invariant subgraphs. code: https://github.com/Nanolbw/SuGAr
Bowen Liu, Mao Sheng
In this paper we complete the study of the Lan-Sheng-Zuo conjecture proposed in arXiv:1210.8280 for the curve case. Precisely, we prove that every semistable Higgs bundle is strongly semistable for curves of genus $g\leq 1$, and over any curves of genus $g\ge2$ construct explicit examples of semistable Higgs bundles of arbitrary big rank (the first example is $p=2,r=3$) which are not strongly semistable. These results are complementary to the strongly semistability theorem of Lan-Sheng-Yang-Zuo and Langer for semistable Higgs bundles of small rank.
Bowen Liu, Pengyue Jia, Wanyu Wang, Derong Xu, Jiawei Cheng, Jiancheng Dong, Xiao Han, Zimo Zhao, Chao Zhang, Bowen Yu, Fangyu Hong, Xiangyu Zhao
The primary objective of cross-view UAV geolocalization is to identify the exact spatial coordinates of drone-captured imagery by aligning it with extensive, geo-referenced satellite databases. Current approaches typically extract features independently from each perspective and rely on basic heuristics to compute similarity, thereby failing to explicitly capture the essential interactions between different views. To address this limitation, we introduce a novel, plug-and-play ranking architecture designed to explicitly perform joint relational modeling for improved UAV-to-satellite image matching. By harnessing the capabilities of a Large Vision-Language Model (LVLM), our framework effectively learns the deep visual-semantic correlations linking UAV and satellite imagery. Furthermore, we present a novel relational-aware loss function to optimize the training phase. By employing soft labels, this loss provides fine-grained supervision that avoids overly penalizing near-positive matches, ultimately boosting both the model's discriminative power and training stability. Comprehensive evaluations across various baseline architectures and standard benchmarks reveal that the proposed method substantially boosts the retrieval accuracy of existing models, yielding superior performance even under highly demanding conditions.
Bowen Liu, Malwane M. A. Ananda
Exponentiated models have been widely used in modeling various types of data such as survival data and insurance claims data. However, the exponentiated composite distribution models have not been explored yet. In this paper, we introduce an improvement of the one-parameter Inverse Gamma-Pareto composite model by exponentiating the random variable associated with the one-parameter Inverse Gamma-Pareto composite distribution function. The goodness-of-fit of the exponentiated Inverse Gamma-Pareto was assessed using three different insurance data sets. The two-parameter exponentiated Inverse Gamma-Pareto model outperforms the one-parameter Inverse Gamma-Pareto model in terms of goodness-of-fit measures for all datasets. In addition, the proposed exponentiated composite Inverse Gamma-Pareto model provides a very good fit with some well-known insurance datasets.
Yuyao Wang, Bowen Liu, Jianheng Tang, Nuo Chen, Yuhan Li, Qifan Zhang, Chenyi Zi, Chen Zhang, Jia Li
Reasoning Large Language Models (RLLMs) have recently achieved remarkable progress on complex reasoning tasks, largely enabled by their long chain-of-thought (Long CoT) capabilities. However, developing these Long CoT behaviors relies heavily on post-training with high-quality datasets, which are typically costly and human-curated (e.g., mathematics and code), leaving scalable alternatives unexplored. In this work, we introduce NP-hard (NPH) graph problems as a novel synthetic training corpus, as they inherently require deep reasoning, extensive exploration, and reflective strategies, which are the core characteristics of Long CoT reasoning. Building on this insight, we develop a two-stage post-training framework: (i) Long-CoT Supervised Fine-Tuning (SFT) on rejection-sampled NPH graph instances, which substantially enhances reasoning depth, and (ii) Reinforcement Learning (RL) with a fine-grained reward design, which sharpens reasoning efficiency. The resulting NPG-Muse-series models exhibit substantially enhanced Long CoT reasoning capabilities, achieving consistent gains across mathematics, coding, logical, and graph reasoning benchmarks. NPG-Muse-7B even surpasses QwQ-32B on NPH graph problems in both accuracy and reasoning efficiency. These results position NPH graph problems as an effective and scalable resource for advancing Long CoT reasoning in LLM post-training. Our implementation is available at https://github.com/littlewyy/NPG-Muse.
Bowen Liu, Malwane M. A. Ananda
In this paper, we propose a new class of distributions by exponentiating the random variables associated with the probability density functions of composite distributions. We also derive some mathematical properties of this new class of distributions including the moments and the limited moments. Specifically, two special models in this family are discussed. Two real data sets were chosen to assess the performance of these two special exponentiated composite models. When fitting to these two data sets, theses two special exponentiated composite distributions demonstrate significantly better performance compared to the original composite distributions.
Bowen Liu, Malwane M. A. Ananda
Modeling excess remains to be an important topic in insurance data modeling. Among the alternatives of modeling excess, the Peaks Over Threshold (POT) framework with Generalized Pareto distribution (GPD) is regarded as an efficient approach due to its flexibility. However, the selection of an appropriate threshold for such framework is a major difficulty. To address such difficulty, we applied several accumulation tests along with Anderson-Darling test to determine an optimal threshold. Based on the selected thresholds, the fitted GPD with the estimated quantiles can be found. We applied the procedure to the well-known Norwegian Fire Insurance data and constructed the confidence intervals for the Value-at-Risks (VaR). The accumulation test approach provides satisfactory performance in modeling the high quantiles of Norwegian Fire Insurance data compared to the previous graphical methods.
Bowen Liu, Pawel Szalachowski, Siwei Sun
The recent development of payment channels and their extensions (e.g., state channels) provides a promising scalability solution for blockchains which allows untrusting parties to transact off-chain and resolve potential disputes via on-chain smart contracts. To protect participants who have no constant access to the blockchain, a watching service named as watchtower is proposed -- a third-party entity obligated to monitor channel states (on behalf of the participants) and correct them on-chain if necessary. Unfortunately, currently proposed watchtower schemes suffer from multiple security and efficiency drawbacks. In this paper, we explore the design space behind watchtowers. We propose a novel watching service named as fail-safe watchtowers. In contrast to prior proposed watching services, our fail-safe watchtower does not watch on-chain smart contracts constantly. Instead, it only sends a single on-chain message periodically confirming or denying the final states of channels being closed. Our watchtowers can easily handle a large number of channels, are privacy-preserving, and fail-safe tolerating multiple attack vectors. Furthermore, we show that watchtowers (in general) may be an option economically unjustified for multiple payment scenarios and we introduce a simple, yet powerful concept of short-lived assertions which can mitigate misbehaving parties in these scenarios.
Bowen Liu, Siwei Sun, Pawel Szalachowski
Although blockchain-based smart contracts promise a ``trustless'' way of enforcing agreements even with monetary consequences, they suffer from multiple security issues. Many of these issues could be mitigated via an effective access control system, however, its realization is challenging due to the properties of current blockchain platforms (like lack of privacy, costly on-chain resources, or latency). To address this problem, we propose the SMACS framework, where updatable and sophisticated Access Control Rules (ACRs)} for smart contracts can be realized with low cost. SMACS shifts the burden of expensive ACRs validation and management operations to an off-chain infrastructure, while implementing on-chain only lightweight token-based access control. SMACS is flexible and in addition to simple access control lists can easily implement rules enhancing the runtime security of smart contracts. With dedicated ACRs backed by vulnerability-detection tools, SMACS can protect vulnerable contracts after deployment. We fully implement SMACS and evaluate it.
Bowen Liu
In this paper, we study the linear stability of the elliptic rhombus solutions, which are the Keplerian homographic solution with the rhombus central configurations in the classical planar four-body problems. Using $ω$-Maslov index theory and trace formula, we prove the linear instability of elliptic rhombus solutions if the shape parameter $u$ and the eccentricity of the elliptic orbit $e$ satisfy $(u,e) \in (1/\sqrt{3}, u_2)\times [0, \hat{f}(\frac{27}{4})^{-1/2})\cup (u_2, 1/u_2) \times [0,1)\cup ( 1/u_2, \sqrt{3})\times [0, \hat{f}(\frac{27}{4})^{-1/2})$ where $u_2\approx 0.6633$ and $\hat{f}(\frac{27}{4})^{-1/2} \approx 0.4454$. Motivated on numerical results of the linear stability to the elliptic Lagrangian solutions in [R. Martínez, A. Samà, and C. Simó, J. Diff. Equa., 226(2006): 619--651.], we further analytically prove the linear instability of elliptic rhombus solutions for $(u,e)\in (1/\sqrt{3}, \sqrt{3}) \times [0,1)$.