Hao He, Connor Bottrell, Christine Wilson, Jorge Moreno, Blakesley Burkhart, Christopher C. Hayward, Lars Hernquist, Angela Twum
Jan 30, 2023·astro-ph.GA·PDF We employ the Feedback In Realistic Environments (FIRE-2) physics model to study how the properties of giant molecular clouds (GMCs) evolve during galaxy mergers. We conduct a pixel-by-pixel analysis of molecular gas properties in both the simulated control galaxies and galaxy major mergers. The simulated GMC-pixels in the control galaxies follow a similar trend in a diagram of velocity dispersion ($σ_v$) versus gas surface density ($Σ_{\mathrm{mol}}$) to the one observed in local spiral galaxies in the Physics at High Angular resolution in Nearby GalaxieS (PHANGS) survey. For GMC-pixels in simulated mergers, we see a significant increase of factor of 5 - 10 in both $Σ_{\mathrm{mol}}$ and $σ_v$, which puts these pixels above the trend of PHANGS galaxies in the $σ_v$ vs $Σ_{\mathrm{mol}}$ diagram. This deviation may indicate that GMCs in the simulated mergers are much less gravitationally bound compared with simulated control galaxies with virial parameter ($α_{\mathrm{vir}}$) reaching 10 - 100. Furthermore, we find that the increase in $α_{\mathrm{vir}}$ happens at the same time as the increase in global star formation rate (SFR), which suggests stellar feedback is responsible for dispersing the gas. We also find that the gas depletion time is significantly lower for high $α_{\mathrm{vir}}$ GMCs during a starburst event. This is in contrast to the simple physical picture that low $α_{\mathrm{vir}}$ GMCs are easier to collapse and form stars on shorter depletion times. This might suggest that some other physical mechanisms besides self-gravity are helping the GMCs in starbursting mergers collapse and form stars.
Yixun Liang, Hao He, Ying-cong Chen
Generalizable neural surface reconstruction techniques have attracted great attention in recent years. However, they encounter limitations of low confidence depth distribution and inaccurate surface reasoning due to the oversimplified volume rendering process employed. In this paper, we present Reconstruction TRansformer (ReTR), a novel framework that leverages the transformer architecture to redesign the rendering process, enabling complex render interaction modeling. It introduces a learnable $\textit{meta-ray token}$ and utilizes the cross-attention mechanism to simulate the interaction of rendering process with sampled points and render the observed color. Meanwhile, by operating within a high-dimensional feature space rather than the color space, ReTR mitigates sensitivity to projected colors in source views. Such improvements result in accurate surface assessment with high confidence. We demonstrate the effectiveness of our approach on various datasets, showcasing how our method outperforms the current state-of-the-art approaches in terms of reconstruction quality and generalization ability. $\textit{Our code is available at }$ https://github.com/YixunLiang/ReTR.
Yintao Wang, Hao He, Shaoyang Wang, Yaohui Liu, Minglie Hu, Youjia Cao, Chingyue Wang
Feb 16, 2015·q-bio.SC·PDF Mitochondrial research is important to ageing, apoptosis, and mitochondrial diseases. In previous works, mitochondria are usually stimulated indirectly by proapoptotic drugs to study mitochondrial development, which is in lack of controllability, or spatial and temporal resolution. These chemicals or even gene techniques regulating mitochondrial dynamics may also activate other inter- or intra-cellular processes simultaneously. Here we demonstrate a photostimulation method on single-mitochondrion level by tightly-focused femtosecond laser that can precisely activate restorable fragmentation of mitochondria which soon recover their original tubular structure after tens of seconds. In this process, series of mitochondrial reactive oxygen species (mROS) flashes are observed and found very critical to mitochondrial fragmentation. Meanwhile, transient openings of mitochondrial permeability transition pores (mPTP), suggested by oscillations of mitochondrial membrane potential, contribute to the scavenging of redundant mROS and recovery of fragmented mitochondria. Those results demonstrate photostimulation as an active, precise and controllable method for the study of mitochondrial oxidative and morphological dynamics or related fields.
Hao He, Bo Xin, David Wipf
The iterations of many first-order algorithms, when applied to minimizing common regularized regression functions, often resemble neural network layers with pre-specified weights. This observation has prompted the development of learning-based approaches that purport to replace these iterations with enhanced surrogates forged as DNN models from available training data. For example, important NP-hard sparse estimation problems have recently benefitted from this genre of upgrade, with simple feedforward or recurrent networks ousting proximal gradient-based iterations. Analogously, this paper demonstrates that more powerful Bayesian algorithms for promoting sparsity, which rely on complex multi-loop majorization-minimization techniques, mirror the structure of more sophisticated long short-term memory (LSTM) networks, or alternative gated feedback networks previously designed for sequence prediction. As part of this development, we examine the parallels between latent variable trajectories operating across multiple time-scales during optimization, and the activations within deep network structures designed to adaptively model such characteristic sequences. The resulting insights lead to a novel sparse estimation system that, when granted training data, can estimate optimal solutions efficiently in regimes where other algorithms fail, including practical direction-of-arrival (DOA) and 3D geometry recovery problems. The underlying principles we expose are also suggestive of a learning process for a richer class of multi-loop algorithms in other domains.
Hao He, Kaiwen Zha, Dina Katabi
Indiscriminate data poisoning attacks are quite effective against supervised learning. However, not much is known about their impact on unsupervised contrastive learning (CL). This paper is the first to consider indiscriminate poisoning attacks of contrastive learning. We propose Contrastive Poisoning (CP), the first effective such attack on CL. We empirically show that Contrastive Poisoning, not only drastically reduces the performance of CL algorithms, but also attacks supervised learning models, making it the most generalizable indiscriminate poisoning attack. We also show that CL algorithms with a momentum encoder are more robust to indiscriminate poisoning, and propose a new countermeasure based on matrix completion. Code is available at: https://github.com/kaiwenzha/contrastive-poisoning.
Zihao Xu, Hao He, Guang-He Lee, Yuyang Wang, Hao Wang
Existing domain adaptation methods tend to treat every domain equally and align them all perfectly. Such uniform alignment ignores topological structures among different domains; therefore it may be beneficial for nearby domains, but not necessarily for distant domains. In this work, we relax such uniform alignment by using a domain graph to encode domain adjacency, e.g., a graph of states in the US with each state as a domain and each edge indicating adjacency, thereby allowing domains to align flexibly based on the graph structure. We generalize the existing adversarial learning framework with a novel graph discriminator using encoding-conditioned graph embeddings. Theoretical analysis shows that at equilibrium, our method recovers classic domain adaptation when the graph is a clique, and achieves non-trivial alignment for other types of graphs. Empirical results show that our approach successfully generalizes uniform alignment, naturally incorporates domain information represented by graphs, and improves upon existing domain adaptation methods on both synthetic and real-world datasets. Code will soon be available at https://github.com/Wang-ML-Lab/GRDA.
Yixun Liang, Hao He, Shishi Xiao, Hao Lu, Yingcong Chen
Point cloud segmentation is a fundamental task in 3D vision that serves a wide range of applications. Although great progresses have been made these years, its practical usability is still limited by the availability of training data. Existing approaches cannot make full use of multiple datasets on hand due to the label mismatch among different datasets. In this paper, we propose a principled approach that supports learning from heterogeneous datasets with different label sets. Our idea is to utilize a pre-trained language model to embed discrete labels to a continuous latent space with the help of their label names. This unifies all labels of different datasets, so that joint training is doable. Meanwhile, classifying points in the continuous 3D space by their vocabulary tokens significantly increase the generalization ability of the model in comparison with existing approaches that have fixed decoder architecture. Besides, we also integrate prompt learning in our framework to alleviate data shifts among different data sources. Extensive experiments demonstrate that our model outperforms the state-of-the-art by a large margin.
He Hao, Borhan M. Sanandaji, Kameshwar Poolla, Tyrone L. Vincent
Residential Thermostatically Controlled Loads (TCLs) such as Air Conditioners (ACs), heat pumps, water heaters, and refrigerators have an enormous thermal storage potential for providing regulation reserve to the grid. In this paper, we study the potential resource and economic analysis of TCLs providing frequency regulation service. In particular, we show that the potential resource of TCLs in California is more than enough for both current and predicted near-future regulation requirements for the California power system. Moreover, we estimate the cost and revenue of TCLs, discuss the qualification requirements, recommended policy changes, and participation incentive methods, and compare TCLs with other energy storage technologies. We show that TCLs are potentially more cost-effective than other energy storage technologies such as flywheels, Li-ion, advanced lead acid, and Zinc Bromide batteries.
Xingwei Gao, Hao He, Weng W. Chow, Alexander Cerjan, Chia Wei Hsu
Recent studies have demonstrated that a laser can self-generate frequency combs when tuned near an exceptional point (EP), where two cavity modes coalesce. These EP combs induce periodic modulation of the population inversion in the gain medium, and their repetition rate is independent of the laser cavity's free spectral range. In this work, we perform a stability analysis that reveals two notable properties of EP combs, bi-stability and a period-doubling cascade. The period-doubling cascade enables halving of the repetition rate while maintaining the comb's total bandwidth, presenting opportunities for the design of highly compact frequency comb generators.
Caoyun Fan, Wenqing Chen, Jidong Tian, Yitian Li, Hao He, Yaohui Jin
When modeling related tasks in computer vision, Multi-Task Learning (MTL) can outperform Single-Task Learning (STL) due to its ability to capture intrinsic relatedness among tasks. However, MTL may encounter the insufficient training problem, i.e., some tasks in MTL may encounter non-optimal situation compared with STL. A series of studies point out that too much gradient noise would lead to performance degradation in STL, however, in the MTL scenario, Inter-Task Gradient Noise (ITGN) is an additional source of gradient noise for each task, which can also affect the optimization process. In this paper, we point out ITGN as a key factor leading to the insufficient training problem. We define the Gradient-to-Noise Ratio (GNR) to measure the relative magnitude of gradient noise and design the MaxGNR algorithm to alleviate the ITGN interference of each task by maximizing the GNR of each task. We carefully evaluate our MaxGNR algorithm on two standard image MTL datasets: NYUv2 and Cityscapes. The results show that our algorithm outperforms the baselines under identical experimental conditions.
Hao He, C. D. Wilson, Kazimierz Sliwa, Daisuke Iono, Toshiki Saito
Jun 22, 2020·astro-ph.GA·PDF We present new high resolution $^{12}$CO $J$=1-0, $J$=2-1, and $^{13}$CO $J$=1-0 maps of the early stage merger Arp 240 (NGC5257/8) obtained with the Atacama Large Millimeter/submillimeter Array (ALMA). Simulations in the literature suggest that the merger has just completed its first passage; however, we find that this system has a lower global gas fraction but a higher star formation efficiency compared to typical close galaxy pairs, which suggests that this system may already be in an advanced merger stage. We combine the ALMA data with $^{12}$CO $J$=3-2 observations from the Submillimeter Array and carry out RADEX modeling on several different regions. Both the RADEX modeling and a local thermal equilibrium (LTE) analysis show that the regions are most likely to have a CO-to-H$_2$ conversion factor $α_{\mathrm{CO}}$ close to or perhaps even smaller than the typical value for (ultra-)luminous infrared galaxies. Using 33 GHz data from the Very Large Array to measure the star formation rate, we find that most star forming regions have molecular gas depletion times of less than 100 Myr. We calculated the star formation efficiency (SFE) per free-fall time for different regions and find some regions appear to have values greater than 100%. We find these regions generally show evidence for young massive clusters (YMCs). After exploring various factors, we argue that this is mainly due to the fact that radio continuum emission in those regions is dominated by that from YMCs, which results in an overestimate of the SFE per free-fall time.
Hao He, Cheng Guo, Meng Xiao
Space-time wave packets can propagate invariantly in free space with arbitrary group velocity thanks to the spatio-temporal correlation. Here it is proved that the space-time wave packets are stable in dispersive media as well and free from the spread in time caused by material dispersion. Furthermore, the law of anomalous refraction for space-time wave packets is generalized to the weakly dispersive situation. These results reveal new potential of space-time wave packets for the applications in real dispersive media.
Tianyi Liu, Zihao Xu, Hao He, Guang-Yuan Hao, Guang-He Lee, Hao Wang
Domain adaptation aims to mitigate distribution shifts among different domains. However, traditional formulations are mostly limited to categorical domains, greatly simplifying nuanced domain relationships in the real world. In this work, we tackle a generalization with taxonomy-structured domains, which formalizes domains with nested, hierarchical similarity structures such as animal species and product catalogs. We build on the classic adversarial framework and introduce a novel taxonomist, which competes with the adversarial discriminator to preserve the taxonomy information. The equilibrium recovers the classic adversarial domain adaptation's solution if given a non-informative domain taxonomy (e.g., a flat taxonomy where all leaf nodes connect to the root node) while yielding non-trivial results with other taxonomies. Empirically, our method achieves state-of-the-art performance on both synthetic and real-world datasets with successful adaptation. Code is available at https://github.com/Wang-ML-Lab/TSDA.
Hao He, William R. F. Dent, Christine Wilson
Nov 22, 2022·astro-ph.IM·PDF The ALMA observatory is now putting more focus on high-frequency observations (frequencies from 275-950 GHz). However, high-frequency observations often suffer from rapid variations in atmospheric opacity that directly affect the system temperature $T_{sys}$. Current observations perform discrete atmospheric calibrations (Atm-cals) every few minutes, with typically 10-20 occurring per hour for high frequency observation and each taking 30-40 seconds. In order to obtain more accurate flux measurements and reduce the number of atmospheric calibrations (Atm-cals), a new method to monitor $T_{sys}$ continuously is proposed using existing data in the measurement set. In this work, we demonstrate the viability of using water vapor radiometer (WVR) data to track the $T_{sys}$ continuously. We find a tight linear correlation between $T_{sys}$ measured using the traditional method and $T_{sys}$ extrapolated based on WVR data with scatter of 0.5-3%. Although the exact form of the linear relation varies among different data sets and spectral windows, we can use a small number of discrete $T_{sys}$ measurements to fit the linear relation and use this heuristic relationship to derive $T_{sys}$ every 10 seconds. Furthermore, we successfully reproduce the observed correlation using atmospheric transmission at microwave (ATM) modeling and demonstrate the viability of a more general method to directly derive the $T_{sys}$ from the modeling. We apply the semi-continuous $T_{sys}$ from heuristic fitting on a few data sets from Band 7 to Band 10 and compare the flux measured using these methods. We find the discrete and continuous $T_{sys}$ methods give us consistent flux measurements with differences up to 5%. Furthermore, this method has significantly reduced the flux uncertainty due to $T_{sys}$ variability for one dataset, which has large precipitable water vapor (PWV) fluctuation, from 10% to 0.7%.
Nathan Brunetti, Christine D. Wilson, Hao He, Jiayi Sun, Adam K. Leroy, Erik Rosolowsky, Ashley Bemis, Frank Bigiel, Brent Groves, Toshiki Saito, Eva Schinnerer
We present observations of the central 9 kpc of the Antennae merger (NGC 4038/9) at 55 pc resolution in the CO 2-1 line obtained with the Atacama Large Millimeter/submillimeter Array (ALMA). We use a pixel-based analysis to compare the gas properties in the Antennae to those in 70 nearby spiral galaxies from the PHANGS-ALMA survey, as well as the merger and nearest luminous infrared galaxy NGC 3256. Compared to PHANGS galaxies at matched spatial resolution, the molecular gas in the Antennae exhibits some of the highest surface densities, velocity dispersions, peak brightness temperatures, and turbulent pressures. However, the virial parameters in the Antennae are consistent with many of the PHANGS galaxies. NGC 3256 has similar gas surface densities but higher nuclear velocity dispersions than the Antennae, as well as higher system-wide peak brightness temperatures and virial parameters. NGC 3256 is at a later stage in the merging process than the Antennae, which may result in more intense merger-driven gas flows that could drive up the turbulence in the gas. The high virial parameters in NGC 3256 may indicate that this increased turbulence is suppressing future star formation as NGC 3256 moves out of the starburst phase. In comparison, the relatively normal virial parameters in the Antennae may imply that it is about to undergo a new burst of star formation.
Runzhi He, Hao He, Yuxia Zhang, Minghui Zhou
Dependency management bots automatically open pull requests to update software dependencies on behalf of developers. Early research shows that developers are suspicious of updates performed by dependency management bots and feel tired of overwhelming notifications from these bots. Despite this, dependency management bots are becoming increasingly popular. Such contrast motivates us to investigate Dependabot, currently the most visible bot on GitHub, to reveal the effectiveness and limitations of state-of-art dependency management bots. We use exploratory data analysis and a developer survey to evaluate the effectiveness of Dependabot in keeping dependencies up-to-date, interacting with developers, reducing update suspicion, and reducing notification fatigue. We obtain mixed findings. On the positive side, projects do reduce technical lag after Dependabot adoption and developers are highly receptive to its pull requests. On the negative side, its compatibility scores are too scarce to be effective in reducing update suspicion; developers tend to configure Dependabot toward reducing the number of notifications; and 11.3% of projects have deprecated Dependabot in favor of other alternatives. The survey confirms our findings and provides insights into the key missing features of Dependabot. Based on our findings, we derive and summarize the key characteristics of an ideal dependency management bot which can be grouped into four dimensions: configurability, autonomy, transparency, and self-adaptability.
Hao He, Xingwei Gao, Alexander Cerjan, Chia Wei Hsu
One of the key features of lasers operating near exceptional points (EPs) is that the gain medium can support an oscillating population inversion above a pump threshold, leading to self-modulated laser dynamics. This unusual behavior opens up new possibilities for frequency comb generation and temporal modulation. However, the dynamic population inversion couples signals with different frequencies and thus cannot be captured by conventional temporal coupled-mode theory (TCMT) based on static saturable gain. In this paper, we develop a perturbative coupled-mode analysis framework to capture the spatial-temporal dynamics of near-EP lasers. By decomposing discrete frequency generation into multiple excitations of resonant modes, our analysis establishes a minimal physical model that translates the local distribution of dynamic population-inversion into a resonant modal interpretation of laser gain. This work enables the exploration of unique properties in this self-time-modulated systems, such as time-varying scattering and non-reciprocal transmission.
Hao He, Bogdan Vasilescu, Christian Kästner
Recent high-profile incidents in open-source software have greatly raised practitioner attention on software supply chain attacks. To guard against potential malicious package updates, security practitioners advocate pinning dependency to specific versions rather than floating in version ranges. However, it remains controversial whether pinning carries a meaningful security benefit that outweighs the cost of maintaining outdated and possibly vulnerable dependencies. In this paper, we quantify, through counterfactual analysis and simulations, the security and maintenance impact of version constraints in the npm ecosystem. By simulating dependency resolutions over historical time points, we find that pinning direct dependencies not only (as expected) increases the cost of maintaining vulnerable and outdated dependencies, but also (surprisingly) even increases the risk of exposure to malicious package updates in larger dependency graphs due to the specifics of npm's dependency resolution mechanism. Finally, we explore collective pinning strategies to secure the ecosystem against supply chain attacks, suggesting specific changes to npm to enable such interventions. Our study provides guidance for practitioners and tool designers to manage their supply chains more securely.
Hao He, Pramod K. Varshney
In this paper, we consider a distributed detection problem for a censoring sensor network where each sensor's communication rate is significantly reduced by transmitting only "informative" observations to the Fusion Center (FC), and censoring those deemed "uninformative". While the independence of data from censoring sensors is often assumed in previous research, we explore spatial dependence among observations. Our focus is on designing the fusion rule under the Neyman-Pearson (NP) framework that takes into account the spatial dependence among observations. Two transmission scenarios are considered, one where uncensored observations are transmitted directly to the FC and second where they are first quantized and then transmitted to further improve transmission efficiency. Copula-based Generalized Likelihood Ratio Test (GLRT) for censored data is proposed with both continuous and discrete messages received at the FC corresponding to different transmission strategies. We address the computational issues of the copula-based GLRTs involving multidimensional integrals by presenting more efficient fusion rules, based on the key idea of injecting controlled noise at the FC before fusion. Although, the signal-to-noise ratio (SNR) is reduced by introducing controlled noise at the receiver, simulation results demonstrate that the resulting noise-aided fusion approach based on adding artificial noise performs very closely to the exact copula-based GLRTs. Copula-based GLRTs and their noise-aided counterparts by exploiting the spatial dependence greatly improve detection performance compared with the fusion rule under independence assumption.
Hao He, Courtney Miller, Shyam Agarwal, Christian Kästner, Bogdan Vasilescu
Large language models (LLMs) have demonstrated the promise to revolutionize the field of software engineering. Among other things, LLM agents are rapidly gaining momentum in software development, with practitioners reporting a multifold increase in productivity after adoption. Yet, empirical evidence is lacking around these claims. In this paper, we estimate the causal effect of adopting a widely popular LLM agent assistant, namely Cursor, on development velocity and software quality. The estimation is enabled by a state-of-the-art difference-in-differences design comparing Cursor-adopting GitHub projects with a matched control group of similar GitHub projects that do not use Cursor. We find that the adoption of Cursor leads to a statistically significant, large, but transient increase in project-level development velocity, along with a substantial and persistent increase in static analysis warnings and code complexity. Further panel generalized-method-of-moments estimation reveals that increases in static analysis warnings and code complexity are major factors driving long-term velocity slowdown. Our study identifies quality assurance as a major bottleneck for early Cursor adopters and calls for it to be a first-class citizen in the design of agentic AI coding tools and AI-driven workflows.