
Alesia Chernikova
I am on the academic and industry job market! I will be grateful for any recommendations and job opportunities!
Postdoctoral Research Associate a.chernikova@northeastern.edu CV
Hi, I am a Postdoctoral Research Associate at the Network Science Institute at Northeastern University.
I work under the supervision of Professor
Tina Eliassi-Rad
and in collaboration with Professor
Dmitri Krioukov.
My research develops foundation that explains how AI models operate and principled methods that ensure their reliability as capabilities continue to accelerate.
Currently I am working on the mechanistic interpretability of Large Language Models. I am developing a fully automated framework for circuit extraction in LLMs using Cross-Layer Transcoders and attribution graphs, combining importance-weighted pruning with a flow-based method to identify compact computational circuits for diagnosing alignment failures and enabling targeted interventions to mitigate misalignment.
I also developed a unified framework for uncertainty quantification, certified adversarial robustness, and generalization in Message-Passing Neural Networks to reason about how they operate under the uncertainty in node features.
Anoter direction of my research focuses on grounding AI architecture design in the geometric principles and hierarchical organization of natural neural networks and offers a principled path toward understanding how the resulting structural properties are related to learning efficiency and generalization.
I completed my Ph.D. in Computer Science at Northeastern University, where I was advised by Professor
Alina Oprea.
I was affiliated with the
NDS2 lab
and was a part of the Cybersecurity and Privacy Institute. My researcg focused on adversarial machine learning and cybernetwork resilience. I developed optimization-based frameworks for feasible evasion attacks under real-world constraints, with applications to cybersecurity. This work informed deployments at the Army Research Laboratory and Perspecta Labs. Our collaboration with Toyota Motor North America R&D Institute resulted in the first demonstrated evasion attacks against deep neural networks used for autonomous vehicle steering prediction. In parallel, I introduced a novel epidemiological model for malware propagation and graph-based defense strategies evaluated on enterprise-scale communication networks.
I spent two summers working as Applied Research Scientist at Amazon Web Services in Amazon Detective team. I created a scalable algorithm for tracing the activity in the AWS cloud represented as a heterogeneous graph to allow further research based on AWS cloud activity data. I also developed the methodology for lateral movement detection in the AWS cloud environment using Bayesian statistics and network science perspectives.
I received my BS degree in Applied Mathematics and Computer Science from
Belarusian State University, where I was affiliated with the Mathematical Modeling and Data Analysis Department
under the supervision of Professor
Vladimir Malugin.
My research focus included the design of hedging algorithms based on derivative contracts. Additionally, I was a
part of the Research Institute of Applied Mathematics and Information Technology Problems, where I participated
in the project for credit rankings estimation and evaluation of national enterprises using mathematical,
statistical, and econometric methods and models.
After my undegraduate studies, I worked as a senior software engineer at IBA Group in Minsk, Belarus. I participated in the development of a large-scale IBM GSAR web portal and maintained its performance.
My other interests include hiking and yoga, film photography and visual art, music and soundscapes.
As modern AI systems, including LLMs, grow in scale and complexity, mechanistic interpretability has become essential to assess whether their internal computations are predictable, reliable, and aligned with human intent. Recent work has shown that sparse coding–based techniques, including sparse autoencoders, transcoders, and crosscoders, can isolate interpretable features in model activations, with many such features corresponding to meaningful semantic concepts. Anthropic’s introduction of Cross-Layer Transcoders (CLTs) and the Circuit Tracer tool represents a significant step forward: CLTs construct attribution graphs that trace how information propagates through an LLM for a given input, revealing groups of interacting neurons that plausibly form computational circuits. However, translating these dense attribution graphs into a concise, mechanistic explanation still relies heavily on human inspection. My research develops a fully automated framework to extract such circuits from CLT-generated attribution graphs. The approach begins with pruning based on importance weights, removing weak or irrelevant edges while preserving the dominant computational backbone. On this reduced graph, I apply a flow-based extraction method grounded in an information-flow analogy to electrical networks, which identifies pathways between selected source and target neurons. The resulting subgraph provides a candidate circuit that can be systematically interpreted by mapping its components to linguistic or semantic functions. This framework transforms a previously manual, ad hoc process into a scalable, principled methodology for uncovering the computational structure within LLMs, offering a path toward more transparent and trustworthy AI systems.
Existing theoretical guarantees for message passing neural networks (MPNNs) assume deterministic node features. We address the more realistic situation where noise or finite measurement precision lead to uncertainties in the values of node features. First, we quantify uncertainty by propagating the moments of node-feature distributions through the MPNN architecture. To propagate the moments through activation functions, we use second-order Taylor expansion and pseudo-Taylor polynomial expansion (PTPE). We use the resulting node embedding distributions to analytically generate probabilistic adversarial robustness certificates for node classification tasks against the L2-bounded perturbations of the node features. Second, we model node features as multivariate random variables and propose a Wasserstein distance-based pseudo-metric, the Feature Convolution Distance FCDp, corresponding to the discriminative power of MPNNs at the node level. We show that MPNNs are global Lipschitz continuous functions with respect to the introduced pseudo-metric FCDp. Using the covering number of the resulting pseudometric space (which is a subset of the Wasserstein space), we derive generalization bounds for MPNNs with uncertainties in the node features. Together, these two complementary approaches --- moment propagation for probabilistic robustness and the FCDp on the subset of the Wasserstein space for generalization --- establish a unified theoretical framework that comprehensively addresses MPNN reliability under node feature uncertainty.
Existing theoretical guarantees for message passing neural networks (MPNNs) assume deterministic node features, whereas in this work we address the more realistic setting where inherent noise or finite measurement precision leads to uncertainty in node features. We assume node features are multivariate Gaussian distributions and propagate their first and second moments through the MPNN architecture. We employ Polynomial Chaos Expansion to approximate nonlinearities, and use the resulting node embedding distributions to analytically produce probabilistic node-wise robustness certificates against L2-bounded node feature perturbations. Moreover, we model node features as multivariate random variables and introduce Feature Convolution Distance, FCDp, a Wasserstein distance-based pseudometric that matches the discriminative power of node-level MPNNs. We show that MPNNs are globally Lipschitz continuous functions with respect to FCDp. Our framework subsumes the deterministic case via Dirac measures and provides a foundation for reasoning about algorithmic stability in MPNNs with uncertainty in node features.
Modern AI faces a critical efficiency crisis: DNNs require billions of parameters and consume enormous energy to achieve state-of-the-art performance, while remaining opaque ``black boxes'' whose decision-making processes cannot be understood or trusted. By grounding DNN architecture design in the geometric principles and hierarchical organization of natural neural networks, this project offers a principled path toward understanding how the resulting structural properties are related to learning efficiency, robustness, and generalization. In our work, we introduce a novel DNN architecture that exploits the geometry of derandomized hyperbolic graphs. Concretely, we begin by generating a derandomized hyperbolic graph that encodes a hierarchical organization among neurons. We then transform this graph into a directed acyclic graph (DAG) that serves as the backbone of our neural architecture. Hidden variables assigned to each node in the graph impose a natural hierarchical order, which we use to direct edges and layer the network in a feed-forward manner. This construction ensures acyclicity and establishes an intrinsic layered structure dictated by the graph geometry. Each neuron (graph node) is assigned to a layer based on the longest path from the input nodes, thereby limiting extraneous connections and redundant message passing. The resulting architecture is a sparse, directed network that retains the expressivity of a dense network, but with far fewer parameters. By reducing overlapping or unimportant paths, the architecture is not only more efficient but also more interpretable, since the hierarchical organization of connections aligns with clearer functional roles for each layer. This work represents a novel synthesis of ideas from statistical physics and deep learning, opening the door to principled new ways to design DNNs.
text
text
text