Hi!

Your personalized paper recommendations for 24 to 28 November, 2025.
๐ŸŽฏ Top Personalized Recommendations
Abstract
Modern cloud databases present scaling as a binary decision: scale-out by adding nodes or scale-up by increasing per-node resources. This one-dimensional view is limiting because database performance, cost, and coordination overhead emerge from the joint interaction of horizontal elasticity and per-node CPU, memory, network bandwidth, and storage IOPS. As a result, systems often overreact to load spikes, underreact to memory pressure, or oscillate between suboptimal states. We introduce the Scaling Plane, a two-dimensional model in which each distributed database configuration is represented as a point (H, V), with H denoting node count and V a vector of resources. Over this plane, we define smooth approximations of latency, throughput, coordination overhead, and monetary cost, providing a unified view of performance trade-offs. We show analytically and empirically that optimal scaling trajectories frequently lie along diagonal paths: sequences of joint horizontal and vertical adjustments that simultaneously exploit cluster parallelism and per-node improvements. To compute such actions, we propose DIAGONALSCALE, a discrete local-search algorithm that evaluates horizontal, vertical, and diagonal moves in the Scaling Plane and selects the configuration minimizing a multi-objective function subject to SLA constraints. Using synthetic surfaces, microbenchmarks, and experiments on distributed SQL and KV systems, we demonstrate that diagonal scaling reduces p95 latency by up to 40 percent, lowers cost-per-query by up to 37 percent, and reduces rebalancing by 2 to 5 times compared to horizontal-only and vertical-only autoscaling. Our results highlight the need for multi-dimensional scaling models and provide a foundation for next-generation autoscaling in cloud database systems.
Why we think this paper is great for you:
This paper offers valuable insights into optimizing performance and scaling for modern distributed and cloud databases, which is crucial for managing large datasets efficiently. You will find its multi-dimensional resource model particularly relevant for advanced database design.
AI Summary
  • DSR-SQL is evaluated on two benchmark datasets: Spider 2.0-Snow and BIRD, and compared with state-of-the-art baselines. [3]
  • The results show that DSR-SQL outperforms the baselines on both datasets, achieving a significant improvement in accuracy and efficiency. [3]
  • Schema-aware alignment: A technique used in DSR-SQL to align the schema of the database with the question being asked, allowing for more accurate SQL queries. [3]
  • DSR-SQL is a highly effective approach to natural language-to-SQL query generation, outperforming state-of-the-art baselines on two benchmark datasets. [3]
  • The paper does not provide a detailed analysis of the error cases where DSR-SQL failed to predict correctly. [3]
  • The evaluation is limited to two benchmark datasets, and it would be beneficial to evaluate DSR-SQL on more diverse datasets. [3]
  • The paper presents a novel approach to natural language-to-SQL query generation called DSR-SQL, which combines schema-aware alignment and adaptive schema selection. [2]
  • Adaptive schema selection: A mechanism in DSR-SQL that selects the most relevant tables and columns from the database schema based on the question being asked. [1]
Abstract
Recent divide-and-conquer reasoning approaches, particularly those based on Chain-of-Thought (CoT), have substantially improved the Text-to-SQL capabilities of Large Language Models (LLMs). However, when applied to complex enterprise databases, such methods struggle to maintain coherent reasoning due to limited context capacity, unreliable schema linking, and weak grounding in database semantics. To overcome these issues, we introduce DSR-SQL, a \textbf{D}ual-\textbf{S}tate \textbf{R}easoning framework that models Text-to-SQL as an interaction between an adaptive context state and a progressive generation state. The first constructs a compact, semantically faithful environment by refining large schemas and selecting relevant structures, while the second formalizes SQL synthesis as feedback-guided state transitions, enabling the model to self-correct and align with user intent. Without any post-training or in-context examples, DSR-SQL achieves competitive performance, reaching 35.28\% execution accuracy on Spider 2.0-Snow and 68.32\% on BIRD development set. Our implementation will be open-sourced at: https://github.com/DMIRLAB-Group/DSR-SQL.
Why we think this paper is great for you:
This research directly addresses enhancing Text-to-SQL capabilities for complex enterprise databases. It provides a cutting-edge approach to improve how you interact with and query large relational systems.
Paper visualization
Rate image: ๐Ÿ‘ ๐Ÿ‘Ž
AI Summary
  • The tool's optimized duplicate detection algorithm uses a three-tier pipeline, including exact byte size matching, sampled fingerprinting, and full-file cryptographic hashing. [3]
  • Fuzzy hashes: similarity-based hashing that captures file similarity rather than identity. [3]
  • Header/extension mismatch: a flag indicating when the file header type disagrees with its extension. [3]
  • SeqManager efficiently identifies and removes duplicate files in sequencing data, reducing storage costs by up to 98%. [2]
Abstract
Motivation: Modern genomics laboratories generate massive volumes of sequencing data, often resulting in significant storage costs. Genomics storage consists of duplicate files, temporary processing files, and redundant intermediate data. Results: We developed SeqManager, a web-based application that provides automated identification, classification, and management of sequencing data files with intelligent duplicate detection. It also detects intermediate sequencing files that can safely be removed. Evaluation across four genomics laboratory settings demonstrate that our tool is fast and has a very low memory footprint.
Why we think this paper is great for you:
You will appreciate this paper's focus on efficient data storage management and duplicate detection, principles that are highly applicable to designing robust data warehousing solutions. It offers practical insights into optimizing storage costs and data quality.
Paper visualization
Rate image: ๐Ÿ‘ ๐Ÿ‘Ž
AI Summary
  • The DataSquad program has been highly effective in providing students with practical experience and skills in data science, software engineering, and project management. [3]
  • The program's emphasis on teamwork, communication, and client interaction has helped students develop valuable soft skills. [3]
  • The DataSquad environment is highly encouraging, with 100% of alumni reporting that they felt encouraged to participate in the program. [3]
  • Statistical Analysis: Collecting, exploring and presenting large amounts of data to discover underlying patterns and trends Database Design/Cloud Systems: Designing a safe place to capture your data (in SQL or other), working with data capture or management tools like Qualtrics or Google Forms Coding, Software Engineering: Using programming languages, such as Python, R, etc., and utilizing file management tools like Git Project Management/Planning: Organizing tasks, managing time, and coordinating resources to achieve goals Effective Teamwork: Collaborating well with others, supporting teammates, and achieving shared objectives [3]
  • Many students experienced multiple roles during their tenure, gaining breadth across the program's offerings. [2]
Abstract
The DataSquad at Carleton College addresses a common problem at small liberal arts colleges: limited capacity for data services and few opportunities for students to gain practical experience with data and software development. Academic Technologist Paula Lackie designed the program as a work-study position that trains undergraduates through structured peer mentorship and real client projects. Students tackle data problems of increasing complexity-from basic data analysis to software development-while learning FAIR data principles and open science practices. The model's core components (peer mentorship structure, project-based learning, and communication training) make it adaptable to other institutions. UCLA and other colleges have adopted the model using openly shared materials through "DataSquad International." This paper describes the program's implementation at Carleton College and examines how structured peer mentorship can simultaneously improve institutional data services and provide students with professional skills and confidence.
Why we think this paper is great for you:
This paper provides an excellent perspective on gaining practical experience with data and software development, which is essential for applying your knowledge of database systems in real-world scenarios. It highlights the importance of hands-on data services.
Paper visualization
Rate image: ๐Ÿ‘ ๐Ÿ‘Ž
AI Summary
  • Explainable Artificial Intelligence (XAI) is rapidly evolving, moving beyond post-hoc methods towards a more comprehensive mechanistic understanding of AI model behavior and decision-making process. [3]
  • Embedding causal reasoning into interpretability pipelines through counterfactuals or mechanistic priors provides the necessary structure to distinguish genuine drivers from spurious correlations. [3]
  • Explainable Artificial Intelligence (XAI): AI that can explain its own decision-making process and provide insights into how it arrived at a particular conclusion. [3]
  • Causality: The relationship between cause and effect, essential for understanding the underlying mechanisms of complex systems. [3]
  • Interpretability: The ability to understand and interpret the results of an algorithm or model. [3]
  • XAI has the potential to contribute to trustworthy and discovery-driven science by providing insights into AI decision-making processes. [3]
  • Causality plays a decisive role in XAI, without causal grounding even stable and reproducible explanations risk being superficially plausible while scientifically misleading. [2]
  • Interpretability results that are accurate but misaligned with domain concepts may fail to resonate with experts, rendering them ineffective in real scientific workflows. [1]
Abstract
Artificial intelligence and machine learning are reshaping how we approach scientific discovery, not by replacing established methods but by extending what researchers can probe, predict, and design. In this roadmap we provide a forward-looking view of AI-enabled science across biology, chemistry, climate science, mathematics, materials science, physics, self-driving laboratories and unconventional computing. Several shared themes emerge: the need for diverse and trustworthy data, transferable electronic-structure and interatomic models, AI systems integrated into end-to-end scientific workflows that connect simulations to experiments and generative systems grounded in synthesisability rather than purely idealised phases. Across domains, we highlight how large foundation models, active learning and self-driving laboratories can close loops between prediction and validation while maintaining reproducibility and physical interpretability. Taken together, these perspectives outline where AI-enabled science stands today, identify bottlenecks in data, methods and infrastructure, and chart concrete directions for building AI systems that are not only more powerful but also more transparent and capable of accelerating discovery in complex real-world environments.
Why we think this paper is great for you:
This roadmap explores how AI is transforming scientific discovery, a domain that heavily relies on the effective management and analysis of vast datasets. You might find the discussion on AI-enabled science relevant to large-scale data infrastructure.
Paper visualization
Rate image: ๐Ÿ‘ ๐Ÿ‘Ž
AI Summary
  • The article discusses the Model Context Protocol (MCP) and its performance evaluation in serving model cards, comparing it to REST protocol. [2]
  • FAIR Signposting Profile: Implementation guidelines for exposing machine-actionable navigation links using standardized HTTP headers and HTML link elements. [1]
Abstract
AI/ML model cards can contain a benchmarked evaluation of an AI/ML model against intended use but a one time assessment during model training does not get at how and where a model is actually used over its lifetime. Through Patra Model Cards embedded in the ICICLE AI Institute software ecosystem we study model cards as dynamic objects. The study reported here assesses the benefits and tradeoffs of adopting the Model Context Protocol (MCP) as an interface to the Patra Model Card server. Quantitative assessment shows the overhead of MCP as compared to a REST interface. The core question however is of active sessions enabled by MCP; this is a qualitative question of fit and use in the context of dynamic model cards that we address as well.
Why we think this paper is great for you:
This paper discusses the lifecycle and evaluation of AI/ML models within software ecosystems, which often involves managing the data used for training and benchmarking. It touches on aspects of data governance and system integration around AI.
Abstract
Edge General Intelligence (EGI) represents a paradigm shift in mobile edge computing, where intelligent agents operate autonomously in dynamic, resource-constrained environments. However, the deployment of advanced agentic AI models on mobile and edge devices faces significant challenges due to limited computation, energy, and storage resources. To address these constraints, this survey investigates the integration of Knowledge Distillation (KD) into EGI, positioning KD as a key enabler for efficient, communication-aware, and scalable intelligence at the wireless edge. In particular, we emphasize KD techniques specifically designed for wireless communication and mobile networking, such as channel-aware self-distillation, cross-model Channel State Information (CSI) feedback distillation, and robust modulation/classification distillation. Furthermore, we review novel architectures natively suited for KD and edge deployment, such as Mamba, RWKV (Receptance, Weight, Key, Value) and Cross-Architecture distillation, which enhance generalization capabilities. Subsequently, we examine diverse applications in which KD-driven architectures enable EGI across vision, speech, and multimodal tasks. Finally, we highlight the key challenges and future directions for KD in EGI. This survey aims to provide a comprehensive reference for researchers exploring KD-driven frameworks for mobile agentic AI in the era of EGI.
Why we think this paper is great for you:
This research explores intelligent agents operating in resource-constrained environments, which implies unique challenges and solutions for data storage and access at the edge. You may find its insights relevant to distributed data management.
AI and Society
Abstract
Algorithms have been estimated to increase AI training FLOP efficiency by a factor of 22,000 between 2012 and 2023 [Ho et al., 2024]. Running small-scale ablation experiments on key innovations from this time period, we are able to account for less than 10x of these gains. Surveying the broader literature, we estimate that additional innovations not included in our ablations account for less than 10x, yielding a total under 100x. This leads us to conduct scaling experiments, which reveal that much of this efficiency gap can be explained by algorithms with scale-dependent efficiency improvements. In particular, we conduct scaling experiments between LSTMs and Transformers, finding exponent differences in their compute-optimal scaling law while finding little scaling difference for many other innovations. These experiments demonstrate that - contrary to standard assumptions - an algorithm's efficiency gains are tied to compute scale. Using experimental extrapolation and literature estimates, we account for 6,930x efficiency gains over the same time period, with the scale-dependent LSTM-to-Transformer transition accounting for the majority of gains. Our results indicate that algorithmic progress for small models has been far slower than previously assumed, and that measures of algorithmic efficiency are strongly reference-dependent.
AI Summary
  • Algorithmic progress in language models exhibits fundamentally different behavior across compute scales. [3]
  • Algorithmic progress: The improvement in training efficiency and capabilities of language models over time. [3]
  • Scale-dependent innovations: Innovations whose impact on efficiency gains varies depending on the compute scale. [3]
  • Algorithmic progress is not a single number, but rather depends on both the reference algorithm and target compute scale. [3]
  • Scale-dependent innovations are critical to understanding algorithmic progress and its implications for the future of AI. [3]
  • The study's experiments are conducted at small scales compared to more recent scaling studies. [3]
  • Scale-dependent innovations, such as the LSTM-to-Transformer transition and Chinchilla rebalancing, account for most of the efficiency gains at frontier scales. [2]
  • The concentration of progress in architectural transitions suggests that future progress may depend on discovering fundamentally new architectures rather than incremental refinements of existing ones. [1]
Abstract
Advances in artificial intelligence now make it possible to simulate the dead through chatbots, voice clones, and video avatars trained on a person's digital traces. These "digital ghosts" are moving from fiction to commercial reality, reshaping how people mourn and remember. This paper offers a conceptual and ethical analysis of AI-mediated digital afterlives. We define what counts as a digital ghost, trace their rise across personal, commercial, and institutional contexts, and identify core ethical tensions around grief and well-being, truthfulness and deception, consent and posthumous privacy, dignity and misrepresentation, and the commercialization of mourning. To analyze these challenges, we propose a nine-dimensional taxonomy of digital afterlife technologies and, building on it, outline the features of an ethically acceptable digital ghost: premortem intent, mutual consent, transparent and limited data use, clear disclosure, restricted purposes and access, family or estate stewardship, and minimal behavioral agency. We argue for targeted regulation and professional guidelines to ensure that digital ghosts can aid remembrance without slipping into forms of deception.
AI Summary
  • The rise of digital ghosts and deadbots forces us to confront fundamental questions about how we remember our dead. [3]
  • Digital ghosts can become the blind spot between memory and trickery, a prolonged mourning disguised as dialogue. [3]
  • Deadbots: AI replicas that simulate the presence of deceased individuals. [3]
  • Digital ghosts: AI-generated representations of deceased people that can interact with the living. [3]
  • The era of AI 'afterlives' is here, and it falls upon us to ensure this technology is used in a way that supports memory without becoming an imposture, and helps heal without betraying the dignity of those we love and lose. [3]
  • The article cites various studies and papers on the topic of digital ghosts and deadbots, including works by authors such as Jed R. [3]
  • From an ethical standpoint, arguably intent and transparency make a substantial difference in creating or engaging with a simulacrum of a loved one. [2]
  • Brubaker and John Danaher. [1]
Research Automation with AI
Abstract
Over the last few months, AI models including large language models have improved greatly. There are now several documented examples where they have helped professional mathematical scientists prove new results, sometimes even helping resolve known open problems. In this short note, we add another example to the list, by documenting how we were able to solve a previously unsolved research problem in robust mathematical statistics with crucial help from GPT-5. Our problem concerns robust density estimation, where the observations are perturbed by Wasserstein-bounded contaminations.In a previous preprint (Chao and Dobriban, 2023, arxiv:2308.01853v2), we have obtained upper and lower bounds on the minimax optimal estimation error; which were, however, not sharp. Starting in October 2025, making significant use of GPT-5 Pro, we were able to derive the minimax optimal error rate (reported in version 3 of the above arxiv preprint). GPT-5 provided crucial help along the way, including by suggesting calculations that we did not think of, and techniques that were not familiar to us, such as the dynamic Benamou-Brenier formulation, for key steps in the analysis. Working with GPT-5 took a few weeks of effort, and we estimate that it could have taken several months to get the same results otherwise. At the same time, there are still areas where working with GPT-5 was challenging: it sometimes provided incorrect references, and glossed over details that sometimes took days of work to fill in. We outline our workflow and steps taken to mitigate issues. Overall, our work can serve as additional documentation for a new age of human-AI collaborative work in mathematical science.
AI Summary
  • The authors highlight the potential benefits of using AI in research, including increased productivity and well-being for mathematicians. [3]
  • They also note that AI can excel at routine but lengthy calculations, freeing up time for more creative work. [3]
  • They also note that human-AI collaboration can lead to new insights and solutions. [3]
  • LLM: Large Language Model AI: Artificial Intelligence The use of AI in research has the potential to increase productivity and well-being for mathematicians. [3]
  • Human-AI collaboration can lead to new insights and solutions. [3]
  • The paper discusses the use of a large language model (LLM) in solving a research problem in mathematical statistics. [2]
AGI: Artificial General Intelligence
Paper visualization
Rate image: ๐Ÿ‘ ๐Ÿ‘Ž
Abstract
As artificial intelligence systems become increasingly integrated into human social contexts, Artificial Social Intelligence (ASI) has emerged as a critical capability that enables AI to perceive, understand, and engage meaningfully in complex human social interactions. This chapter introduces a comprehensive framework for Human-Centered Artificial Social Intelligence (HC-ASI), built upon the Technology-Human Factors-Ethics (THE) Triangle, which systematically addresses both technical foundations and human-centered design principles necessary for developing socially intelligent AI systems. This chapter provides a comprehensive overview of current ASI research. This chapter begins by establishing the theoretical foundations of ASI, tracing its evolution from classical psychological theories of human social intelligence to contemporary computational models, then examines the mechanisms underlying human-AI social interaction with particular emphasis on establishing shared social understanding and appropriate role positioning. The chapter further explores ASI's practical implications for individuals and groups through comprehensive evaluation frameworks that combine technical benchmarks with human-centered experiential assessments, demonstrating real-world applications through detailed case studies spanning healthcare, companionship, education, and customer service domains. Building on the overview and the framework of HC -ASI, this chapter articulates core HC-ASI design principles and translates them into actionable methodologies and implementation guidelines that provide practical guidance for researchers and practitioners. This chapter concludes with a critical discussion of current challenges and promising directions for developing comprehensive HC-ASI ecosystems.
AI Summary
  • Social AI is a type of artificial intelligence that can interact with humans in a way that simulates human-like conversation and behavior. [3]
  • The development of Social AI requires the integration of multiple disciplines, including computer science, psychology, sociology, and philosophy. [3]
  • Social AI has the potential to revolutionize various industries, such as healthcare, education, and customer service, by providing personalized support and improving user experience. [3]
  • Social AI: A type of artificial intelligence that can interact with humans in a way that simulates human-like conversation and behavior. [3]
  • Human-AI interaction: The process by which humans interact with artificial intelligence systems, such as chatbots or virtual assistants. [3]
  • The development of Social AI has the potential to transform various industries and improve user experience. [3]
  • However, the development of Social AI also raises several challenges and concerns, including ensuring transparency, accountability, and ethics in AI decision-making. [2]
Deep Learning
Abstract
In a study, published in \emph{Nature}, researchers from DeepMind and mathematicians demonstrated a general framework using machine learning to make conjectures in pure mathematics. Their work uses neural networks and attribution techniques to guide human intuition towards making provable conjectures. Here, we build upon this framework to develop a method for identifying sufficient conditions that imply a given mathematical statement. Our approach trains neural networks with a custom loss function that prioritizes high precision. Then uses attribution techniques and exploratory data analysis to make conjectures. As a demonstration, we apply this process to Stanley's problem of $e$-positivity of graphs--a problem that has been at the center of algebraic combinatorics for the past three decades. Guided by AI, we rediscover that one sufficient condition for a graph to be $e$-positive is that it is co-triangle-free, and that the number of claws is the most important factor for $e$-positivity. Based on the most important factors in Saliency Map analysis of neural networks, we suggest that the classification of $e$-positive graphs is more related to continuous graph invariants rather than the discrete ones. Furthermore, using neural networks and exploratory data analysis, we show that the claw-free and claw-contractible-free graphs with $10$ and $11$ vertices are $e$-positive, resolving a conjecture by Dahlberg, Foley, and van Willigenburg.
AI Summary
  • The authors used a precision-optimized model to identify the top four features that impact e-positivity in graphs. [3]
  • The model achieved 100% precision on the test set, indicating high reliability for its positive predictions. [3]
  • The study demonstrates how AI can be used to guide human intuition and advance mathematics by identifying underlying patterns associated with e-positivity. [3]
  • The authors' approach can be applied to other areas of mathematics where pattern recognition is crucial. [3]
  • E-positivity: a property of graphs that refers to the existence of certain combinatorial structures, such as cycles or paths. [3]
  • Chromatic symmetric function: a polynomial invariant of graphs that encodes information about their coloring properties. [3]
  • The study demonstrates the potential of AI in advancing mathematics by identifying underlying patterns associated with e-positivity. [3]
  • The precision-optimized model achieved high reliability for its positive predictions, indicating that it can be trusted with high confidence when classifying graphs as e-positive. [3]
  • The approach used in this study can be applied to other areas of mathematics where pattern recognition is crucial. [3]
  • Saliency Map analysis: a technique used to identify the most important features or variables in a dataset by computing the average gradient of the model's output with respect to its input features. [2]
Abstract
This paper argues that DNNs implement a computational Occam's razor -- finding the `simplest' algorithm that fits the data -- and that this could explain their incredible and wide-ranging success over more traditional statistical methods. We start with the discovery that the set of real-valued function $f$ that can be $ฮต$-approximated with a binary circuit of size at most $cฮต^{-ฮณ}$ becomes convex in the `Harder than Monte Carlo' (HTMC) regime, when $ฮณ>2$, allowing for the definition of a HTMC norm on functions. In parallel one can define a complexity measure on the parameters of a ResNets (a weighted $\ell_1$ norm of the parameters), which induce a `ResNet norm' on functions. The HTMC and ResNet norms can then be related by an almost matching sandwich bound. Thus minimizing this ResNet norm is equivalent to finding a circuit that fits the data with an almost minimal number of nodes (within a power of 2 of being optimal). ResNets thus appear as an alternative model for computation of real functions, better adapted to the HTMC regime and its convexity.
AI Summary
  • The HTMC norm is a measure of the complexity of a function, and it has several useful properties, including compositionality and convexity. [3]
  • The construction of the ResNet involves two main parts: first, the input is mapped to the weighted binary representations of its surrounding vertices; second, a sorting algorithm is used to recover the simplex that contains the input. [3]
  • The Lipschitz constant of this network is bounded by cp doutb|C|p=2/3. [3]
  • The ResNet representation of Tetrakis functions has several useful properties, including compositionality and convexity. [3]
  • HTMC norm: a measure of the complexity of a function. [3]
  • Hยจ older continuous: a property of a function that implies it can be represented as a sum of Tetrakis functions. [3]
  • Tetrakis function: a type of function that is both HTMC computable and Hยจ older continuous. [3]
  • ResNet: a type of neural network that can represent functions that are both HTMC computable and Hยจ older continuous. [3]
  • ResNets can be used to represent functions that are both HTMC computable and Hยจ older continuous. [2]

We did not find tons of content matching your interests we've included some additional topics that are popular. Also be aware that if the topics is not present in arxiv we wont be able to recommend it.

AI Agents
Abstract
AI/ML model cards can contain a benchmarked evaluation of an AI/ML model against intended use but a one time assessment during model training does not get at how and where a model is actually used over its lifetime. Through Patra Model Cards embedded in the ICICLE AI Institute software ecosystem we study model cards as dynamic objects. The study reported here assesses the benefits and tradeoffs of adopting the Model Context Protocol (MCP) as an interface to the Patra Model Card server. Quantitative assessment shows the overhead of MCP as compared to a REST interface. The core question however is of active sessions enabled by MCP; this is a qualitative question of fit and use in the context of dynamic model cards that we address as well.
AI Summary
  • The article discusses the Model Context Protocol (MCP) and its performance evaluation in serving model cards, comparing it to REST protocol. [2]
  • FAIR Signposting Profile: Implementation guidelines for exposing machine-actionable navigation links using standardized HTTP headers and HTML link elements. [1]
Abstract
Edge General Intelligence (EGI) represents a paradigm shift in mobile edge computing, where intelligent agents operate autonomously in dynamic, resource-constrained environments. However, the deployment of advanced agentic AI models on mobile and edge devices faces significant challenges due to limited computation, energy, and storage resources. To address these constraints, this survey investigates the integration of Knowledge Distillation (KD) into EGI, positioning KD as a key enabler for efficient, communication-aware, and scalable intelligence at the wireless edge. In particular, we emphasize KD techniques specifically designed for wireless communication and mobile networking, such as channel-aware self-distillation, cross-model Channel State Information (CSI) feedback distillation, and robust modulation/classification distillation. Furthermore, we review novel architectures natively suited for KD and edge deployment, such as Mamba, RWKV (Receptance, Weight, Key, Value) and Cross-Architecture distillation, which enhance generalization capabilities. Subsequently, we examine diverse applications in which KD-driven architectures enable EGI across vision, speech, and multimodal tasks. Finally, we highlight the key challenges and future directions for KD in EGI. This survey aims to provide a comprehensive reference for researchers exploring KD-driven frameworks for mobile agentic AI in the era of EGI.
AI and Society
Abstract
Algorithms have been estimated to increase AI training FLOP efficiency by a factor of 22,000 between 2012 and 2023 [Ho et al., 2024]. Running small-scale ablation experiments on key innovations from this time period, we are able to account for less than 10x of these gains. Surveying the broader literature, we estimate that additional innovations not included in our ablations account for less than 10x, yielding a total under 100x. This leads us to conduct scaling experiments, which reveal that much of this efficiency gap can be explained by algorithms with scale-dependent efficiency improvements. In particular, we conduct scaling experiments between LSTMs and Transformers, finding exponent differences in their compute-optimal scaling law while finding little scaling difference for many other innovations. These experiments demonstrate that - contrary to standard assumptions - an algorithm's efficiency gains are tied to compute scale. Using experimental extrapolation and literature estimates, we account for 6,930x efficiency gains over the same time period, with the scale-dependent LSTM-to-Transformer transition accounting for the majority of gains. Our results indicate that algorithmic progress for small models has been far slower than previously assumed, and that measures of algorithmic efficiency are strongly reference-dependent.
AI Summary
  • Algorithmic progress in language models exhibits fundamentally different behavior across compute scales. [3]
  • Algorithmic progress: The improvement in training efficiency and capabilities of language models over time. [3]
  • Scale-dependent innovations: Innovations whose impact on efficiency gains varies depending on the compute scale. [3]
  • Algorithmic progress is not a single number, but rather depends on both the reference algorithm and target compute scale. [3]
  • Scale-dependent innovations are critical to understanding algorithmic progress and its implications for the future of AI. [3]
  • The study's experiments are conducted at small scales compared to more recent scaling studies. [3]
  • Scale-dependent innovations, such as the LSTM-to-Transformer transition and Chinchilla rebalancing, account for most of the efficiency gains at frontier scales. [2]
  • The concentration of progress in architectural transitions suggests that future progress may depend on discovering fundamentally new architectures rather than incremental refinements of existing ones. [1]
Abstract
Advances in artificial intelligence now make it possible to simulate the dead through chatbots, voice clones, and video avatars trained on a person's digital traces. These "digital ghosts" are moving from fiction to commercial reality, reshaping how people mourn and remember. This paper offers a conceptual and ethical analysis of AI-mediated digital afterlives. We define what counts as a digital ghost, trace their rise across personal, commercial, and institutional contexts, and identify core ethical tensions around grief and well-being, truthfulness and deception, consent and posthumous privacy, dignity and misrepresentation, and the commercialization of mourning. To analyze these challenges, we propose a nine-dimensional taxonomy of digital afterlife technologies and, building on it, outline the features of an ethically acceptable digital ghost: premortem intent, mutual consent, transparent and limited data use, clear disclosure, restricted purposes and access, family or estate stewardship, and minimal behavioral agency. We argue for targeted regulation and professional guidelines to ensure that digital ghosts can aid remembrance without slipping into forms of deception.
AI Summary
  • The rise of digital ghosts and deadbots forces us to confront fundamental questions about how we remember our dead. [3]
  • Digital ghosts can become the blind spot between memory and trickery, a prolonged mourning disguised as dialogue. [3]
  • Deadbots: AI replicas that simulate the presence of deceased individuals. [3]
  • Digital ghosts: AI-generated representations of deceased people that can interact with the living. [3]
  • The era of AI 'afterlives' is here, and it falls upon us to ensure this technology is used in a way that supports memory without becoming an imposture, and helps heal without betraying the dignity of those we love and lose. [3]
  • The article cites various studies and papers on the topic of digital ghosts and deadbots, including works by authors such as Jed R. [3]
  • From an ethical standpoint, arguably intent and transparency make a substantial difference in creating or engaging with a simulacrum of a loved one. [2]
  • Brubaker and John Danaher. [1]
Research Automation with AI
Abstract
Over the last few months, AI models including large language models have improved greatly. There are now several documented examples where they have helped professional mathematical scientists prove new results, sometimes even helping resolve known open problems. In this short note, we add another example to the list, by documenting how we were able to solve a previously unsolved research problem in robust mathematical statistics with crucial help from GPT-5. Our problem concerns robust density estimation, where the observations are perturbed by Wasserstein-bounded contaminations.In a previous preprint (Chao and Dobriban, 2023, arxiv:2308.01853v2), we have obtained upper and lower bounds on the minimax optimal estimation error; which were, however, not sharp. Starting in October 2025, making significant use of GPT-5 Pro, we were able to derive the minimax optimal error rate (reported in version 3 of the above arxiv preprint). GPT-5 provided crucial help along the way, including by suggesting calculations that we did not think of, and techniques that were not familiar to us, such as the dynamic Benamou-Brenier formulation, for key steps in the analysis. Working with GPT-5 took a few weeks of effort, and we estimate that it could have taken several months to get the same results otherwise. At the same time, there are still areas where working with GPT-5 was challenging: it sometimes provided incorrect references, and glossed over details that sometimes took days of work to fill in. We outline our workflow and steps taken to mitigate issues. Overall, our work can serve as additional documentation for a new age of human-AI collaborative work in mathematical science.
AI Summary
  • The authors highlight the potential benefits of using AI in research, including increased productivity and well-being for mathematicians. [3]
  • They also note that AI can excel at routine but lengthy calculations, freeing up time for more creative work. [3]
  • They also note that human-AI collaboration can lead to new insights and solutions. [3]
  • LLM: Large Language Model AI: Artificial Intelligence The use of AI in research has the potential to increase productivity and well-being for mathematicians. [3]
  • Human-AI collaboration can lead to new insights and solutions. [3]
  • The paper discusses the use of a large language model (LLM) in solving a research problem in mathematical statistics. [2]
Abstract
Artificial intelligence and machine learning are reshaping how we approach scientific discovery, not by replacing established methods but by extending what researchers can probe, predict, and design. In this roadmap we provide a forward-looking view of AI-enabled science across biology, chemistry, climate science, mathematics, materials science, physics, self-driving laboratories and unconventional computing. Several shared themes emerge: the need for diverse and trustworthy data, transferable electronic-structure and interatomic models, AI systems integrated into end-to-end scientific workflows that connect simulations to experiments and generative systems grounded in synthesisability rather than purely idealised phases. Across domains, we highlight how large foundation models, active learning and self-driving laboratories can close loops between prediction and validation while maintaining reproducibility and physical interpretability. Taken together, these perspectives outline where AI-enabled science stands today, identify bottlenecks in data, methods and infrastructure, and chart concrete directions for building AI systems that are not only more powerful but also more transparent and capable of accelerating discovery in complex real-world environments.
AI Summary
  • Explainable Artificial Intelligence (XAI) is rapidly evolving, moving beyond post-hoc methods towards a more comprehensive mechanistic understanding of AI model behavior and decision-making process. [3]
  • Embedding causal reasoning into interpretability pipelines through counterfactuals or mechanistic priors provides the necessary structure to distinguish genuine drivers from spurious correlations. [3]
  • Explainable Artificial Intelligence (XAI): AI that can explain its own decision-making process and provide insights into how it arrived at a particular conclusion. [3]
  • Causality: The relationship between cause and effect, essential for understanding the underlying mechanisms of complex systems. [3]
  • Interpretability: The ability to understand and interpret the results of an algorithm or model. [3]
  • XAI has the potential to contribute to trustworthy and discovery-driven science by providing insights into AI decision-making processes. [3]
  • Causality plays a decisive role in XAI, without causal grounding even stable and reproducible explanations risk being superficially plausible while scientifically misleading. [2]
  • Interpretability results that are accurate but misaligned with domain concepts may fail to resonate with experts, rendering them ineffective in real scientific workflows. [1]
AGI: Artificial General Intelligence
Abstract
As artificial intelligence systems become increasingly integrated into human social contexts, Artificial Social Intelligence (ASI) has emerged as a critical capability that enables AI to perceive, understand, and engage meaningfully in complex human social interactions. This chapter introduces a comprehensive framework for Human-Centered Artificial Social Intelligence (HC-ASI), built upon the Technology-Human Factors-Ethics (THE) Triangle, which systematically addresses both technical foundations and human-centered design principles necessary for developing socially intelligent AI systems. This chapter provides a comprehensive overview of current ASI research. This chapter begins by establishing the theoretical foundations of ASI, tracing its evolution from classical psychological theories of human social intelligence to contemporary computational models, then examines the mechanisms underlying human-AI social interaction with particular emphasis on establishing shared social understanding and appropriate role positioning. The chapter further explores ASI's practical implications for individuals and groups through comprehensive evaluation frameworks that combine technical benchmarks with human-centered experiential assessments, demonstrating real-world applications through detailed case studies spanning healthcare, companionship, education, and customer service domains. Building on the overview and the framework of HC -ASI, this chapter articulates core HC-ASI design principles and translates them into actionable methodologies and implementation guidelines that provide practical guidance for researchers and practitioners. This chapter concludes with a critical discussion of current challenges and promising directions for developing comprehensive HC-ASI ecosystems.
AI Summary
  • Social AI is a type of artificial intelligence that can interact with humans in a way that simulates human-like conversation and behavior. [3]
  • The development of Social AI requires the integration of multiple disciplines, including computer science, psychology, sociology, and philosophy. [3]
  • Social AI has the potential to revolutionize various industries, such as healthcare, education, and customer service, by providing personalized support and improving user experience. [3]
  • Social AI: A type of artificial intelligence that can interact with humans in a way that simulates human-like conversation and behavior. [3]
  • Human-AI interaction: The process by which humans interact with artificial intelligence systems, such as chatbots or virtual assistants. [3]
  • The development of Social AI has the potential to transform various industries and improve user experience. [3]
  • However, the development of Social AI also raises several challenges and concerns, including ensuring transparency, accountability, and ethics in AI decision-making. [2]
Deep Learning
Abstract
In a study, published in \emph{Nature}, researchers from DeepMind and mathematicians demonstrated a general framework using machine learning to make conjectures in pure mathematics. Their work uses neural networks and attribution techniques to guide human intuition towards making provable conjectures. Here, we build upon this framework to develop a method for identifying sufficient conditions that imply a given mathematical statement. Our approach trains neural networks with a custom loss function that prioritizes high precision. Then uses attribution techniques and exploratory data analysis to make conjectures. As a demonstration, we apply this process to Stanley's problem of $e$-positivity of graphs--a problem that has been at the center of algebraic combinatorics for the past three decades. Guided by AI, we rediscover that one sufficient condition for a graph to be $e$-positive is that it is co-triangle-free, and that the number of claws is the most important factor for $e$-positivity. Based on the most important factors in Saliency Map analysis of neural networks, we suggest that the classification of $e$-positive graphs is more related to continuous graph invariants rather than the discrete ones. Furthermore, using neural networks and exploratory data analysis, we show that the claw-free and claw-contractible-free graphs with $10$ and $11$ vertices are $e$-positive, resolving a conjecture by Dahlberg, Foley, and van Willigenburg.
AI Summary
  • The authors used a precision-optimized model to identify the top four features that impact e-positivity in graphs. [3]
  • The model achieved 100% precision on the test set, indicating high reliability for its positive predictions. [3]
  • The study demonstrates how AI can be used to guide human intuition and advance mathematics by identifying underlying patterns associated with e-positivity. [3]
  • The authors' approach can be applied to other areas of mathematics where pattern recognition is crucial. [3]
  • E-positivity: a property of graphs that refers to the existence of certain combinatorial structures, such as cycles or paths. [3]
  • Chromatic symmetric function: a polynomial invariant of graphs that encodes information about their coloring properties. [3]
  • The study demonstrates the potential of AI in advancing mathematics by identifying underlying patterns associated with e-positivity. [3]
  • The precision-optimized model achieved high reliability for its positive predictions, indicating that it can be trusted with high confidence when classifying graphs as e-positive. [3]
  • The approach used in this study can be applied to other areas of mathematics where pattern recognition is crucial. [3]
  • Saliency Map analysis: a technique used to identify the most important features or variables in a dataset by computing the average gradient of the model's output with respect to its input features. [2]
Abstract
This paper argues that DNNs implement a computational Occam's razor -- finding the `simplest' algorithm that fits the data -- and that this could explain their incredible and wide-ranging success over more traditional statistical methods. We start with the discovery that the set of real-valued function $f$ that can be $ฮต$-approximated with a binary circuit of size at most $cฮต^{-ฮณ}$ becomes convex in the `Harder than Monte Carlo' (HTMC) regime, when $ฮณ>2$, allowing for the definition of a HTMC norm on functions. In parallel one can define a complexity measure on the parameters of a ResNets (a weighted $\ell_1$ norm of the parameters), which induce a `ResNet norm' on functions. The HTMC and ResNet norms can then be related by an almost matching sandwich bound. Thus minimizing this ResNet norm is equivalent to finding a circuit that fits the data with an almost minimal number of nodes (within a power of 2 of being optimal). ResNets thus appear as an alternative model for computation of real functions, better adapted to the HTMC regime and its convexity.
AI Summary
  • The HTMC norm is a measure of the complexity of a function, and it has several useful properties, including compositionality and convexity. [3]
  • The construction of the ResNet involves two main parts: first, the input is mapped to the weighted binary representations of its surrounding vertices; second, a sorting algorithm is used to recover the simplex that contains the input. [3]
  • The Lipschitz constant of this network is bounded by cp doutb|C|p=2/3. [3]
  • The ResNet representation of Tetrakis functions has several useful properties, including compositionality and convexity. [3]
  • HTMC norm: a measure of the complexity of a function. [3]
  • Hยจ older continuous: a property of a function that implies it can be represented as a sum of Tetrakis functions. [3]
  • Tetrakis function: a type of function that is both HTMC computable and Hยจ older continuous. [3]
  • ResNet: a type of neural network that can represent functions that are both HTMC computable and Hยจ older continuous. [3]
  • ResNets can be used to represent functions that are both HTMC computable and Hยจ older continuous. [2]

Interests not found

We did not find any papers that match the below interests. Try other terms also consider if the content exists in arxiv.org.
  • Database Design
  • Relational Databases
You can edit or add more interests any time.