Papers from 22 to 26 September, 2025

Here are the personalized paper recommendations sorted by most relevant
Data Science Career Advice
👍 👎 ♥ Save
Abstract
The IT industry provides supportive pathways such as returnship programs, coding boot camps, and buddy systems for women re-entering their job after a career break. Academia, however, offers limited opportunities to motivate women to return. We propose a diverse multicultural research project investigating the challenges faced by women with software engineering (SE) backgrounds re-entering academia or related research roles after a career break. Career disruptions due to pregnancy, immigration status, or lack of flexible work options can significantly impact women's career progress, creating barriers for returning as lecturers, professors, or senior researchers. Although many companies promote gender diversity policies, such measures are less prominent and often under-recognized within academic institutions. Our goal is to explore the specific challenges women encounter when re-entering academic roles compared to industry roles; to understand the institutional perspective, including a comparative analysis of existing policies and opportunities in different countries for women to return to the field; and finally, to provide recommendations that support transparent hiring practices. The research project will be carried out in multiple universities and in multiple countries to capture the diverse challenges and policies that vary by location.
Data Careers
👍 👎 ♥ Save
The Pennsylvania State Un
Paper visualization
Rate this image: 😍 👍 👎
Abstract
The rapid advancement of Large Language Models (LLMs) has enabled the generation of highly realistic synthetic data. We identify a new vulnerability, LLMs generating convincing career trajectories in fake resumes and explore effective detection methods. To address this challenge, we construct a dataset of machine-generated career trajectories using LLMs and various methods, and demonstrate that conventional text-based detectors perform poorly on structured career data. We propose CareerScape, a novel heterogeneous, hierarchical multi-layer graph framework that models career entities and their relations in a unified global graph built from genuine resumes. Unlike conventional classifiers that treat each instance independently, CareerScape employs a structure-aware framework that augments user-specific subgraphs with trusted neighborhood information from a global graph, enabling the model to capture both global structural patterns and local inconsistencies indicative of synthetic career paths. Experimental results show that CareerScape outperforms state-of-the-art baselines by 5.8-85.0% relatively, highlighting the importance of structure-aware detection for machine-generated content.
AI Insights
  • An Agent‑Based Generator lets two LLMs converse, refining career paths until they read like real resumes.
  • A Critic Agent flags improbable moves, guiding the Generator toward authentic transitions.
  • The authors released 4,000 synthetic career trajectories, a ready‑made benchmark for LLM evaluation.
  • A template turns structured data into natural‑language narratives, easing LLM testing.
  • CareerScape’s multi‑layer graph fuses user subgraphs with a global resume network, spotting subtle inconsistencies.
  • Building on prior synthetic‑data work, this paper models complex industry shifts that earlier methods missed.
  • GitHub hosts the Agent‑Based Generator code and dataset, inviting researchers to extend the benchmark.
Data Career Development
👍 👎 ♥ Save
Indian Institute of Scine
Abstract
Scientific mobility shapes individual research careers and national innovation by enabling knowledge exchange, fostering collaborations, and providing access to leading research environments. Studying international mobility patterns of researchers from developing countries offers insights into strengthening domestic scientific ecosystems and addressing talent migration. We analyze the international mobility of India-affiliated researchers using longitudinal affiliation trajectories from the OpenAlex database, covering 157,471 researchers categorized as immobile, returnees, or settled abroad after moving to the US, EU, or other high-income countries. Our analysis shows that 28% experience at least one international move, yet over 73% never return, highlighting persistent brain drain. Internationally mobile researchers predominantly originate from premier Indian institutions. Matched pair analyses demonstrate that mobility yields lasting benefits: citation impact increases, publication rates align with immobile peers, and international collaboration rises-foreign co-author share grows from 52% to 83-87% at transition abroad and remains elevated among returnees (32-40 percentage points across disciplines). Returnees maintain global networks, bridging Indian science with global research. These patterns are consistent across major research disciplines, emphasizing that scientific mobility drives excellence and engagement while posing challenges for developing nations seeking to reintegrate talent.
AI Insights
  • OpenAlex data on 157k Indian researchers shows 28% moved abroad, yet 73% never returned, underscoring brain drain.
  • Returnees’ citation impact climbs 30–40% and international co‑authorship rises from 52% to 83–87%, while publication rates equal immobile peers.
  • Researchers from premier Indian institutions dominate the abroad cohort, revealing elite pipelines.
  • Mobility benefits persist across all major disciplines, indicating universal career gains.
  • Citation impact and international collaboration predict return likelihood, serving as both outcome and driver.
  • Policy implication: incentives are needed to retain talent and strengthen returnee networks linking India to global science.

We did not find tons of content matching your interests we've included some additional topics that are popular. Also be aware that if the topics is not present in arxiv we wont be able to recommend it.

AI Agents
👍 👎 ♥ Save
Ixent Games
Abstract
Current Large Reasoning Models (LRMs) exhibit significant limitations in reliability and transparency, often showing a collapse in reasoning capabilities when faced with high-complexity, long-horizon tasks. This "illusion of thinking" is frequently an artifact of non-agentic, black-box evaluation paradigms that fail to cultivate robust problem-solving processes. In response, we introduce The STAR-XAI Protocol (Socratic, Transparent, Agentic, Reasoning - for eXplainable Artificial Intelligence), a novel methodology for training and operating verifiably reliable AI agents. Our method reframes the human-AI interaction as a structured, Socratic dialogue, governed by an explicit and evolving rulebook, the Consciousness Transfer Package (CTP). Through an interactive Gameplay Cycle that enforces ante-hoc strategic justification and a state-locking Checksum that prevents error accumulation, the protocol transforms a powerful but opaque LRM into a disciplined "Clear Box" agent. We demonstrate the efficacy of this method through an exhaustive 25-move case study in the complex strategic game "Caps i Caps". The agent not only solved the high-complexity puzzle but also demonstrated Second-Order Agency, identifying flaws in its own supervisor-approved plans and adapting its core integrity protocols mid-task. The STAR-XAI Protocol offers a practical pathway to creating AI agents that are not just high-performing, but also transparent, auditable, and trustworthy by design.
AI Insights
  • The Consciousness Transfer Package (CTP) is a step‑by‑step manual for mastering gear‑based board games, covering placement, rotation, and vector mechanics.
  • CTP provides concrete examples of successful moves, letting Gems learn proven strategies instead of trial‑and‑error.
  • The package is designed for seamless handoff, so one Gem can train an agent that another can use without knowledge loss.
  • Recommended literature includes reasoning classics, game‑theory treatises, and studies on gear‑placement efficiency.
  • Online forums and simulation tools are highlighted as practical resources for testing and refining gear‑game tactics.
  • A caveat: the CTP’s depth may overwhelm novices and assumes baseline familiarity with gear‑game mechanics.
  • Core definitions—gear, placement, rotation, vector, base—ensure consistent terminology across training sessions.
👍 👎 ♥ Save
Abstract
Modern socio-economic systems are undergoing deep integration with artificial intelligence technologies. This paper constructs a heterogeneous agent-based modeling framework that incorporates both human workers and autonomous AI agents, to study the impact of AI collaboration under resource constraints on aggregate social output. We build five progressively extended models: Model 1 serves as the baseline of pure human collaboration; Model 2 introduces AI as collaborators; Model 3 incorporates network effects among agents; Model 4 treats agents as independent producers; and Model 5 integrates both network effects and independent agent production. Through theoretical derivation and simulation analysis, we find that the introduction of AI agents can significantly increase aggregate social output. When considering network effects among agents, this increase exhibits nonlinear growth far exceeding the simple sum of individual contributions. Under the same resource inputs, treating agents as independent producers provides higher long-term growth potential; introducing network effects further demonstrates strong characteristics of increasing returns to scale.
AI and Society
👍 👎 ♥ Save
University of Buenos Ares
Paper visualization
Rate this image: 😍 👍 👎
Abstract
This paper develops a taxonomy of expert perspectives on the risks and likely consequences of artificial intelligence, with particular focus on Artificial General Intelligence (AGI) and Artificial Superintelligence (ASI). Drawing from primary sources, we identify three predominant doctrines: (1) The dominance doctrine, which predicts that the first actor to create sufficiently advanced AI will attain overwhelming strategic superiority sufficient to cheaply neutralize its opponents' defenses; (2) The extinction doctrine, which anticipates that humanity will likely lose control of ASI, leading to the extinction of the human species or its permanent disempowerment; (3) The replacement doctrine, which forecasts that AI will automate a large share of tasks currently performed by humans, but will not be so transformative as to fundamentally reshape or bring an end to human civilization. We examine the assumptions and arguments underlying each doctrine, including expectations around the pace of AI progress and the feasibility of maintaining advanced AI under human control. While the boundaries between doctrines are sometimes porous and many experts hedge across them, this taxonomy clarifies the core axes of disagreement over the anticipated scale and nature of the consequences of AI development.
Research Automation with AI
👍 👎 ♥ Save
Paper visualization
Rate this image: 😍 👍 👎
Abstract
Climate data science faces persistent barriers stemming from the fragmented nature of data sources, heterogeneous formats, and the steep technical expertise required to identify, acquire, and process datasets. These challenges limit participation, slow discovery, and reduce the reproducibility of scientific workflows. In this paper, we present a proof of concept for addressing these barriers through the integration of a curated knowledge graph (KG) with AI agents designed for cloud-native scientific workflows. The KG provides a unifying layer that organizes datasets, tools, and workflows, while AI agents -- powered by generative AI services -- enable natural language interaction, automated data access, and streamlined analysis. Together, these components drastically lower the technical threshold for engaging in climate data science, enabling non-specialist users to identify and analyze relevant datasets. By leveraging existing cloud-ready API data portals, we demonstrate that "a knowledge graph is all you need" to unlock scalable and agentic workflows for scientific inquiry. The open-source design of our system further supports community contributions, ensuring that the KG and associated tools can evolve as a shared commons. Our results illustrate a pathway toward democratizing access to climate data and establishing a reproducible, extensible framework for human--AI collaboration in scientific research.
👍 👎 ♥ Save
KT
Abstract
KT developed a Responsible AI (RAI) assessment methodology and risk mitigation technologies to ensure the safety and reliability of AI services. By analyzing the Basic Act on AI implementation and global AI governance trends, we established a unique approach for regulatory compliance and systematically identify and manage all potential risk factors from AI development to operation. We present a reliable assessment methodology that systematically verifies model safety and robustness based on KT's AI risk taxonomy tailored to the domestic environment. We also provide practical tools for managing and mitigating identified AI risks. With the release of this report, we also release proprietary Guardrail : SafetyGuard that blocks harmful responses from AI models in real-time, supporting the enhancement of safety in the domestic AI development ecosystem. We also believe these research outcomes provide valuable insights for organizations seeking to develop Responsible AI.
AI Insights
  • The risk taxonomy categorizes threats into data, model, deployment, and societal dimensions, each with measurable indicators.
  • A multi‑stage assessment pipeline integrates static code analysis, adversarial testing, and human‑in‑the‑loop audits to quantify robustness.
  • SafetyGuard employs a lightweight transformer‑based policy network that intercepts outputs in real time, achieving <5 ms latency on edge devices.
  • Compliance mapping aligns each risk factor with specific clauses of the Basic Act on AI, enabling automated audit reports.
  • Pilot deployments in Korean telecom and finance sectors demonstrated a 30 % reduction in policy‑violating incidents after Guardrail integration.
  • The report proposes a future research agenda on explainable mitigation strategies and cross‑border data‑sharing protocols.
AGI: Artificial General Intelligence
👍 👎 ♥ Save
Abstract
Safety, trust and Artificial General Intelligence (AGI) are aspirational goals in artificial intelligence (AI) systems, and there are several informal interpretations of these notions. In this paper, we propose strict, mathematical definitions of safety, trust, and AGI, and demonstrate a fundamental incompatibility between them. We define safety of a system as the property that it never makes any false claims, trust as the assumption that the system is safe, and AGI as the property of an AI system always matching or exceeding human capability. Our core finding is that -- for our formal definitions of these notions -- a safe and trusted AI system cannot be an AGI system: for such a safe, trusted system there are task instances which are easily and provably solvable by a human but not by the system. We note that we consider strict mathematical definitions of safety and trust, and it is possible for real-world deployments to instead rely on alternate, practical interpretations of these notions. We show our results for program verification, planning, and graph reachability. Our proofs draw parallels to G\"odel's incompleteness theorems and Turing's proof of the undecidability of the halting problem, and can be regarded as interpretations of G\"odel's and Turing's results.
Deep Learning
👍 👎 ♥ Save
University of Pennsylvann
Paper visualization
Rate this image: 😍 👍 👎
Abstract
Given the widespread use of deep learning models in safety-critical applications, ensuring that the decisions of such models are robust against adversarial exploitation is of fundamental importance. In this thesis, we discuss recent progress toward designing algorithms that exhibit desirable robustness properties. First, we discuss the problem of adversarial examples in computer vision, for which we introduce new technical results, training paradigms, and certification algorithms. Next, we consider the problem of domain generalization, wherein the task is to train neural networks to generalize from a family of training distributions to unseen test distributions. We present new algorithms that achieve state-of-the-art generalization in medical imaging, molecular identification, and image classification. Finally, we study the setting of jailbreaking large language models (LLMs), wherein an adversarial user attempts to design prompts that elicit objectionable content from an LLM. We propose new attacks and defenses, which represent the frontier of progress toward designing robust language-based agents.
AI Insights
  • Random erasing data augmentation injects stochastic occlusions during training, boosting pixel‑level robustness.
  • Stability training enforces Lipschitz continuity across layers, yielding provable robustness margins.
  • Robust prompt optimization tailors LLM inputs to shrink jailbreak‑induced decision space.
  • Universal adversarial attacks generate a single perturbation that transfers across many inputs, breaking input‑specific defenses.
  • Randomness in SGD can amplify or dampen adversarial vulnerability, depending on learning‑rate schedules.
  • Tooling—automated augmentation pipelines and reproducibility frameworks—drives consistent robustness across labs.
  • “Essentials of Robust Control” links classical control theory to deep learning, providing a rigorous basis for safe neural systems.
👍 👎 ♥ Save
Istanbul Medeniyet Univer
Abstract
Deep learning optimizers are optimization algorithms that enable deep neural networks to learn. The effectiveness of learning is highly dependent on the optimizer employed in the training process. Alongside the rapid advancement of deep learning, a wide range of optimizers with different approaches have been developed. This study aims to provide a review of various optimizers that have been proposed and received attention in the literature. From Stochastic gradient descent to the most recent ones such as Momentum, AdamW, Sophia, and Muon in chronological order, optimizers are examined individually, and their distinctive features are highlighted in the study. The update rule of each optimizer is presented in detail, with an explanation of the associated concepts and variables. The techniques applied by these optimizers, their contributions to the optimization process, and their default hyperparameter settings are also discussed. In addition, insights are offered into the open challenges encountered in the optimization of deep learning models. Thus, a comprehensive resource is provided both for understanding the current state of optimizers and for identifying potential areas of future development.

Interests not found

We did not find any papers that match the below interests. Try other terms also consider if the content exists in arxiv.org.
  • Data Science Career Guidance
  • Data Career Path
You can edit or add more interests any time.

Unsubscribe from these updates