Hi!

Your personalized paper recommendations for 26 to 30 January, 2026.
Harvard University
AI Insights
  • Milestones serve dual pedagogical and validation purposes, providing motivation through historical framing and demonstrating implementation correctness through real-world task performance. (ML: 0.98)πŸ‘πŸ‘Ž
  • Each module concludes with systems reasoning prompts measuring conceptual understanding beyond syntactic correctness. (ML: 0.97)πŸ‘πŸ‘Ž
  • Milestones are designed to be challenging but achievable, allowing students to demonstrate their understanding of complex concepts through real-world tasks. (ML: 0.96)πŸ‘πŸ‘Ž
  • Assessment validates both isolated correctness and cross-module integration. (ML: 0.96)πŸ‘πŸ‘Ž
  • The TinyTorch framework is designed for teaching machine learning concepts through hands-on implementation and analysis. (ML: 0.95)πŸ‘πŸ‘Ž
  • Reflect: Systems Analysis Questions. (ML: 0.94)πŸ‘πŸ‘Ž
  • TinyTorch follows a consistent Build-Use-Reflect cycle, integrating implementation, application, and systems reasoning to address multiple learning objectives. (ML: 0.94)πŸ‘πŸ‘Ž
  • It's a pedagogical tool aimed at bridging the gap between theoretical understanding and practical application. (ML: 0.94)πŸ‘πŸ‘Ž
  • Students implement components in Jupyter notebooks with scaffolded guidance. (ML: 0.91)πŸ‘πŸ‘Ž
  • TinyTorch's design emphasizes systems thinking, encouraging students to analyze and understand the relationships between components, rather than just focusing on individual functions. (ML: 0.87)πŸ‘πŸ‘Ž
  • The framework includes six historical milestones that recreate actual breakthroughs using exclusively student code, validating success through task-appropriate performance. (ML: 0.85)πŸ‘πŸ‘Ž
  • The framework is built with a focus on explicit dependencies, making it easier for students to understand where each module fits in the larger architecture. (ML: 0.83)πŸ‘πŸ‘Ž
  • Use: Integration Testing Beyond Unit Tests. (ML: 0.77)πŸ‘πŸ‘Ž
  • Build: Implementation with Explicit Dependencies. (ML: 0.66)πŸ‘πŸ‘Ž
Abstract
Machine learning education faces a fundamental gap: students learn algorithms without understanding the systems that execute them. They study gradient descent without measuring memory, attention mechanisms without analyzing O(N^2) scaling, optimizer theory without knowing why Adam requires 3x the memory of SGD. This "algorithm-systems divide" produces practitioners who can train models but cannot debug memory failures, optimize inference latency, or reason about deployment trade-offs--the very skills industry demands as "ML systems engineering." We present TinyTorch, a 20-module curriculum that closes this gap through "implementation-based systems pedagogy": students construct PyTorch's core components (tensors, autograd, optimizers, CNNs, transformers) in pure Python, building a complete framework where every operation they invoke is code they wrote. The design employs three patterns: "progressive disclosure" of complexity, "systems-first integration" of profiling from the first module, and "build-to-validate milestones" recreating 67 years of ML breakthroughs--from Perceptron (1958) through Transformers (2017) to MLPerf-style benchmarking. Requiring only 4GB RAM and no GPU, TinyTorch demonstrates that deep ML systems understanding is achievable without specialized hardware. The curriculum is available open-source at mlsysbook.ai/tinytorch.
Why we are recommending this paper?
Because deep learning is a popular topic and you have less than 3 interests with available recommendations

This paper directly addresses a critical gap in machine learning education – the lack of understanding of underlying systems. Given your interest in data science career development, this work provides a foundational understanding of how ML systems operate, which is essential for informed career choices.
xbenchai
Paper visualization
Rate image: πŸ‘ πŸ‘Ž
AI Insights
  • The results highlight the need for further research in instruction-following tasks and the development of more effective language models. (ML: 0.98)πŸ‘πŸ‘Ž
  • The authors acknowledge that the proposed benchmark may not capture all aspects of real-world instruction-following tasks. (ML: 0.98)πŸ‘πŸ‘Ž
  • The synthetic data generation approach relies on human-made questions and may not accurately reflect real-world scenarios. (ML: 0.96)πŸ‘πŸ‘Ž
  • The paper proposes a new benchmark for evaluating the ability of language models to follow complex instructions and perform tasks that require multiple steps. (ML: 0.96)πŸ‘πŸ‘Ž
  • The proposed benchmark provides a more comprehensive evaluation of language models' ability to follow instructions and perform tasks. (ML: 0.96)πŸ‘πŸ‘Ž
  • Synthetic data generation: The process of creating artificial data that mimics real-world scenarios, used to train and evaluate language models. (ML: 0.96)πŸ‘πŸ‘Ž
  • The results show that state-of-the-art language models struggle to perform well on this benchmark, highlighting the need for more research in this area. (ML: 0.95)πŸ‘πŸ‘Ž
  • The authors introduce a novel approach to generating synthetic data for instruction-following tasks, which allows them to create a large-scale dataset with diverse and realistic scenarios. (ML: 0.94)πŸ‘πŸ‘Ž
  • Instruction-following task: A task that requires a model to follow a set of instructions to complete a specific goal or achieve a certain outcome. (ML: 0.93)πŸ‘πŸ‘Ž
  • The synthetic data generation approach allows for the creation of diverse and realistic scenarios, making it easier to train and evaluate language models. (ML: 0.93)πŸ‘πŸ‘Ž
Abstract
The capacity of AI agents to effectively handle tasks of increasing duration and complexity continues to grow, demonstrating exceptional performance in coding, deep research, and complex problem-solving evaluations. However, in daily scenarios, the perception of these advanced AI capabilities among general users remains limited. We argue that current evaluations prioritize increasing task difficulty without sufficiently addressing the diversity of agentic tasks necessary to cover the daily work, life, and learning activities of a broad demographic. To address this, we propose AgentIF-OneDay, aimed at determining whether general users can utilize natural language instructions and AI agents to complete a diverse array of daily tasks. These tasks require not only solving problems through dialogue but also understanding various attachment types and delivering tangible file-based results. The benchmark is structured around three user-centric categories: Open Workflow Execution, which assesses adherence to explicit and complex workflows; Latent Instruction, which requires agents to infer implicit instructions from attachments; and Iterative Refinement, which involves modifying or expanding upon ongoing work. We employ instance-level rubrics and a refined evaluation pipeline that aligns LLM-based verification with human judgment, achieving an 80.1% agreement rate using Gemini-3-Pro. AgentIF-OneDay comprises 104 tasks covering 767 scoring points. We benchmarked four leading general AI agents and found that agent products built based on APIs and ChatGPT agents based on agent RL remain in the first tier simultaneously. Leading LLM APIs and open-source models have internalized agentic capabilities, enabling AI application teams to develop cutting-edge Agent products.
Why we are recommending this paper?
Because ai agents is a popular topic and you have less than 3 interests with available recommendations

This research focuses on the development of AI agents capable of handling complex tasks, aligning with the evolving landscape of AI and its potential impact on various industries. Understanding agent-based systems is increasingly relevant for those pursuing careers in data science and AI.
University of Tartu
Paper visualization
Rate image: πŸ‘ πŸ‘Ž
AI Insights
  • It emphasizes the need for a more comprehensive understanding of the complex relationships between technological, social, and economic factors in AI development and deployment. (ML: 0.99)πŸ‘πŸ‘Ž
  • It cites studies on AI bias, transparency, accountability, and the need for responsible AI development and deployment. (ML: 0.99)πŸ‘πŸ‘Ž
  • The study acknowledges that it has limitations due to the complexity of the topic and the need for further research. (ML: 0.98)πŸ‘πŸ‘Ž
  • It also recognizes that the development of AI governance frameworks is a dynamic process that requires ongoing evaluation and refinement. (ML: 0.98)πŸ‘πŸ‘Ž
  • The study highlights the importance of considering paradoxes when developing AI governance frameworks. (ML: 0.97)πŸ‘πŸ‘Ž
  • The paper explores these kinds of complexities in AI governance and why they need to be considered when developing frameworks for responsible AI development and deployment. (ML: 0.97)πŸ‘πŸ‘Ž
  • The paper explores the concept of paradox in the context of artificial intelligence (AI) and its governance, highlighting the importance of considering these complexities when developing AI governance frameworks. (ML: 0.97)πŸ‘πŸ‘Ž
  • The paper discusses the concept of paradox in management and organization theories, specifically focusing on artificial intelligence (AI) and its governance. (ML: 0.96)πŸ‘πŸ‘Ž
  • The paper draws on existing literature in management, organization theory, and AI ethics to inform its discussion of paradoxes in AI governance. (ML: 0.95)πŸ‘πŸ‘Ž
  • Imagine you're trying to create a system that can make decisions without being biased, but at the same time, you want it to be transparent so people understand how those decisions are made. (ML: 0.93)πŸ‘πŸ‘Ž
  • That's a paradox! (ML: 0.91)πŸ‘πŸ‘Ž
  • Paradox: a situation or condition that is contradictory or opposite to what would be expected. (ML: 0.86)πŸ‘πŸ‘Ž
Abstract
The rapid proliferation of artificial intelligence across organizational contexts has generated profound strategic opportunities while introducing significant ethical and operational risks. Despite growing scholarly attention to responsible AI, extant literature remains fragmented and is often adopting either an optimistic stance emphasizing value creation or an excessively cautious perspective fixated on potential harms. This paper addresses this gap by presenting a comprehensive examination of AI's dual nature through the lens of strategic information systems. Drawing upon a systematic synthesis of the responsible AI literature and grounded in paradox theory, we develop the Paradox-based Responsible AI Governance (PRAIG) framework that articulates: (1) the strategic benefits of AI adoption, (2) the inherent risks and unintended consequences, and (3) governance mechanisms that enable organizations to navigate these tensions. Our framework advances theoretical understanding by conceptualizing responsible AI governance as the dynamic management of paradoxical tensions between value creation and risk mitigation. We provide formal propositions demonstrating that trade-off approaches amplify rather than resolve these tensions, and we develop a taxonomy of paradox management strategies with specified contingency conditions. For practitioners, we offer actionable guidance for developing governance structures that neither stifle innovation nor expose organizations to unacceptable risks. The paper concludes with a research agenda for advancing responsible AI governance scholarship.
Why we are recommending this paper?
Because ai agents is a popular topic and you have less than 3 interests with available recommendations

Given your interest in data careers, it's crucial to understand the ethical considerations surrounding AI development and deployment. This paper offers a valuable perspective on responsible AI practices, a key area for future data professionals.
University of Bochum
AI Insights
  • The concept of human-centered AI may be challenging to implement in organizations with rigid structures or cultures. (ML: 0.98)πŸ‘πŸ‘Ž
  • Collaboration between experts, 7. (ML: 0.98)πŸ‘πŸ‘Ž
  • Human-centered AI may not be suitable for all types of tasks or industries, requiring careful consideration and evaluation. (ML: 0.98)πŸ‘πŸ‘Ž
  • Continuous learning and improvement, 2. (ML: 0.98)πŸ‘πŸ‘Ž
  • Monitoring and evaluation, and 10. (ML: 0.97)πŸ‘πŸ‘Ž
  • Clear roles and responsibilities, 5. (ML: 0.97)πŸ‘πŸ‘Ž
  • Effective communication, 6. (ML: 0.96)πŸ‘πŸ‘Ž
  • The concept of keeping the organization in the loop is crucial for the successful implementation of human-centered AI. (ML: 0.96)πŸ‘πŸ‘Ž
  • Interacting organizational practices require significant resources and effort to establish and maintain. (ML: 0.96)πŸ‘πŸ‘Ž
  • Ten types of interacting organizational practices are identified as essential to accompany human-centered AI: 1. (ML: 0.96)πŸ‘πŸ‘Ž
  • Case B substantiates this concept by highlighting the collaboration between technical and analytical experts, anchored in systematic communication structures. (ML: 0.96)πŸ‘πŸ‘Ž
  • Interacting organizational practices are essential to accompany human-centered AI and ensure its effectiveness and adaptability. (ML: 0.95)πŸ‘πŸ‘Ž
  • Adaptation processes, 8. (ML: 0.95)πŸ‘πŸ‘Ž
  • Human-centered AI: An approach to keeping the human in the loop by emphasizing the importance of interacting organizational practices. (ML: 0.95)πŸ‘πŸ‘Ž
  • The concept of keeping the organization in the loop is developed based on case A, which emphasizes the importance of human-centered AI and the need for interacting organizational practices. (ML: 0.95)πŸ‘πŸ‘Ž
  • Interacting organizational practices: The essential types of practices that need to accompany human-centered AI, including continuous learning and improvement, feedback mechanisms, regular meetings and updates, clear roles and responsibilities, effective communication, collaboration between experts, adaptation processes, documentation and knowledge management, monitoring and evaluation, and continuous refinement of the AI system. (ML: 0.94)πŸ‘πŸ‘Ž
  • Feedback mechanisms, 3. (ML: 0.93)πŸ‘πŸ‘Ž
  • Continuous refinement of the AI system. (ML: 0.91)πŸ‘πŸ‘Ž
  • Documentation and knowledge management, 9. (ML: 0.89)πŸ‘πŸ‘Ž
  • Regular meetings and updates, 4. (ML: 0.89)πŸ‘πŸ‘Ž
Abstract
This contribution explores how the integration of Artificial Intelligence (AI) into organizational practices can be effectively framed through a socio-technical perspective to comply with the requirements of Human-centered AI (HCAI). Instead of viewing AI merely as a technical tool, the analysis emphasizes the importance of embedding AI into communication, collaboration, and decision-making processes within organizations from a human-centered perspective. Ten case-based patterns illustrate how AI support of predictive maintenance can be organized to address quality assurance and continuous improvement and to provide different types of sup-port for HCAI. The analysis shows that AI adoption often requires and enables new forms of organizational learning, where specialists jointly interpret AI output, adapt workflows, and refine rules for system improve-ment. Different dimensions and levels of socio-technical integration of AI are considered to reflect the effort and benefits of keeping the organization in the loop.
Why we are recommending this paper?
Because ai and society is a popular topic and you have less than 3 interests with available recommendations

This paper explores the integration of AI into organizational contexts, which is a growing area of focus for those considering careers in data science and strategy. Understanding how AI impacts workflows is vital for successful career transitions.
NYU Grossman School of Medicine
AI Insights
  • The dataset includes images of skin lesions, which are annotated with labels indicating the type of lesion and its severity. (ML: 0.94)πŸ‘πŸ‘Ž
  • The dataset can be used for training machine learning models to diagnose skin conditions from images. (ML: 0.91)πŸ‘πŸ‘Ž
  • Digital pathology: The practice of using digital tools and techniques to analyze and interpret histopathological images. (ML: 0.91)πŸ‘πŸ‘Ž
  • The dataset is available for download and use by researchers and clinicians. (ML: 0.91)πŸ‘πŸ‘Ž
  • The preprocessed dataset ensures that patient privacy is maintained while still allowing for meaningful analysis and annotation. (ML: 0.90)πŸ‘πŸ‘Ž
  • The dataset is a valuable resource for researchers and clinicians working in the field of dermatopathology. (ML: 0.90)πŸ‘πŸ‘Ž
  • The dataset has been preprocessed to remove identifying information and ensure patient privacy. (ML: 0.89)πŸ‘πŸ‘Ž
  • The dataset is a collection of images from various sources, including publicly available datasets and private collections. (ML: 0.86)πŸ‘πŸ‘Ž
  • De-identification: The process of removing identifying information from medical images to ensure patient privacy. (ML: 0.84)πŸ‘πŸ‘Ž
  • Whole-slide imaging (WSI): A technique that captures high-resolution images of entire slides, allowing for detailed analysis and annotation. (ML: 0.80)πŸ‘πŸ‘Ž
Abstract
Accessing high-quality, open-access dermatopathology image datasets for learning and cross-referencing is a common challenge for clinicians and dermatopathology trainees. To establish a comprehensive open-access dermatopathology dataset for educational, cross-referencing, and machine-learning purposes, we employed a hybrid workflow to curate and categorize images from the PubMed Central (PMC) repository. We used specific keywords to extract relevant images, and classified them using a novel hybrid method that combined deep learning-based image modality classification with figure caption analyses. Validation on 651 manually annotated images demonstrated the robustness of our workflow, with an F-score of 89.6\% for the deep learning approach, 61.0\% for the keyword-based retrieval method, and 90.4\% for the hybrid approach. We retrieved over 7,772 images across 166 diagnoses and released this fully annotated dataset, reviewed by board-certified dermatopathologists. Using our dataset as a challenging task, we found the current image analysis algorithm from OpenAI inadequate for analyzing dermatopathology images. In conclusion, we have developed a large, peer-reviewed, open-access dermatopathology image dataset, DermpathNet, which features a semi-automated curation workflow.
Why we are recommending this paper?
Because research automation with ai is a popular topic and you have less than 3 interests with available recommendations

This paper explores the application of AI in medical diagnostics, a field with significant growth potential for data scientists. The focus on dermatopathology demonstrates how AI is transforming healthcare, a domain of increasing interest for career-focused individuals.
University of WisconsinMadison
Paper visualization
Rate image: πŸ‘ πŸ‘Ž
AI Insights
  • A generic workflow for EduDB's benchmark has been introduced to evaluate and rank students' optimizations. (ML: 0.93)πŸ‘πŸ‘Ž
  • EduDB: A simple database prototype designed for educational purposes. (ML: 0.92)πŸ‘πŸ‘Ž
  • Parser: The component responsible for parsing SQL queries into an abstract syntax tree (AST). (ML: 0.92)πŸ‘πŸ‘Ž
  • EduDB is a simple database prototype designed for educational purposes. (ML: 0.91)πŸ‘πŸ‘Ž
  • The system consists of several components, including the parser, query executor, buffer manager, file manager, concurrency manager, and transaction manager. (ML: 0.89)πŸ‘πŸ‘Ž
  • Query Executor: The component that executes the parsed query on the database. (ML: 0.84)πŸ‘πŸ‘Ž
  • File Manager: Responsible for managing files and directories on disk. (ML: 0.81)πŸ‘πŸ‘Ž
  • Buffer Manager: Manages the buffer cache, which stores frequently accessed data in memory. (ML: 0.80)πŸ‘πŸ‘Ž
  • Concurrency Manager: Ensures that multiple transactions can be executed concurrently without conflicts. (ML: 0.79)πŸ‘πŸ‘Ž
  • Transaction Manager: Manages the execution of transactions, including commit and rollback operations. (ML: 0.76)πŸ‘πŸ‘Ž
Abstract
Database Management System (DBMS) is designed to help store and process large collections of data, and is incredibly flexible to perform various kinds of optimizations as long as it achieves serializability with a high-level interface available. The current undergraduate level DBMS course in UW-Madison (i.e., CS564) involves implementing specific modules of DB architecture, including B+ tree, but students may end up spending numerous amounts of effort on corner cases and not gaining a more comprehensive understanding of the internal design. Thus, we present EduDB, a simple database prototype for educational purposes that provides students a clean, concise, and comprehensive overview of the database system. We also attempt to develop an integrative series of course projects based on EduDB, which offers a platform for students to perform any optimization learned during the semester.
Why we are recommending this paper?
Due to your Interest in Data Career Development
Radboud University
AI Insights
  • Value-Sensitive Design (VSD) can help the community reflect on how genAI tools affect learning, expertise development, and knowledge passing, and make choices about tools, incentives, and evaluation practices with explicit attention to agency, responsibility, and learning. (ML: 0.99)πŸ‘πŸ‘Ž
  • The community should support practices that preserve technical grounding, encourage verification, and maintain human responsibility, rather than relying solely on efficiency and automation. (ML: 0.98)πŸ‘πŸ‘Ž
  • Researchers are losing technical grounding and becoming dependent on automated systems, which can lead to errors and accountability issues. (ML: 0.98)πŸ‘πŸ‘Ž
  • A renewed emphasis on first-principles thinking, verification literacy, and resilience in research practice is necessary to ensure that researchers remain responsible for their outputs and capable of intervening when AI tools fail. (ML: 0.98)πŸ‘πŸ‘Ž
  • The software engineering research community must prioritize human responsibility and accountability in the face of widespread AI assistance. (ML: 0.98)πŸ‘πŸ‘Ž
  • Resilience in Research Practice: The capacity to reason about and solve problems without continuous reliance on automated systems. (ML: 0.97)πŸ‘πŸ‘Ž
  • Value-Sensitive Design (VSD): A research and design approach that explicitly accounts for human values throughout the design, development, and use of technology. (ML: 0.97)πŸ‘πŸ‘Ž
  • Verification Literacy: The ability to audit and test AI-generated outputs in a rigorous way, requiring deeper technical understanding. (ML: 0.97)πŸ‘πŸ‘Ž
  • The software engineering research community is at a crossroads as AI-based tools become more prevalent in their work. (ML: 0.96)πŸ‘πŸ‘Ž
  • Design Fiction: A narrative technique used to explore the implications of emerging technologies on society. (ML: 0.92)πŸ‘πŸ‘Ž
Abstract
Rising publication pressure and the routine use of generative AI tools are reshaping how software engineering research is produced, assessed, and taught. While these developments promise efficiency, they also raise concerns about skill degradation, responsibility, and trust in scholarly outputs. This vision paper employs Design Fiction as a methodological lens to examine how such concerns might materialise if current practices persist. Drawing on themes reported in a recent community survey, we construct a speculative artifact situated in a near future research setting. The fiction is used as an analytical device rather than a forecast, enabling reflection on how automated assistance might impede domain knowledge competence, verification, and mentoring practices. By presenting an intentionally unsettling scenario, the paper invites discussion on how the software engineering research community in the future will define proficiency, allocate responsibility, and support learning.
Why we are recommending this paper?
Because research automation with ai is a popular topic and you have less than 3 interests with available recommendations
Syracuse University
Paper visualization
Rate image: πŸ‘ πŸ‘Ž
AI Insights
  • The framework's performance may be affected by the quality of input data and the complexity of the query. (ML: 0.95)πŸ‘πŸ‘Ž
  • The benchmark evaluation only considers a limited set of intelligence modes, which may not fully represent the range of possible designs. (ML: 0.93)πŸ‘πŸ‘Ž
  • Agentic AI: An artificial intelligence that can act on behalf of humans to achieve specific goals. (ML: 0.91)πŸ‘πŸ‘Ž
  • Previous studies have shown that agentic AI can improve building energy management through optimized control policies and coordinated control strategies. (ML: 0.85)πŸ‘πŸ‘Ž
  • The benchmark evaluation highlights the importance of intelligence mode design in achieving balanced performance, with centralized two-stage mode providing the most reliable results while maintaining low execution latency and inference cost. (ML: 0.85)πŸ‘πŸ‘Ž
  • The results show the impact of system upgrades on thermal and electrical domain performance. (ML: 0.83)πŸ‘πŸ‘Ž
  • The proposed agentic AI framework demonstrates its ability to execute complex building energy management queries through sequential simulations. (ML: 0.81)πŸ‘πŸ‘Ž
  • The proposed agentic AI framework executes a realistic building energy management query through three sequential simulations: baseline configuration, system upgrade, and system upgrade with control upgrade. (ML: 0.78)πŸ‘πŸ‘Ž
  • MCP (Model-Component Platform): A platform for integrating models and components. (ML: 0.76)πŸ‘πŸ‘Ž
  • DER (Distributed Energy Resources): Devices or systems that generate, store, or manage energy locally. (ML: 0.75)πŸ‘πŸ‘Ž
  • PIML (Physical Information Modeling Language): A tool for modeling and simulating physical systems. (ML: 0.61)πŸ‘πŸ‘Ž
Abstract
The urgent need for building decarbonization calls for a paradigm shift in future autonomous building energy operation, from human-intensive engineering workflows toward intelligent agents that interact with physics-grounded digital environments. This study proposes an end-to-end agentic AI-enabled Physics-Informed Machine Learning (PIML) environment for scalable building energy modeling, simulation, control, and automation. The framework consists of (1) a modular and physics-consistent PIML digital environment spanning building thermal dynamics, Heating, Ventilation, and Air Conditioning (HVAC), and distributed energy resources (DER) for grid-interactive energy management; and (2) an agentic AI layer with 11 specialist agents and 72 Model Context Protocol (MCP) tools that enable end-to-end execution of multi-step energy analytics. A representative case study demonstrates multi-domain, multi-agent coordination for assessing how system and control upgrades affect energy use, operating cost, thermal comfort, and flexibility. In addition, a large-scale benchmark (about 4000 runs) systematically evaluates workflow performance in terms of accuracy, token consumption, execution time, and inference cost. The results quantify the impacts of intelligence mode design, model size, task complexity, and orchestrator-specialist coordination, and provide key lessons for building future agentic AI systems in real-world building energy applications. This work establishes a scalable, physics-grounded foundation for deploying agentic AI in decarbonized and grid-interactive building operations.
Why we are recommending this paper?
Because agi: artificial general intelligence is a popular topic and you have less than 3 interests with available recommendations
Universit Paris Cit , CNRS
AI Insights
  • The authors show that any feedforward network with ReLU activations can be viewed as a place-independent IFS, and they extend this result to other types of neural networks, including residual blocks and MoE models. (ML: 0.92)πŸ‘πŸ‘Ž
  • The paper discusses the interpretation of deep neural networks as iterated function systems (IFSs) and provides a general framework for analyzing their convergence properties. (ML: 0.89)πŸ‘πŸ‘Ž
  • The paper provides several examples of neural network architectures that can be interpreted as IFSs, including ResNet with Softplus activation, Transformer block, and MoE model. (ML: 0.88)πŸ‘πŸ‘Ž
  • The authors use the Hutchinson operator to analyze the convergence properties of IFSs and show that they can be used to bound the Wasserstein distance between the output of a neural network and its fixed point. (ML: 0.87)πŸ‘πŸ‘Ž
  • Definition 1: A Markov recursion is a sequence of random variables {Xn} defined by X0 = x and Xt+1 = w(Xt, Θ), where w is a function that depends on the current state Xt and the parameter Θ. (ML: 0.82)πŸ‘πŸ‘Ž
  • Definition 3: A place-dependent IFS (P-IFS) is an IFS {wΞΎ} where each wΞΎ depends on the current state x and the parameter Θ. (ML: 0.80)πŸ‘πŸ‘Ž
  • Definition 2: An iterated function system (IFS) is a collection of functions {wΞΎ} indexed by ΞΎ ∈ I, where each wΞΎ is a Lipschitz map from X to itself. (ML: 0.80)πŸ‘πŸ‘Ž
  • They also introduce the concept of strong average Lipschitz contractivity for place-dependent IFSs and provide conditions under which it holds. (ML: 0.75)πŸ‘πŸ‘Ž
  • Definition 5: A P-IFS {wΞΎ} is strongly average-contractive if sup_x∈X βˆ‘_{ξ∈I} pΞΎ(x)cΞΎ ≀ c < 1. (ML: 0.67)πŸ‘πŸ‘Ž
  • Definition 4: The Hutchinson operator T is a contraction on the space of probability measures PP(X) with respect to the Wasserstein distance W2 if there exists a constant c < 1 such that W2(T(Β΅), T(Ξ½)) ≀ cW2(Β΅, Ξ½) for all Β΅, Ξ½ ∈ PP(X). (ML: 0.67)πŸ‘πŸ‘Ž
  • The Hutchinson operator T is defined as T(Β΅) = βˆ‘_{ξ∈I} pwΞΎ#Β΅q. (ML: 0.49)πŸ‘πŸ‘Ž
Abstract
Deep neural networks (DNNs) achieve remarkable performance on a wide range of tasks, yet their mathematical analysis remains fragmented: stability and generalization are typically studied in disparate frameworks and on a case-by-case basis. Architecturally, DNNs rely on the recursive application of parametrized functions, a mechanism that can be unstable and difficult to train, making stability a primary concern. Even when training succeeds, there are few rigorous results on how well such models generalize beyond the observed data, especially in the generative setting. In this work, we leverage the theory of stochastic Iterated Function Systems (IFS) and show that two important deep architectures can be viewed as, or canonically associated with, place-dependent IFS. This connection allows us to import results from random dynamical systems to (i) establish the existence and uniqueness of invariant measures under suitable contractivity assumptions, and (ii) derive a Wasserstein generalization bound for generative modeling. The bound naturally leads to a new training objective that directly controls the collage-type approximation error between the data distribution and its image under the learned transfer operator. We illustrate the theory on a controlled 2D example and empirically evaluate the proposed objective on standard image datasets (MNIST, CelebA, CIFAR-10).
Why we are recommending this paper?
Because deep learning is a popular topic and you have less than 3 interests with available recommendations
Southeast University
AI Insights
  • Artifact-aware evaluation framework: A comprehensive framework for evaluating video generation models based on their ability to detect and correct artifacts. (ML: 0.94)πŸ‘πŸ‘Ž
  • The proposed framework provides a comprehensive evaluation of video generation models, enabling researchers to identify areas for improvement. (ML: 0.94)πŸ‘πŸ‘Ž
  • Video quality assessment methods: Techniques for evaluating the quality of generated videos based on various metrics, such as peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM). (ML: 0.93)πŸ‘πŸ‘Ž
  • The novel DV AR (Dense Video Artifact Recognition) framework is proposed, which leverages the strengths of both text-to-video diffusion models and video quality assessment methods. (ML: 0.91)πŸ‘πŸ‘Ž
  • The paper proposes a comprehensive artifact-aware evaluation framework for video generation, focusing on fine-grained artifact detection across Appearance, Motion, and Camera dimensions. (ML: 0.91)πŸ‘πŸ‘Ž
  • Experimental results demonstrate the accuracy of the approach in identifying artifacts in generated videos, with a significant improvement over state-of-the-art methods. (ML: 0.90)πŸ‘πŸ‘Ž
  • Text-to-video diffusion models: A type of deep learning model that generates videos from text descriptions. (ML: 0.88)πŸ‘πŸ‘Ž
  • Fine-grained artifact detection: The process of detecting specific types of artifacts in videos, such as noise, blurriness, or color distortion. (ML: 0.87)πŸ‘πŸ‘Ž
  • The large-scale GenVID dataset is introduced, which includes a wide range of videos with diverse content, styles, and quality levels. (ML: 0.87)πŸ‘πŸ‘Ž
  • The FMG-DFS strategy enhances temporal localization, enabling more precise and efficient artifact detection. (ML: 0.68)πŸ‘πŸ‘Ž
Abstract
With the rapid advancement of video generation techniques, evaluating and auditing generated videos has become increasingly crucial. Existing approaches typically offer coarse video quality scores, lacking detailed localization and categorization of specific artifacts. In this work, we introduce a comprehensive evaluation protocol focusing on three key aspects affecting human perception: Appearance, Motion, and Camera. We define these axes through a taxonomy of 10 prevalent artifact categories reflecting common generative failures observed in video generation. To enable robust artifact detection and categorization, we introduce GenVID, a large-scale dataset of 80k videos generated by various state-of-the-art video generation models, each carefully annotated for the defined artifact categories. Leveraging GenVID, we develop DVAR, a Dense Video Artifact Recognition framework for fine-grained identification and classification of generative artifacts. Extensive experiments show that our approach significantly improves artifact detection accuracy and enables effective filtering of low-quality content.
Why we are recommending this paper?
Because image and video generation is a popular topic and you have less than 3 interests with available recommendations
Rutgers University
AI Insights
  • Generative models: A class of machine learning algorithms that can generate new data samples based on a given dataset. (ML: 0.97)πŸ‘πŸ‘Ž
  • The paper assumes that the input text is well-formed and does not contain any errors or ambiguities. (ML: 0.97)πŸ‘πŸ‘Ž
  • Diffusion prior constraints: A type of regularization technique used in generative models that encourages the model to produce samples that are similar to the data distribution. (ML: 0.92)πŸ‘πŸ‘Ž
  • The paper discusses the concept of ConceptLab, a creative generation system that uses diffusion prior constraints to generate novel images. (ML: 0.90)πŸ‘πŸ‘Ž
  • ConceptLab is based on a combination of generative models and text-image harmony, which allows it to produce high-quality images with specific styles and concepts. (ML: 0.88)πŸ‘πŸ‘Ž
  • ConceptLab is based on a combination of generative models and text-image harmony, which allows it to produce high-quality images with specific styles and concepts. (ML: 0.88)πŸ‘πŸ‘Ž
  • The paper presents ConceptLab, a creative generation system that uses diffusion prior constraints to generate novel images. (ML: 0.86)πŸ‘πŸ‘Ž
  • The authors propose a new method for training generative models using diffusion prior constraints, which enables the model to learn from the data distribution and generate novel samples. (ML: 0.84)πŸ‘πŸ‘Ž
  • The authors propose a new method for training generative models using diffusion prior constraints, which enables the model to learn from the data distribution and generate novel samples. (ML: 0.84)πŸ‘πŸ‘Ž
  • Text-image harmony: A method for combining text and image features to generate images with specific styles and concepts. (ML: 0.76)πŸ‘πŸ‘Ž
Abstract
Creative image generation has emerged as a compelling area of research, driven by the need to produce novel and high-quality images that expand the boundaries of imagination. In this work, we propose a novel framework for creative generation using diffusion models, where creativity is associated with the inverse probability of an image's existence in the CLIP embedding space. Unlike prior approaches that rely on a manual blending of concepts or exclusion of subcategories, our method calculates the probability distribution of generated images and drives it towards low-probability regions to produce rare, imaginative, and visually captivating outputs. We also introduce pullback mechanisms, achieving high creativity without sacrificing visual fidelity. Extensive experiments on text-to-image diffusion models demonstrate the effectiveness and efficiency of our creative generation framework, showcasing its ability to produce unique, novel, and thought-provoking images. This work provides a new perspective on creativity in generative models, offering a principled method to foster innovation in visual content synthesis.
Why we are recommending this paper?
Because image and video generation is a popular topic and you have less than 3 interests with available recommendations

We did not find tons of content matching your interests we've included some additional topics that are popular. Also be aware that if the topics is not present in arxiv we wont be able to recommend it.

University of Tartu
Paper visualization
Rate image: πŸ‘ πŸ‘Ž
AI Insights
  • It emphasizes the need for a more comprehensive understanding of the complex relationships between technological, social, and economic factors in AI development and deployment. (ML: 0.99)πŸ‘πŸ‘Ž
  • It cites studies on AI bias, transparency, accountability, and the need for responsible AI development and deployment. (ML: 0.99)πŸ‘πŸ‘Ž
  • The study acknowledges that it has limitations due to the complexity of the topic and the need for further research. (ML: 0.98)πŸ‘πŸ‘Ž
  • It also recognizes that the development of AI governance frameworks is a dynamic process that requires ongoing evaluation and refinement. (ML: 0.98)πŸ‘πŸ‘Ž
  • The study highlights the importance of considering paradoxes when developing AI governance frameworks. (ML: 0.97)πŸ‘πŸ‘Ž
  • The paper explores these kinds of complexities in AI governance and why they need to be considered when developing frameworks for responsible AI development and deployment. (ML: 0.97)πŸ‘πŸ‘Ž
  • The paper explores the concept of paradox in the context of artificial intelligence (AI) and its governance, highlighting the importance of considering these complexities when developing AI governance frameworks. (ML: 0.97)πŸ‘πŸ‘Ž
  • The paper discusses the concept of paradox in management and organization theories, specifically focusing on artificial intelligence (AI) and its governance. (ML: 0.96)πŸ‘πŸ‘Ž
  • The paper draws on existing literature in management, organization theory, and AI ethics to inform its discussion of paradoxes in AI governance. (ML: 0.95)πŸ‘πŸ‘Ž
  • Imagine you're trying to create a system that can make decisions without being biased, but at the same time, you want it to be transparent so people understand how those decisions are made. (ML: 0.93)πŸ‘πŸ‘Ž
  • That's a paradox! (ML: 0.91)πŸ‘πŸ‘Ž
  • Paradox: a situation or condition that is contradictory or opposite to what would be expected. (ML: 0.86)πŸ‘πŸ‘Ž
Abstract
The rapid proliferation of artificial intelligence across organizational contexts has generated profound strategic opportunities while introducing significant ethical and operational risks. Despite growing scholarly attention to responsible AI, extant literature remains fragmented and is often adopting either an optimistic stance emphasizing value creation or an excessively cautious perspective fixated on potential harms. This paper addresses this gap by presenting a comprehensive examination of AI's dual nature through the lens of strategic information systems. Drawing upon a systematic synthesis of the responsible AI literature and grounded in paradox theory, we develop the Paradox-based Responsible AI Governance (PRAIG) framework that articulates: (1) the strategic benefits of AI adoption, (2) the inherent risks and unintended consequences, and (3) governance mechanisms that enable organizations to navigate these tensions. Our framework advances theoretical understanding by conceptualizing responsible AI governance as the dynamic management of paradoxical tensions between value creation and risk mitigation. We provide formal propositions demonstrating that trade-off approaches amplify rather than resolve these tensions, and we develop a taxonomy of paradox management strategies with specified contingency conditions. For practitioners, we offer actionable guidance for developing governance structures that neither stifle innovation nor expose organizations to unacceptable risks. The paper concludes with a research agenda for advancing responsible AI governance scholarship.
Why we are recommending this paper?
Because ai agents is a popular topic and you have less than 3 interests with available recommendations
xbenchai
AI Insights
  • The results highlight the need for further research in instruction-following tasks and the development of more effective language models. (ML: 0.98)πŸ‘πŸ‘Ž
  • The authors acknowledge that the proposed benchmark may not capture all aspects of real-world instruction-following tasks. (ML: 0.98)πŸ‘πŸ‘Ž
  • The synthetic data generation approach relies on human-made questions and may not accurately reflect real-world scenarios. (ML: 0.96)πŸ‘πŸ‘Ž
  • The paper proposes a new benchmark for evaluating the ability of language models to follow complex instructions and perform tasks that require multiple steps. (ML: 0.96)πŸ‘πŸ‘Ž
  • The proposed benchmark provides a more comprehensive evaluation of language models' ability to follow instructions and perform tasks. (ML: 0.96)πŸ‘πŸ‘Ž
  • Synthetic data generation: The process of creating artificial data that mimics real-world scenarios, used to train and evaluate language models. (ML: 0.96)πŸ‘πŸ‘Ž
  • The results show that state-of-the-art language models struggle to perform well on this benchmark, highlighting the need for more research in this area. (ML: 0.95)πŸ‘πŸ‘Ž
  • The authors introduce a novel approach to generating synthetic data for instruction-following tasks, which allows them to create a large-scale dataset with diverse and realistic scenarios. (ML: 0.94)πŸ‘πŸ‘Ž
  • Instruction-following task: A task that requires a model to follow a set of instructions to complete a specific goal or achieve a certain outcome. (ML: 0.93)πŸ‘πŸ‘Ž
  • The synthetic data generation approach allows for the creation of diverse and realistic scenarios, making it easier to train and evaluate language models. (ML: 0.93)πŸ‘πŸ‘Ž
Abstract
The capacity of AI agents to effectively handle tasks of increasing duration and complexity continues to grow, demonstrating exceptional performance in coding, deep research, and complex problem-solving evaluations. However, in daily scenarios, the perception of these advanced AI capabilities among general users remains limited. We argue that current evaluations prioritize increasing task difficulty without sufficiently addressing the diversity of agentic tasks necessary to cover the daily work, life, and learning activities of a broad demographic. To address this, we propose AgentIF-OneDay, aimed at determining whether general users can utilize natural language instructions and AI agents to complete a diverse array of daily tasks. These tasks require not only solving problems through dialogue but also understanding various attachment types and delivering tangible file-based results. The benchmark is structured around three user-centric categories: Open Workflow Execution, which assesses adherence to explicit and complex workflows; Latent Instruction, which requires agents to infer implicit instructions from attachments; and Iterative Refinement, which involves modifying or expanding upon ongoing work. We employ instance-level rubrics and a refined evaluation pipeline that aligns LLM-based verification with human judgment, achieving an 80.1% agreement rate using Gemini-3-Pro. AgentIF-OneDay comprises 104 tasks covering 767 scoring points. We benchmarked four leading general AI agents and found that agent products built based on APIs and ChatGPT agents based on agent RL remain in the first tier simultaneously. Leading LLM APIs and open-source models have internalized agentic capabilities, enabling AI application teams to develop cutting-edge Agent products.
Why we are recommending this paper?
Because ai agents is a popular topic and you have less than 3 interests with available recommendations
University of Bochum
AI Insights
  • The concept of human-centered AI may be challenging to implement in organizations with rigid structures or cultures. (ML: 0.98)πŸ‘πŸ‘Ž
  • Collaboration between experts, 7. (ML: 0.98)πŸ‘πŸ‘Ž
  • Human-centered AI may not be suitable for all types of tasks or industries, requiring careful consideration and evaluation. (ML: 0.98)πŸ‘πŸ‘Ž
  • Continuous learning and improvement, 2. (ML: 0.98)πŸ‘πŸ‘Ž
  • Monitoring and evaluation, and 10. (ML: 0.97)πŸ‘πŸ‘Ž
  • Clear roles and responsibilities, 5. (ML: 0.97)πŸ‘πŸ‘Ž
  • Effective communication, 6. (ML: 0.96)πŸ‘πŸ‘Ž
  • The concept of keeping the organization in the loop is crucial for the successful implementation of human-centered AI. (ML: 0.96)πŸ‘πŸ‘Ž
  • Interacting organizational practices require significant resources and effort to establish and maintain. (ML: 0.96)πŸ‘πŸ‘Ž
  • Ten types of interacting organizational practices are identified as essential to accompany human-centered AI: 1. (ML: 0.96)πŸ‘πŸ‘Ž
  • Case B substantiates this concept by highlighting the collaboration between technical and analytical experts, anchored in systematic communication structures. (ML: 0.96)πŸ‘πŸ‘Ž
  • Interacting organizational practices are essential to accompany human-centered AI and ensure its effectiveness and adaptability. (ML: 0.95)πŸ‘πŸ‘Ž
  • Adaptation processes, 8. (ML: 0.95)πŸ‘πŸ‘Ž
  • Human-centered AI: An approach to keeping the human in the loop by emphasizing the importance of interacting organizational practices. (ML: 0.95)πŸ‘πŸ‘Ž
  • The concept of keeping the organization in the loop is developed based on case A, which emphasizes the importance of human-centered AI and the need for interacting organizational practices. (ML: 0.95)πŸ‘πŸ‘Ž
  • Interacting organizational practices: The essential types of practices that need to accompany human-centered AI, including continuous learning and improvement, feedback mechanisms, regular meetings and updates, clear roles and responsibilities, effective communication, collaboration between experts, adaptation processes, documentation and knowledge management, monitoring and evaluation, and continuous refinement of the AI system. (ML: 0.94)πŸ‘πŸ‘Ž
  • Feedback mechanisms, 3. (ML: 0.93)πŸ‘πŸ‘Ž
  • Continuous refinement of the AI system. (ML: 0.91)πŸ‘πŸ‘Ž
  • Documentation and knowledge management, 9. (ML: 0.89)πŸ‘πŸ‘Ž
  • Regular meetings and updates, 4. (ML: 0.89)πŸ‘πŸ‘Ž
Abstract
This contribution explores how the integration of Artificial Intelligence (AI) into organizational practices can be effectively framed through a socio-technical perspective to comply with the requirements of Human-centered AI (HCAI). Instead of viewing AI merely as a technical tool, the analysis emphasizes the importance of embedding AI into communication, collaboration, and decision-making processes within organizations from a human-centered perspective. Ten case-based patterns illustrate how AI support of predictive maintenance can be organized to address quality assurance and continuous improvement and to provide different types of sup-port for HCAI. The analysis shows that AI adoption often requires and enables new forms of organizational learning, where specialists jointly interpret AI output, adapt workflows, and refine rules for system improve-ment. Different dimensions and levels of socio-technical integration of AI are considered to reflect the effort and benefits of keeping the organization in the loop.
Why we are recommending this paper?
Because ai and society is a popular topic and you have less than 3 interests with available recommendations
NYU Grossman School of Medicine
AI Insights
  • The dataset includes images of skin lesions, which are annotated with labels indicating the type of lesion and its severity. (ML: 0.94)πŸ‘πŸ‘Ž
  • The dataset can be used for training machine learning models to diagnose skin conditions from images. (ML: 0.91)πŸ‘πŸ‘Ž
  • Digital pathology: The practice of using digital tools and techniques to analyze and interpret histopathological images. (ML: 0.91)πŸ‘πŸ‘Ž
  • The dataset is available for download and use by researchers and clinicians. (ML: 0.91)πŸ‘πŸ‘Ž
  • The preprocessed dataset ensures that patient privacy is maintained while still allowing for meaningful analysis and annotation. (ML: 0.90)πŸ‘πŸ‘Ž
  • The dataset is a valuable resource for researchers and clinicians working in the field of dermatopathology. (ML: 0.90)πŸ‘πŸ‘Ž
  • The dataset has been preprocessed to remove identifying information and ensure patient privacy. (ML: 0.89)πŸ‘πŸ‘Ž
  • The dataset is a collection of images from various sources, including publicly available datasets and private collections. (ML: 0.86)πŸ‘πŸ‘Ž
  • De-identification: The process of removing identifying information from medical images to ensure patient privacy. (ML: 0.84)πŸ‘πŸ‘Ž
  • Whole-slide imaging (WSI): A technique that captures high-resolution images of entire slides, allowing for detailed analysis and annotation. (ML: 0.80)πŸ‘πŸ‘Ž
Abstract
Accessing high-quality, open-access dermatopathology image datasets for learning and cross-referencing is a common challenge for clinicians and dermatopathology trainees. To establish a comprehensive open-access dermatopathology dataset for educational, cross-referencing, and machine-learning purposes, we employed a hybrid workflow to curate and categorize images from the PubMed Central (PMC) repository. We used specific keywords to extract relevant images, and classified them using a novel hybrid method that combined deep learning-based image modality classification with figure caption analyses. Validation on 651 manually annotated images demonstrated the robustness of our workflow, with an F-score of 89.6\% for the deep learning approach, 61.0\% for the keyword-based retrieval method, and 90.4\% for the hybrid approach. We retrieved over 7,772 images across 166 diagnoses and released this fully annotated dataset, reviewed by board-certified dermatopathologists. Using our dataset as a challenging task, we found the current image analysis algorithm from OpenAI inadequate for analyzing dermatopathology images. In conclusion, we have developed a large, peer-reviewed, open-access dermatopathology image dataset, DermpathNet, which features a semi-automated curation workflow.
Why we are recommending this paper?
Because research automation with ai is a popular topic and you have less than 3 interests with available recommendations
Radboud University
AI Insights
  • Value-Sensitive Design (VSD) can help the community reflect on how genAI tools affect learning, expertise development, and knowledge passing, and make choices about tools, incentives, and evaluation practices with explicit attention to agency, responsibility, and learning. (ML: 0.99)πŸ‘πŸ‘Ž
  • The community should support practices that preserve technical grounding, encourage verification, and maintain human responsibility, rather than relying solely on efficiency and automation. (ML: 0.98)πŸ‘πŸ‘Ž
  • Researchers are losing technical grounding and becoming dependent on automated systems, which can lead to errors and accountability issues. (ML: 0.98)πŸ‘πŸ‘Ž
  • A renewed emphasis on first-principles thinking, verification literacy, and resilience in research practice is necessary to ensure that researchers remain responsible for their outputs and capable of intervening when AI tools fail. (ML: 0.98)πŸ‘πŸ‘Ž
  • The software engineering research community must prioritize human responsibility and accountability in the face of widespread AI assistance. (ML: 0.98)πŸ‘πŸ‘Ž
  • Resilience in Research Practice: The capacity to reason about and solve problems without continuous reliance on automated systems. (ML: 0.97)πŸ‘πŸ‘Ž
  • Value-Sensitive Design (VSD): A research and design approach that explicitly accounts for human values throughout the design, development, and use of technology. (ML: 0.97)πŸ‘πŸ‘Ž
  • Verification Literacy: The ability to audit and test AI-generated outputs in a rigorous way, requiring deeper technical understanding. (ML: 0.97)πŸ‘πŸ‘Ž
  • The software engineering research community is at a crossroads as AI-based tools become more prevalent in their work. (ML: 0.96)πŸ‘πŸ‘Ž
  • Design Fiction: A narrative technique used to explore the implications of emerging technologies on society. (ML: 0.92)πŸ‘πŸ‘Ž
Abstract
Rising publication pressure and the routine use of generative AI tools are reshaping how software engineering research is produced, assessed, and taught. While these developments promise efficiency, they also raise concerns about skill degradation, responsibility, and trust in scholarly outputs. This vision paper employs Design Fiction as a methodological lens to examine how such concerns might materialise if current practices persist. Drawing on themes reported in a recent community survey, we construct a speculative artifact situated in a near future research setting. The fiction is used as an analytical device rather than a forecast, enabling reflection on how automated assistance might impede domain knowledge competence, verification, and mentoring practices. By presenting an intentionally unsettling scenario, the paper invites discussion on how the software engineering research community in the future will define proficiency, allocate responsibility, and support learning.
Why we are recommending this paper?
Because research automation with ai is a popular topic and you have less than 3 interests with available recommendations
Syracuse University
AI Insights
  • The framework's performance may be affected by the quality of input data and the complexity of the query. (ML: 0.95)πŸ‘πŸ‘Ž
  • The benchmark evaluation only considers a limited set of intelligence modes, which may not fully represent the range of possible designs. (ML: 0.93)πŸ‘πŸ‘Ž
  • Agentic AI: An artificial intelligence that can act on behalf of humans to achieve specific goals. (ML: 0.91)πŸ‘πŸ‘Ž
  • Previous studies have shown that agentic AI can improve building energy management through optimized control policies and coordinated control strategies. (ML: 0.85)πŸ‘πŸ‘Ž
  • The benchmark evaluation highlights the importance of intelligence mode design in achieving balanced performance, with centralized two-stage mode providing the most reliable results while maintaining low execution latency and inference cost. (ML: 0.85)πŸ‘πŸ‘Ž
  • The results show the impact of system upgrades on thermal and electrical domain performance. (ML: 0.83)πŸ‘πŸ‘Ž
  • The proposed agentic AI framework demonstrates its ability to execute complex building energy management queries through sequential simulations. (ML: 0.81)πŸ‘πŸ‘Ž
  • The proposed agentic AI framework executes a realistic building energy management query through three sequential simulations: baseline configuration, system upgrade, and system upgrade with control upgrade. (ML: 0.78)πŸ‘πŸ‘Ž
  • MCP (Model-Component Platform): A platform for integrating models and components. (ML: 0.76)πŸ‘πŸ‘Ž
  • DER (Distributed Energy Resources): Devices or systems that generate, store, or manage energy locally. (ML: 0.75)πŸ‘πŸ‘Ž
  • PIML (Physical Information Modeling Language): A tool for modeling and simulating physical systems. (ML: 0.61)πŸ‘πŸ‘Ž
Abstract
The urgent need for building decarbonization calls for a paradigm shift in future autonomous building energy operation, from human-intensive engineering workflows toward intelligent agents that interact with physics-grounded digital environments. This study proposes an end-to-end agentic AI-enabled Physics-Informed Machine Learning (PIML) environment for scalable building energy modeling, simulation, control, and automation. The framework consists of (1) a modular and physics-consistent PIML digital environment spanning building thermal dynamics, Heating, Ventilation, and Air Conditioning (HVAC), and distributed energy resources (DER) for grid-interactive energy management; and (2) an agentic AI layer with 11 specialist agents and 72 Model Context Protocol (MCP) tools that enable end-to-end execution of multi-step energy analytics. A representative case study demonstrates multi-domain, multi-agent coordination for assessing how system and control upgrades affect energy use, operating cost, thermal comfort, and flexibility. In addition, a large-scale benchmark (about 4000 runs) systematically evaluates workflow performance in terms of accuracy, token consumption, execution time, and inference cost. The results quantify the impacts of intelligence mode design, model size, task complexity, and orchestrator-specialist coordination, and provide key lessons for building future agentic AI systems in real-world building energy applications. This work establishes a scalable, physics-grounded foundation for deploying agentic AI in decarbonized and grid-interactive building operations.
Why we are recommending this paper?
Because agi: artificial general intelligence is a popular topic and you have less than 3 interests with available recommendations
Universit Paris Cit , CNRS
AI Insights
  • The authors show that any feedforward network with ReLU activations can be viewed as a place-independent IFS, and they extend this result to other types of neural networks, including residual blocks and MoE models. (ML: 0.92)πŸ‘πŸ‘Ž
  • The paper discusses the interpretation of deep neural networks as iterated function systems (IFSs) and provides a general framework for analyzing their convergence properties. (ML: 0.89)πŸ‘πŸ‘Ž
  • The paper provides several examples of neural network architectures that can be interpreted as IFSs, including ResNet with Softplus activation, Transformer block, and MoE model. (ML: 0.88)πŸ‘πŸ‘Ž
  • The authors use the Hutchinson operator to analyze the convergence properties of IFSs and show that they can be used to bound the Wasserstein distance between the output of a neural network and its fixed point. (ML: 0.87)πŸ‘πŸ‘Ž
  • Definition 1: A Markov recursion is a sequence of random variables {Xn} defined by X0 = x and Xt+1 = w(Xt, Θ), where w is a function that depends on the current state Xt and the parameter Θ. (ML: 0.82)πŸ‘πŸ‘Ž
  • Definition 3: A place-dependent IFS (P-IFS) is an IFS {wΞΎ} where each wΞΎ depends on the current state x and the parameter Θ. (ML: 0.80)πŸ‘πŸ‘Ž
  • Definition 2: An iterated function system (IFS) is a collection of functions {wΞΎ} indexed by ΞΎ ∈ I, where each wΞΎ is a Lipschitz map from X to itself. (ML: 0.80)πŸ‘πŸ‘Ž
  • They also introduce the concept of strong average Lipschitz contractivity for place-dependent IFSs and provide conditions under which it holds. (ML: 0.75)πŸ‘πŸ‘Ž
  • Definition 5: A P-IFS {wΞΎ} is strongly average-contractive if sup_x∈X βˆ‘_{ξ∈I} pΞΎ(x)cΞΎ ≀ c < 1. (ML: 0.67)πŸ‘πŸ‘Ž
  • Definition 4: The Hutchinson operator T is a contraction on the space of probability measures PP(X) with respect to the Wasserstein distance W2 if there exists a constant c < 1 such that W2(T(Β΅), T(Ξ½)) ≀ cW2(Β΅, Ξ½) for all Β΅, Ξ½ ∈ PP(X). (ML: 0.67)πŸ‘πŸ‘Ž
  • The Hutchinson operator T is defined as T(Β΅) = βˆ‘_{ξ∈I} pwΞΎ#Β΅q. (ML: 0.49)πŸ‘πŸ‘Ž
Abstract
Deep neural networks (DNNs) achieve remarkable performance on a wide range of tasks, yet their mathematical analysis remains fragmented: stability and generalization are typically studied in disparate frameworks and on a case-by-case basis. Architecturally, DNNs rely on the recursive application of parametrized functions, a mechanism that can be unstable and difficult to train, making stability a primary concern. Even when training succeeds, there are few rigorous results on how well such models generalize beyond the observed data, especially in the generative setting. In this work, we leverage the theory of stochastic Iterated Function Systems (IFS) and show that two important deep architectures can be viewed as, or canonically associated with, place-dependent IFS. This connection allows us to import results from random dynamical systems to (i) establish the existence and uniqueness of invariant measures under suitable contractivity assumptions, and (ii) derive a Wasserstein generalization bound for generative modeling. The bound naturally leads to a new training objective that directly controls the collage-type approximation error between the data distribution and its image under the learned transfer operator. We illustrate the theory on a controlled 2D example and empirically evaluate the proposed objective on standard image datasets (MNIST, CelebA, CIFAR-10).
Why we are recommending this paper?
Because deep learning is a popular topic and you have less than 3 interests with available recommendations
Harvard University
AI Insights
  • Milestones serve dual pedagogical and validation purposes, providing motivation through historical framing and demonstrating implementation correctness through real-world task performance. (ML: 0.98)πŸ‘πŸ‘Ž
  • Each module concludes with systems reasoning prompts measuring conceptual understanding beyond syntactic correctness. (ML: 0.97)πŸ‘πŸ‘Ž
  • Milestones are designed to be challenging but achievable, allowing students to demonstrate their understanding of complex concepts through real-world tasks. (ML: 0.96)πŸ‘πŸ‘Ž
  • Assessment validates both isolated correctness and cross-module integration. (ML: 0.96)πŸ‘πŸ‘Ž
  • The TinyTorch framework is designed for teaching machine learning concepts through hands-on implementation and analysis. (ML: 0.95)πŸ‘πŸ‘Ž
  • Reflect: Systems Analysis Questions. (ML: 0.94)πŸ‘πŸ‘Ž
  • TinyTorch follows a consistent Build-Use-Reflect cycle, integrating implementation, application, and systems reasoning to address multiple learning objectives. (ML: 0.94)πŸ‘πŸ‘Ž
  • It's a pedagogical tool aimed at bridging the gap between theoretical understanding and practical application. (ML: 0.94)πŸ‘πŸ‘Ž
  • Students implement components in Jupyter notebooks with scaffolded guidance. (ML: 0.91)πŸ‘πŸ‘Ž
  • TinyTorch's design emphasizes systems thinking, encouraging students to analyze and understand the relationships between components, rather than just focusing on individual functions. (ML: 0.87)πŸ‘πŸ‘Ž
  • The framework includes six historical milestones that recreate actual breakthroughs using exclusively student code, validating success through task-appropriate performance. (ML: 0.85)πŸ‘πŸ‘Ž
  • The framework is built with a focus on explicit dependencies, making it easier for students to understand where each module fits in the larger architecture. (ML: 0.83)πŸ‘πŸ‘Ž
  • Use: Integration Testing Beyond Unit Tests. (ML: 0.77)πŸ‘πŸ‘Ž
  • Build: Implementation with Explicit Dependencies. (ML: 0.66)πŸ‘πŸ‘Ž
Abstract
Machine learning education faces a fundamental gap: students learn algorithms without understanding the systems that execute them. They study gradient descent without measuring memory, attention mechanisms without analyzing O(N^2) scaling, optimizer theory without knowing why Adam requires 3x the memory of SGD. This "algorithm-systems divide" produces practitioners who can train models but cannot debug memory failures, optimize inference latency, or reason about deployment trade-offs--the very skills industry demands as "ML systems engineering." We present TinyTorch, a 20-module curriculum that closes this gap through "implementation-based systems pedagogy": students construct PyTorch's core components (tensors, autograd, optimizers, CNNs, transformers) in pure Python, building a complete framework where every operation they invoke is code they wrote. The design employs three patterns: "progressive disclosure" of complexity, "systems-first integration" of profiling from the first module, and "build-to-validate milestones" recreating 67 years of ML breakthroughs--from Perceptron (1958) through Transformers (2017) to MLPerf-style benchmarking. Requiring only 4GB RAM and no GPU, TinyTorch demonstrates that deep ML systems understanding is achievable without specialized hardware. The curriculum is available open-source at mlsysbook.ai/tinytorch.
Why we are recommending this paper?
Because deep learning is a popular topic and you have less than 3 interests with available recommendations
Southeast University
AI Insights
  • Artifact-aware evaluation framework: A comprehensive framework for evaluating video generation models based on their ability to detect and correct artifacts. (ML: 0.94)πŸ‘πŸ‘Ž
  • The proposed framework provides a comprehensive evaluation of video generation models, enabling researchers to identify areas for improvement. (ML: 0.94)πŸ‘πŸ‘Ž
  • Video quality assessment methods: Techniques for evaluating the quality of generated videos based on various metrics, such as peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM). (ML: 0.93)πŸ‘πŸ‘Ž
  • The novel DV AR (Dense Video Artifact Recognition) framework is proposed, which leverages the strengths of both text-to-video diffusion models and video quality assessment methods. (ML: 0.91)πŸ‘πŸ‘Ž
  • The paper proposes a comprehensive artifact-aware evaluation framework for video generation, focusing on fine-grained artifact detection across Appearance, Motion, and Camera dimensions. (ML: 0.91)πŸ‘πŸ‘Ž
  • Experimental results demonstrate the accuracy of the approach in identifying artifacts in generated videos, with a significant improvement over state-of-the-art methods. (ML: 0.90)πŸ‘πŸ‘Ž
  • Text-to-video diffusion models: A type of deep learning model that generates videos from text descriptions. (ML: 0.88)πŸ‘πŸ‘Ž
  • Fine-grained artifact detection: The process of detecting specific types of artifacts in videos, such as noise, blurriness, or color distortion. (ML: 0.87)πŸ‘πŸ‘Ž
  • The large-scale GenVID dataset is introduced, which includes a wide range of videos with diverse content, styles, and quality levels. (ML: 0.87)πŸ‘πŸ‘Ž
  • The FMG-DFS strategy enhances temporal localization, enabling more precise and efficient artifact detection. (ML: 0.68)πŸ‘πŸ‘Ž
Abstract
With the rapid advancement of video generation techniques, evaluating and auditing generated videos has become increasingly crucial. Existing approaches typically offer coarse video quality scores, lacking detailed localization and categorization of specific artifacts. In this work, we introduce a comprehensive evaluation protocol focusing on three key aspects affecting human perception: Appearance, Motion, and Camera. We define these axes through a taxonomy of 10 prevalent artifact categories reflecting common generative failures observed in video generation. To enable robust artifact detection and categorization, we introduce GenVID, a large-scale dataset of 80k videos generated by various state-of-the-art video generation models, each carefully annotated for the defined artifact categories. Leveraging GenVID, we develop DVAR, a Dense Video Artifact Recognition framework for fine-grained identification and classification of generative artifacts. Extensive experiments show that our approach significantly improves artifact detection accuracy and enables effective filtering of low-quality content.
Why we are recommending this paper?
Because image and video generation is a popular topic and you have less than 3 interests with available recommendations
Rutgers University
AI Insights
  • Generative models: A class of machine learning algorithms that can generate new data samples based on a given dataset. (ML: 0.97)πŸ‘πŸ‘Ž
  • The paper assumes that the input text is well-formed and does not contain any errors or ambiguities. (ML: 0.97)πŸ‘πŸ‘Ž
  • Diffusion prior constraints: A type of regularization technique used in generative models that encourages the model to produce samples that are similar to the data distribution. (ML: 0.92)πŸ‘πŸ‘Ž
  • The paper discusses the concept of ConceptLab, a creative generation system that uses diffusion prior constraints to generate novel images. (ML: 0.90)πŸ‘πŸ‘Ž
  • ConceptLab is based on a combination of generative models and text-image harmony, which allows it to produce high-quality images with specific styles and concepts. (ML: 0.88)πŸ‘πŸ‘Ž
  • ConceptLab is based on a combination of generative models and text-image harmony, which allows it to produce high-quality images with specific styles and concepts. (ML: 0.88)πŸ‘πŸ‘Ž
  • The paper presents ConceptLab, a creative generation system that uses diffusion prior constraints to generate novel images. (ML: 0.86)πŸ‘πŸ‘Ž
  • The authors propose a new method for training generative models using diffusion prior constraints, which enables the model to learn from the data distribution and generate novel samples. (ML: 0.84)πŸ‘πŸ‘Ž
  • The authors propose a new method for training generative models using diffusion prior constraints, which enables the model to learn from the data distribution and generate novel samples. (ML: 0.84)πŸ‘πŸ‘Ž
  • Text-image harmony: A method for combining text and image features to generate images with specific styles and concepts. (ML: 0.76)πŸ‘πŸ‘Ž
Abstract
Creative image generation has emerged as a compelling area of research, driven by the need to produce novel and high-quality images that expand the boundaries of imagination. In this work, we propose a novel framework for creative generation using diffusion models, where creativity is associated with the inverse probability of an image's existence in the CLIP embedding space. Unlike prior approaches that rely on a manual blending of concepts or exclusion of subcategories, our method calculates the probability distribution of generated images and drives it towards low-probability regions to produce rare, imaginative, and visually captivating outputs. We also introduce pullback mechanisms, achieving high creativity without sacrificing visual fidelity. Extensive experiments on text-to-image diffusion models demonstrate the effectiveness and efficiency of our creative generation framework, showcasing its ability to produce unique, novel, and thought-provoking images. This work provides a new perspective on creativity in generative models, offering a principled method to foster innovation in visual content synthesis.
Why we are recommending this paper?
Because image and video generation is a popular topic and you have less than 3 interests with available recommendations

Interests not found

We did not find any papers that match the below interests. Try other terms also consider if the content exists in arxiv.org.
  • Data Career Path
  • Data Science Career Guidance
  • Data Science Career Advice
  • Data Careers
You can edit or add more interests any time.