AIS Logo
Living knowledge for digital leadership
All AI Governance & Ethics Digital Transformation & Innovation Supply Chain & IoT SME & IT Management Platform Ecosystems & Strategy Cybersecurity & Risk AI Applications & Technologies Healthcare & Well-being Digital Work & Collaboration
Analyzing German Parliamentary Speeches: A Machine Learning Approach for Topic and Sentiment Classification

Analyzing German Parliamentary Speeches: A Machine Learning Approach for Topic and Sentiment Classification

Lukas Pätz, Moritz Beyer, Jannik Späth, Lasse Bohlen, Patrick Zschech, Mathias Kraus, and Julian Rosenberger
This study investigates political discourse in the German parliament (the Bundestag) by applying machine learning to analyze approximately 28,000 speeches from the last five years. The researchers developed and trained two separate models to classify the topic and the sentiment (positive or negative tone) of each speech. These models were then used to identify trends in topics and sentiment across different political parties and over time.

Problem In recent years, Germany has experienced a growing public distrust in political institutions and a perceived divide between politicians and the general population. While much political discussion is analyzed from social media, understanding the formal, unfiltered debates within parliament is crucial for transparency and for assessing the dynamics of political communication. This study addresses the need for tools to systematically analyze this large volume of political speech to uncover patterns in parties' priorities and rhetorical strategies.

Outcome - Debates are dominated by three key policy areas: Economy and Finance, Social Affairs and Education, and Foreign and Security Policy, which together account for about 70% of discussions.
- A party's role as either government or opposition strongly influences its tone; parties in opposition use significantly more negative language than those in government, and this tone shifts when their role changes after an election.
- Parties on the political extremes (AfD and Die Linke) consistently use a much higher percentage of negative language compared to centrist parties.
- Parties tend to be most critical (i.e., use more negative sentiment) when discussing their own core policy areas, likely as a strategy to emphasize their priorities and the need for action.
- The developed machine learning models proved highly effective, demonstrating that this computational approach is a feasible and valuable method for large-scale analysis of political discourse.
Natural Language Processing, German Parliamentary, Discourse Analysis, Bundestag, Machine Learning, Sentiment Analysis, Topic Classification
Challenges and Mitigation Strategies for AI Startups: Leveraging Effectuation Theory in a Dynamic Environment

Challenges and Mitigation Strategies for AI Startups: Leveraging Effectuation Theory in a Dynamic Environment

Marleen Umminger, Alina Hafner
This study investigates the unique benefits and obstacles encountered by Artificial Intelligence (AI) startups. Through ten semi-structured interviews with founders in the DACH region, the research identifies key challenges and applies effectuation theory to explore effective strategies for navigating the uncertain and dynamic high-tech field.

Problem While investment in AI startups is surging, founders face unique challenges related to data acquisition, talent recruitment, regulatory hurdles, and intense competition. Existing literature often groups AI startups with general digital ventures, overlooking the specific difficulties stemming from AI's complexity and data dependency, which creates a need for tailored mitigation strategies.

Outcome - AI startups face core resource challenges in securing high-quality data, accessing affordable AI models, and hiring skilled technical staff like CTOs.
- To manage costs, founders often use publicly available data, form partnerships with customers for data access, and start with open-source or low-cost MVP models.
- Founders navigate competition by tailoring solutions to specific customer needs and leveraging personal networks, while regulatory uncertainty is managed by either seeking legal support or framing compliance as a competitive advantage to attract enterprise customers.
- Effectuation theory proves to be a relevant framework, as successful founders tend to leverage existing resources and networks (bird-in-hand), form strategic partnerships (crazy quilt), and adapt flexibly to unforeseen events (lemonade) rather than relying on long-term prediction.
Artificial intelligence, Entrepreneurial challenge, Effectuation theory, Qualitative research, AI startups, Mitigation strategies
BPMN4CAI: A BPMN Extension for Modeling Dynamic Conversational AI

BPMN4CAI: A BPMN Extension for Modeling Dynamic Conversational AI

Björn-Lennart Eger, Daniel Rose, and Barbara Dinter
This study develops and evaluates a standard-compliant extension for Business Process Model and Notation (BPMN) called BPMN4CAI. Using a Design Science Research methodology, the paper creates a framework that systematically extends existing BPMN elements to better model the dynamic and context-sensitive interactions of Conversational AI systems. The applicability of the BPMN4CAI framework is demonstrated through a case study in the insurance industry.

Problem Conversational AI systems like chatbots are increasingly integrated into business processes, but the standard modeling language, BPMN, is designed for predictable, deterministic processes. This creates a gap, as traditional BPMN cannot adequately represent the dynamic, context-aware dialogues and flexible decision-making inherent to modern AI. Businesses lack a standardized method to formally and accurately model processes involving these advanced AI agents.

Outcome - The study successfully developed BPMN4CAI, an extension to the standard BPMN, which allows for the formal modeling of Conversational AI in business processes.
- The new extension elements (e.g., Conversational Task, AI Decision Gateway, Human Escalation Event) facilitate the representation of adaptive decision-making, context management, and transparent interactions.
- A proof-of-concept demonstrated that BPMN4CAI improves model clarity and provides a semantic bridge for technical implementation compared to standard BPMN.
- The evaluation also identified limitations, noting that modeling highly dynamic, non-deterministic process paths and visualizing complex context transfers remains a challenge.
Conversational AI, BPMN, Business Process Modeling, Chatbots, Conversational Agent
Generative Al in Business Process Optimization: A Maturity Analysis of Business Applications

Generative Al in Business Process Optimization: A Maturity Analysis of Business Applications

Ralf Mengele
This study analyzes the current state of Generative AI (GAI) in the business world by systematically reviewing scientific literature. It identifies where GAI applications have been explored or implemented across the value chain and evaluates the maturity of these use cases. The goal is to provide managers and researchers with a clear overview of which business areas can already benefit from GAI and which require further development.

Problem While Generative AI holds enormous potential for companies, its recent emergence means it is often unclear where the technology can be most effectively applied. Businesses lack a comprehensive, systematic overview that evaluates the maturity of GAI use cases across different business processes, making it difficult to prioritize investment and adoption.

Outcome - The most mature and well-researched applications of Generative AI are in product development and in maintenance and repair within the manufacturing sector.
- The manufacturing segment as a whole exhibits the most mature GAI use cases compared to other parts of the business value chain.
- Technical domains show a higher level of GAI maturity and successful implementation than process areas dominated by interpersonal interactions, such as marketing and sales.
- GAI models like Generative Adversarial Networks (GANs) are particularly mature, proving highly effective for tasks like generating synthetic data for early damage detection in machinery.
- Research into GAI is still in its early stages for many business areas, with fields like marketing, sales, and human resources showing low implementation and maturity.
Generative AI, Business Processes, Optimization, Maturity Analysis, Literature Review, Manufacturing
Successfully Organizing AI Innovation Through Collaboration with Startups

Successfully Organizing AI Innovation Through Collaboration with Startups

Jana Oehmichen, Alexander Schult, John Qi Dong
This study examines how established firms can successfully partner with Artificial Intelligence (AI) startups to foster innovation. Based on an in-depth analysis of six real-world AI implementation projects across two startups, the research identifies five key challenges and provides corresponding recommendations for navigating these collaborations effectively.

Problem Established companies often lack the specialized expertise needed to leverage AI technologies, leading them to partner with startups. However, these collaborations introduce unique difficulties, such as assessing a startup's true capabilities, identifying high-impact AI applications, aligning commercial interests, and managing organizational change, which can derail innovation efforts.

Outcome - Challenge 1: Finding the right AI startup. Firms should overcome the inscrutability of AI startups by assessing credible quality signals, such as investor backing, academic achievements of staff, and success in prior contests, rather than relying solely on product demos.
- Challenge 2: Identifying the right AI use case. Instead of focusing on data availability, companies should collaborate with startups in workshops to identify use cases with the highest potential for value creation and business impact.
- Challenge 3: Agreeing on commercial terms. To align incentives and reduce information asymmetry, contracts should include performance-based or usage-based compensation, linking the startup's payment to the value generated by the AI solution.
- Challenge 4: Considering the impact on people. Firms must manage user acceptance by carefully selecting the degree of AI autonomy, involving employees in the design process, and clarifying the startup's role to mitigate fears of job displacement.
- Challenge 5: Overcoming implementation roadblocks. Depending on the company's organizational maturity, it should either facilitate deep collaboration between the startup and all internal stakeholders or use the startup to build new systems that bypass internal roadblocks entirely.
Artificial Intelligence, AI Innovation, Corporate-startup collaboration, Open Innovation, Digital Transformation, AI Startups
How Siemens Democratized Artificial Intelligence

How Siemens Democratized Artificial Intelligence

Benjamin van Giffen, Helmuth Ludwig
This paper presents an in-depth case study on how the global technology company Siemens successfully moved artificial intelligence (AI) projects from pilot stages to full-scale, value-generating applications. The study analyzes Siemens' journey through three evolutionary stages, focusing on the concept of 'AI democratization', which involves integrating the unique skills of domain experts, data scientists, and IT professionals. The findings provide a framework for how other organizations can build the necessary capabilities to adopt and scale AI technologies effectively.

Problem Many companies invest in artificial intelligence but struggle to progress beyond small-scale prototypes and pilot projects. This failure to scale prevents them from realizing the full business value of AI. The core problem is the difficulty in making modern AI technologies broadly accessible to employees, which is necessary to identify, develop, and implement valuable applications across the organization.

Outcome - Siemens successfully scaled AI by evolving through three stages: 1) Tactical AI pilots, 2) Strategic AI enablement, and 3) AI democratization for business transformation.
- Democratizing AI, defined as the collaborative integration of domain experts, data scientists, and IT professionals, is crucial for overcoming key adoption challenges such as defining AI tasks, managing data, accepting probabilistic outcomes, and addressing 'black-box' fears.
- Key initiatives that enabled this transformation included establishing a central AI Lab to foster co-creation, an AI Academy for upskilling employees, and developing a global AI platform to support scaling.
- This approach allowed Siemens to transform manufacturing processes with predictive quality control and create innovative healthcare products like the AI-Rad Companion.
- The study concludes that democratizing AI creates value by rooting AI exploration in deep domain knowledge and reduces costs by creating scalable infrastructures and processes.
Artificial Intelligence, AI Democratization, Digital Transformation, Organizational Capability, Case Study, AI Adoption, Siemens
How Boards of Directors Govern Artificial Intelligence

How Boards of Directors Govern Artificial Intelligence

Benjamin van Giffen, Helmuth Ludwig
This study investigates how corporate boards of directors oversee and integrate Artificial Intelligence (AI) into their governance practices. Based on in-depth interviews with high-profile board members from diverse industries, the research identifies common challenges and provides examples of effective strategies for board-level AI governance.

Problem Despite the transformative impact of AI on the business landscape, the majority of corporate boards struggle to understand its implications and their role in governing it. This creates a significant gap, as boards have a fiduciary responsibility to oversee strategy, risk, and investment related to critical technologies, yet AI is often not a mainstream boardroom topic.

Outcome - Identified four key groups of board-level AI governance issues: Strategy and Firm Competitiveness, Capital Allocation, AI Risks, and Technology Competence.
- Boards should ensure AI is integrated into the company's core business strategy by evaluating its impact on the competitive landscape and making it a key topic in annual strategy meetings.
- Effective capital allocation involves encouraging AI experimentation, securing investments in foundational AI capabilities, and strategically considering external partnerships and acquisitions.
- To manage risks, boards must engage with experts, integrate AI-specific risks into Enterprise Risk Management (ERM) frameworks, and address ethical, reputational, and legal challenges.
- Enhancing technology competence requires boards to develop their own AI literacy, review board and committee composition for relevant expertise, and include AI competency in executive succession planning.
AI governance, board of directors, corporate governance, artificial intelligence, strategic management, risk management, technology competence
Promises and Perils of Generative AI in Cybersecurity

Promises and Perils of Generative AI in Cybersecurity

Pratim Datta, Tom Acton
This paper presents a case study of a fictional insurance company, based on real-life events, to illustrate how generative artificial intelligence (GenAI) can be used for both offensive and defensive cybersecurity purposes. It explores the dual nature of GenAI as a tool for both attackers and defenders, presenting a significant dilemma for IT executives. The study provides actionable recommendations for developing a comprehensive cybersecurity strategy in the age of GenAI.

Problem With the rapid adoption of Generative AI by both cybersecurity defenders and malicious actors, IT leaders face a critical challenge. GenAI significantly enhances the capabilities of attackers to create sophisticated, large-scale, and automated cyberattacks, while also offering powerful new tools for defense. This creates a high-stakes 'AI arms race,' forcing organizations to decide how to strategically embrace GenAI for defense without being left vulnerable to adversaries armed with the same technology.

Outcome - GenAI is a double-edged sword, capable of both triggering and defending against sophisticated cyberattacks, requiring a proactive, not reactive, security posture.
- Organizations must integrate a 'Defense in Depth' (DiD) strategy that extends beyond technology to include processes, a security-first culture, and continuous employee education.
- Robust data governance is crucial to manage and protect data, the primary target of attacks, by classifying its value and implementing security controls accordingly.
- A culture of continuous improvement is essential, involving regular simulations of real-world attacks (red-team/blue-team exercises) and maintaining a zero-trust mindset.
- Companies must fortify defenses against AI-powered social engineering by combining advanced technical filtering with employee training focused on skepticism and verification.
- Businesses should embrace proactive, AI-driven defense mechanisms like AI-powered threat hunting and adaptive honeypots to anticipate and neutralize threats before they escalate.
Generative AI, Cybersecurity, Black-hat AI, White-hat AI, Threat Hunting, Social Engineering, Defense in Depth
Successfully Mitigating AI Management Risks to Scale AI Globally

Successfully Mitigating AI Management Risks to Scale AI Globally

Thomas Hutzschenreuter, Tim Lämmermann, Alexander Sake, Helmuth Ludwig
This study presents an in-depth case study of the industrial AI pioneer Siemens AG to understand how companies can effectively scale artificial intelligence systems. It identifies five critical technology management risks associated with both generative and predictive AI and provides practical recommendations for mitigating them to create company-wide business impact.

Problem Many companies struggle to effectively scale modern AI systems, with over 70% of implementation projects failing to create a measurable business impact. These failures stem from machine learning's unique characteristics, which amplify existing technology management challenges and introduce entirely new ones that firms are often unprepared to handle.

Outcome - Missing or falsely evaluated potential AI use case opportunities.
- Algorithmic training and data quality issues.
- Task-specific system complexities.
- Mismanagement of system stakeholders.
- Threats from provider and system dependencies.
AI management, risk mitigation, scaling AI, generative AI, predictive AI, technology management, case study
How Audi Scales Artificial Intelligence in Manufacturing

How Audi Scales Artificial Intelligence in Manufacturing

André Sagodi, Benjamin van Giffen, Johannes Schniertshauer, Klemens Niehues, Jan vom Brocke
This paper presents a case study on how the automotive manufacturer Audi successfully scaled an artificial intelligence (AI) solution for quality inspection in its manufacturing press shops. It analyzes Audi's four-year journey, from initial exploration to multi-site deployment, to identify key strategies and challenges. The study provides actionable recommendations for senior leaders aiming to capture business value by scaling AI innovations.

Problem Many organizations struggle to move their AI initiatives from the pilot phase to full-scale operational use, failing to realize the technology's full economic potential. This is a particular challenge in manufacturing, where integrating AI with legacy systems and processes presents significant barriers. This study addresses how a company can overcome these challenges to successfully scale an AI solution and unlock long-term business value.

Outcome - Audi successfully scaled an AI-based system to automate the detection of cracks in sheet metal parts, a crucial quality control step in its press shops.
- The success was driven by a strategic four-stage approach: Exploring, Developing, Implementing, and Scaling, with a focus on designing for scalability from the outset.
- Key success factors included creating a single, universal AI model for multiple deployments, leveraging data from various sources to improve the model, and integrating the solution into the broader Volkswagen Group's digital production platform to create synergies.
- The study highlights the importance of decoupling value from cost, which Audi achieved by automating monitoring and deployment pipelines, thereby scaling operations without proportionally increasing expenses.
- Recommendations for other businesses include making AI scaling a strategic priority, fostering collaboration between AI experts and domain specialists, and streamlining operations through automation and robust governance.
Artificial Intelligence, AI Scaling, Manufacturing, Automotive Industry, Case Study, Digital Transformation, Quality Inspection
The Promise and Perils of Low-Code AI Platforms

The Promise and Perils of Low-Code AI Platforms

Maria Kandaurova, Daniel A. Skog, Petra M. Bosch-Sijtsema
This study investigates the adoption of a low-code conversational Artificial Intelligence (AI) platform within four multinational corporations. Through a case study approach, the research identifies significant challenges that arise from fundamental, yet incorrect, assumptions about low-code technologies. The paper offers recommendations for companies to better navigate the implementation process and unlock the full potential of these platforms.

Problem As businesses increasingly turn to AI for process automation, they often encounter significant hurdles during adoption. Low-code AI platforms are marketed as a solution to simplify this process, but there is limited research on their real-world application. This study addresses the gap by showing how companies' false assumptions about the ease of use, adaptability, and integration of these platforms can limit their effectiveness and return on investment.

Outcome - The usability of low-code AI platforms is often overestimated; non-technical employees typically face a much steeper learning curve than anticipated and still require a foundational level of coding and AI knowledge.
- Adapting low-code AI applications to specific, complex business contexts is challenging and time-consuming, contrary to the assumption of easy tailoring. It often requires significant investment in standardizing existing business processes first.
- Integrating low-code platforms with existing legacy systems and databases is not a simple 'plug-and-play' process. Companies face significant challenges due to incompatible data formats, varied interfaces, and a lack of a comprehensive data strategy.
- Successful implementation requires cross-functional collaboration between IT and business teams, thorough platform testing before procurement, and a strategic approach to reengineering business processes to align with AI capabilities.
Low-Code AI Platforms, Artificial Intelligence, Conversational AI, Implementation Challenges, Digital Transformation, Business Process Automation, Case Study
How GuideCom Used the Cognigy.AI Low-Code Platform to Develop an AI-Based Smart Assistant

How GuideCom Used the Cognigy.AI Low-Code Platform to Develop an AI-Based Smart Assistant

Imke Grashoff, Jan Recker
This case study investigates how GuideCom, a medium-sized German software provider, utilized the Cognigy.AI low-code platform to create an AI-based smart assistant. The research follows the company's entire development process to identify the key ways in which low-code platforms enable and constrain AI development. The study illustrates the strategic trade-offs companies face when adopting this approach.

Problem Small and medium-sized enterprises (SMEs) often lack the extensive resources and specialized expertise required for in-house AI development, while off-the-shelf solutions can be too rigid. Low-code platforms are presented as a solution to democratize AI, but there is a lack of understanding regarding their real-world impact. This study addresses the gap by examining the practical enablers and constraints that firms encounter when using these platforms for AI product development.

Outcome - Low-code platforms enable AI development by reducing complexity through visual interfaces, facilitating cross-functional collaboration between IT and business experts, and preserving resources.
- Key constraints of using low-code AI platforms include challenges with architectural integration into existing systems, ensuring the product is expandable for different clients and use cases, and managing security and data privacy concerns.
- Contrary to the 'no-code' implication, existing software development skills are still critical for customizing solutions, re-engineering code, and overcoming platform limitations, especially during testing and implementation.
- Establishing a strong knowledge network with the platform provider (for technical support) and innovation partners like clients (for domain expertise and data) is a crucial factor for success.
- The decision to use a low-code platform is a strategic trade-off; it significantly lowers the barrier to entry for AI innovation but requires careful management of platform dependencies and inherent constraints.
low-code development, AI development, smart assistant, conversational AI, case study, digital transformation, SME
Showing all 48 podcasts