Networking in AI: A Perspective for Business Track Students
Picture
Member for
1 year 1 month
Real name
David O'Neill
Bio
David O’Neill is a Professor of Finance and Data Analytics at the Gordon School of Business, SIAI. A Swiss-based researcher, his work explores the intersection of quantitative finance, AI, and educational innovation, particularly in designing executive-level curricula for AI-driven investment strategy. In addition to teaching, he manages the operational and financial oversight of SIAI’s education programs in Europe, contributing to the institute’s broader initiatives in hedge fund research and emerging market financial systems.
Published
Modified
A recent discussion on GIAI Square brought up concerns about networking opportunities in the SIAI 2.0 AI MBA program. While technical students focus on engineering and quantitative finance, business track students need a different kind of networking—one that connects them to venture capitalists, private equity firms, and AI-driven business leaders.
Unlike traditional business MBAs, where networking revolves around corporate job placements and HR-driven recruitment, the AI/DS industry demands a deeper understanding of technological realities. This article explores how business track students can build a valuable professional network that extends beyond superficial industry connections.
AI Investment and the Role of Business Track Students
SIAI’s business track does not train software engineers, but it does ensure that students understand the true mechanics of AI/DS projects. This exposure is critical for those aiming to work in:
Venture Capital (VC): Assessing AI startups requires more than just reviewing pitch decks. Business track students should develop the ability to distinguish real AI capabilities from hype-driven marketing.
Private Equity Funds (PEF): AI-focused PEFs need professionals who understand how AI can enhance operational efficiencies and financial performance, rather than just investing in ‘trendy’ AI companies.
AI Strategy & Consulting: Business leaders who understand the limitations of AI can provide more effective strategic guidance than those who rely solely on buzzwords.
The Difference Between SIAI’s Business Track and Traditional STEM MBAs
There has been a rise in STEM MBA programs in the U.S., many of which provide only bootcamp-level AI training with little depth. Some prospective students might wonder: If business track students are not gaining hands-on AI/DS skills, how is this different from other MBA programs?
The key distinction is that SIAI’s business track provides exposure to real AI/DS research and development. Students do not merely learn surface-level programming or attend AI workshops; they engage with real AI/DS researchers and witness what separates serious AI/DS projects from bootcamp-level development.
This exposure allows them to:
Discern the difference between commercialized, shallow AI solutions and research-driven AI models.
Recognize what real AI/DS teams need in terms of business support and strategic planning.
Avoid common investment pitfalls by understanding the depth required to execute successful AI-driven businesses.
Unlike traditional MBA graduates who may overestimate their AI literacy, SIAI business track students will not be fooled by superficial AI projects—they will develop a refined sense of what truly constitutes an AI-driven innovation.
Current Problems in AI Investment and How SIAI Graduates Can Stand Out
Many VCs and PEFs today—especially in less mature markets—lack the technical depth to evaluate AI startups properly. Instead, they:
Follow the herd, investing in companies based on hype rather than substance.
Rely on media narratives, without critically assessing the technology’s viability.
Ignore deep industry research, instead making decisions based on networking events and investor consensus.
This lack of technical literacy often leads to poor investment choices, funding companies that lack true AI innovation while overlooking startups with real technical potential.
SIAI’s business track aims to close this knowledge gap, producing professionals who can:
Assess AI/DS startups based on real technological value.
Guide AI firms with strategic, well-informed business insights.
Recognize AI-driven business models that are sustainable rather than speculative.
SIAI’s Approach to Business Networking
Networking in AI-driven industries is not just about knowing the right people—it’s about having the credibility to engage with top AI professionals and investors. To ensure business track students develop this credibility, SIAI’s networking approach includes:
Connections with senior AI researchers and investors: Instead of focusing solely on HR networking events, students gain access to scientists and executives who shape AI business trends.
Industry-based case studies: Students analyze real AI business models, learning how to separate meaningful innovations from unsustainable hype.
GIAI’s Future Investment Arm: As part of its long-term vision, GIAI—the mother institution of SIAI—plans to launch its own investment vehicles, including a PEF/VC firm specializing in AI-driven businesses. This will provide business track students with real-world exposure to AI investments and potential career opportunities.
What Really Matters for Business Track Students?
Unlike traditional MBAs, success in AI/DS business leadership is not based on prestige or social capital alone. The best business professionals in this field:
Understand AI/DS at a fundamental level.
Can differentiate between real innovation and overhyped technology.
Leverage technical credibility to earn the trust of AI founders and investors.
If you want to succeed in AI-driven business roles, your network must be built on knowledge and value, not just professional titles.
David O’Neill is a Professor of Finance and Data Analytics at the Gordon School of Business, SIAI. A Swiss-based researcher, his work explores the intersection of quantitative finance, AI, and educational innovation, particularly in designing executive-level curricula for AI-driven investment strategy. In addition to teaching, he manages the operational and financial oversight of SIAI’s education programs in Europe, contributing to the institute’s broader initiatives in hedge fund research and emerging market financial systems.
Networking in AI: A Perspective for Technical Track Students
Picture
Member for
1 year 1 month
Real name
Ethan McGowan
Bio
Ethan McGowan is a Professor of Financial Technology and Legal Analytics at the Gordon School of Business, SIAI. Originally from the United Kingdom, he works at the frontier of AI applications in financial regulation and institutional strategy, advising on governance and legal frameworks for next-generation investment vehicles. McGowan plays a key role in SIAI’s expansion into global finance hubs, including oversight of the institute’s initiatives in the Middle East and its emerging hedge fund operations.
Published
Modified
In a recent discussion on GIAI Square, a student raised concerns about networking opportunities in the SIAI 2.0 AI MBA program, particularly about the strength of the alumni network and its impact on career opportunities post-graduation. As a professor and industry professional, I provided my perspective based on both academic experience and real-world industry exposure.
With the AI MBA program now divided into technical and business tracks, it is essential to address both career paths. This article focuses on networking for technical students—those who aim to work in AI/DS engineering roles or quantitative finance, where expertise matters more than connections alone.
Networking in STEM: Beyond Job Fairs and School Prestige
During my own education, job fairs were frequent, with Fortune 500 companies sending HR teams to present their hiring strategies. However, let me be blunt—these events were not about hiring top talent. They were primarily about company branding and industry presence. 99.9% of attendees did not walk away with job offers.
For technical professionals, the challenge is clear: your value must be demonstrated through skill, not just credentials. Schools provide a foundation, but true expertise comes from independent work. The best networking strategy for technical students involves:
Building a strong GitHub presence: Employers often search candidates online. A well-documented portfolio of math-heavy AI/DS projects will do more for your career than a LinkedIn profile alone.
Engaging in open-source AI projects: Actively contributing to repositories increases visibility among technical hiring managers.
Participating in AI/DS research communities: Whether through Kaggle competitions, research publications, or AI forums, showcasing expertise is critical.
Technical blog writing and discussions: Sharing insights on AI/DS applications and mathematical concepts demonstrates thought leadership.
SIAI’s Approach to Technical Networking
Unlike traditional MBA programs, where networking often means connecting with HR teams, SIAI’s technical track focuses on peer and mentor-driven networking. Our approach emphasizes:
Direct engagement with senior AI researchers: Instead of prioritizing HR-led job fairs, SIAI hosts expert-led discussions on hiring preferences.
Research-driven networking: SIAI students gain access to real-world AI/DS projects, allowing them to work alongside experienced professionals.
GIAI’s Future Investment Arm: As part of its long-term vision, GIAI—the mother institution of SIAI—plans to launch its own investment vehicles, including a hedge fund focused on computational finance. This initiative will provide high-performing technical students with opportunities to transition into quantitative finance roles and AI-driven investment research.
What Really Matters for AI/DS Professionals?
For technical track students, networking is not about the quantity of connections but the quality of expertise. AI hiring decisions are often made by technical leaders, not HR representatives. Companies look for candidates who demonstrate:
A deep mathematical and computational foundation
Strong problem-solving abilities in AI and Data Science
Independent research or engineering contributions
If you focus on becoming an expert in your domain, networking will follow naturally—senior researchers and industry professionals will recognize your work and recommend you for opportunities.
Ethan McGowan is a Professor of Financial Technology and Legal Analytics at the Gordon School of Business, SIAI. Originally from the United Kingdom, he works at the frontier of AI applications in financial regulation and institutional strategy, advising on governance and legal frameworks for next-generation investment vehicles. McGowan plays a key role in SIAI’s expansion into global finance hubs, including oversight of the institute’s initiatives in the Middle East and its emerging hedge fund operations.
Beyond Bootcamps – A Rigorous AI Education Rooted in Science and Business
Picture
Member for
1 year 1 month
Real name
Catherine Maguire
Bio
Catherine Maguire is a Professor of Computer Science and AI Systems at the Gordon School of Business, part of the Swiss Institute of Artificial Intelligence (SIAI). She specializes in machine learning infrastructure and applied data engineering, with a focus on bridging research and large-scale deployment of AI tools in financial and policy contexts. Based in the United States (with summer in Berlin and Zurich), she co-leads SIAI’s technical operations, overseeing the institute’s IT architecture and supporting its research-to-production pipeline for AI-driven finance.
Published
Modified
Unlike typical AI bootcamps, SIAI offers in-depth AI education with a strong foundation in mathematics, statistics, and real-world business applications.
The MSc AI/Data Science program at SIAI emphasizes rigorous scientific studies, ensuring students master the theoretical and practical aspects of AI.
SIAI’s MBA AI programs incorporate extensive business case studies, with a new MBA AI/Finance track focusing on corporate finance and financial investments.
Beyond AI Bootcamp
AI bootcamps have become a popular way to enter the field, promising job-ready skills in a matter of months. However, these programs often emphasize coding without the necessary depth in mathematical reasoning, algorithmic theory, or real-world application complexities. While they may provide an entry point, they fall short in developing expertise necessary for advanced AI research, business strategy, and financial decision-making.
At the Swiss Institute of Artificial Intelligence (SIAI), we go beyond the standard bootcamp approach. Our programs are built on a foundation of rigorous academic principles, blending mathematical and statistical rigor with AI-driven business applications. We train professionals to understand AI at its core, rather than just using pre-built libraries and models.
Recognizing the need for structured AI education, SIAI’s mother institution, the Global Institute of Artificial Intelligence (GIAI), offers a 'free' AI bootcamp course on GIAI LMS(https://lms.giai.org). This courseware serves as an introductory learning platform, providing accessible AI and data science fundamentals for beginners.
While the free bootcamp offers valuable foundational training, it is designed only as a stepping stone to more advanced studies. For those seeking deeper expertise, SIAI’s MSc and MBA AI programs provide the next level of education.
MSc AI/Data Science: The Scientific Approach
SIAI’s MSc AI/Data Science program is built on the pillars of mathematics, statistics, and scientific computing. Unlike bootcamps that focus mainly on coding skills, our MSc program ensures students develop a strong understanding of:
Advanced mathematical modeling for AI
Statistical inference and probability theory
Computational optimization and algorithmic design
Theoretical and applied machine learning
AI research methodologies and scientific experimentation
Graduates of this program are equipped not only to implement AI models but to develop new AI techniques and contribute to scientific advancements in artificial intelligence and data science.
MBA AI/Big Data: Business-Driven AI Case Studies
While the MSc program takes a research-oriented approach, SIAI’s MBA AI/Big Data program focuses on real-world business applications. This program is structured around in-depth AI and data science case studies, helping executives and business professionals understand:
How AI is applied in marketing, operations, and strategy
The role of data-driven decision-making in business transformation
Ethical and regulatory challenges of AI deployment in enterprises
Case studies of AI implementation across diverse industries
Unlike AI bootcamps that offer surface-level exposure to business analytics, SIAI’s MBA AI/Big Data program ensures professionals gain practical insights into AI’s role in corporate decision-making.
Introducing MBA AI/Finance: AI in Corporate Finance and Investment
Building upon the success of MBA AI/Big Data, SIAI is launching MBA AI/Finance, a specialized track integrating AI with corporate finance and financial investment strategies. This program provides:
AI-driven corporate financial analysis: Understanding how AI can optimize budgeting, forecasting, and risk management in enterprises.
AI applications in investment strategies: Learning how hedge funds, asset managers, and financial institutions leverage AI to enhance portfolio management, algorithmic trading, and risk assessment.
Case studies on AI in financial decision-making: Reviewing how major firms have successfully integrated AI into financial operations and strategic investments.
This program is designed for finance professionals, investment analysts, and corporate executives looking to harness AI in financial decision-making. Unlike bootcamp courses that barely scratch the surface, MBA AI/Finance provides deep, case-based learning tailored for real-world applications.
Why SIAI? A Path Beyond the Bootcamp Mentality
SIAI stands apart from standard AI bootcamps by emphasizing:
Scientific Depth: Mathematical and statistical foundations critical for true AI expertise.
Real-World Case Studies: Business-oriented applications that translate AI into tangible business results.
Specialized Tracks: Focused programs in AI/Data Science, AI/Big Data, and AI/Finance to meet diverse career needs.
For those looking to develop a genuine AI expertise beyond a crash course, SIAI offers a structured, rigorous, and research-driven educational experience. Whether through the MSc AI/Data Science track for scientific mastery or the MBA AI programs for business and finance applications, SIAI ensures that students receive an education that truly sets them apart in the AI industry.
Picture
Member for
1 year 1 month
Real name
Catherine Maguire
Bio
Catherine Maguire is a Professor of Computer Science and AI Systems at the Gordon School of Business, part of the Swiss Institute of Artificial Intelligence (SIAI). She specializes in machine learning infrastructure and applied data engineering, with a focus on bridging research and large-scale deployment of AI tools in financial and policy contexts. Based in the United States (with summer in Berlin and Zurich), she co-leads SIAI’s technical operations, overseeing the institute’s IT architecture and supporting its research-to-production pipeline for AI-driven finance.
[MSc Research topic 2025-2026] Shapley value with graph models for HR
Picture
Member for
1 year 2 months
Real name
Keith Lee
Bio
Keith Lee is a Professor of AI and Data Science at the Gordon School of Business, part of the Swiss Institute of Artificial Intelligence (SIAI), where he leads research and teaching on AI-driven finance and data science. He is also a Senior Research Fellow with the GIAI Council, advising on the institute’s global research and financial strategy, including initiatives in Asia and the Middle East.
Published
Modified
GIAI's primary research objective with the coming cycle's of MSc AI/Data Science is to build a graph-based Shapley Value for HR contribution analysis. In case you are not familiar with Shapley Value, it is a game-theory concept for properly allocating group project's gains/costs, which was first introduced in 1951 and awarded Nobel Prize in 2012.
The idea for this model originally came from one of the business case study classes(BUS501) in the MBA AI/Big Data program. In the class, students were given the task of testing a model to measure each student's contribution to group projects. Some students wanted to extend the model by incorporating participation in forum discussions as an additional metric.
This idea gained traction and has since been integrated into all course evaluations at the Swiss Institute of Artificial Intelligence (SIAI). Now, we aim to take this model beyond the classroom and make it more general and business-friendly. The goal is to refine it into a structured, scalable framework that can address a key challenge in corporate HR analytics: how to accurately measure multi-stage and indirect contributions in large organizations.
Understanding Team Contribution in Multi-Staged Work Environments
Traditional regression-based models for performance evaluation assign proportional credit based on direct contributions. While useful, they assume that all contributions are immediate and directly observable within a single stage of work. However, in real-world workplaces:
Projects are multi-staged and often take months or years to complete.
Some contributions emerge over time, rather than being immediately visible.
Key individuals may act as connectors or enablers, rather than direct output producers.
To address these challenges, I am developing a new model that leverages graph-based Shapley value calculations. Unlike conventional models, this approach:
Captures contributions that unfold over multiple project cycles.
Identifies knowledge-sharing roles that support long-term success.
Quantifies the impact of ‘helpers’ who enable others to succeed without always producing measurable outputs themselves.
Leveraging Communication Data to Measure Contribution
To make this model applicable in business settings, I plan to incorporate email and chat data as key sources of information. These internal communication networks serve as vital indicators of:
How knowledge flows within an organization.
Who provides critical insights, guidance, and solutions.
Which employees are silent contributors who strengthen a team’s efficiency over time.
This naturally raises concerns about privacy, and I want to emphasize that ethical implementation is a key priority. While companies may find it reasonable to analyze work-related communication, employees must also have the right to:
Opt out if they do not wish to be evaluated using this model.
Maintain separate communication channels—one strictly for business, another for personal interactions.
Building on Traditional Contribution Models
This model does not aim to replace existing HR analytics but rather to complement them. Traditional evaluation methods already track:
✅ Task completion and project logs (Jira, Trello, Asana) ✅ Document collaboration (Google Docs, Notion, Confluence) ✅ Meeting participation and scheduling (Google Calendar, Outlook) ✅ Code commits and technical contributions (GitHub, GitLab)
However, these approaches primarily measure direct, immediate contributions. By integrating a graph-based structure, this model adds an extra rung on the ladder, allowing us to:
Identify individuals whose contributions emerge across multiple projects.
Detect key connectors and enablers within an organization.
Assign Shapley value-based credit to those who facilitate success beyond direct outputs.
Why Does This Matter? The Role of 'Helpers' in Teams
Many workplaces unintentionally overlook contributors who are not direct project leaders. These individuals—whom I call 'helpers'—are vital in ensuring long-term efficiency, knowledge-sharing, and problem-solving.
Traditional performance metrics reward project leaders, often missing those who facilitate success behind the scenes.
A graph-based evaluation helps reveal these hidden contributors, ensuring fair recognition.
Large-scale organizations rely on cross-team knowledge flow, which is difficult to quantify with traditional models.
By refining this methodology, we aim to provide a more balanced and fair assessment of who truly drives organizational success.
A Practical Application: Fairer Bonus Allocation
A major application of this research is in corporate HR, where annual bonus allocation is often based on direct deliverables. However:
❌ Employees who create long-term strategic advantages often go unnoticed. ❌ Those who enable cross-team collaboration are rarely rewarded. ❌ Many companies struggle to identify silent contributors who significantly impact multiple projects.
Our model seeks to address this by providing data-driven, fairer evaluations that recognize both direct and indirect contributions. This could help businesses:
Improve bonus distribution fairness.
Identify emerging leaders within the company.
Strengthen team efficiency and collaboration.
Next steps after computational multi-stage cooperative game ― Auto driving and squadron drones
Game theory models are often hard to solve, but it is much harder to design a set-up for closed form solutions as well as desired equilibrium paths. After all, this is why not 'mathematical' but 'computational' approach is expected to be much more industry-friendly and we also expect to solve it within a reasonable amount of time and effort, if we can be free from theoretically robust mathematical model.
One other reason SIAI is focused on this topic is to extend the model for coordinated group behaviors in response to counterparties. Current self-driving mechanism only passively updates information from surrounding cars on the road, to the best of my knowledge. But when other cars move around with erratic behavior, for example if the driver is drunk, then evasive driving will perform far better off if the algorithm can confirm that the erratic driving is not a mistake by a sober driver but a failure of correction by a drunk driver. The same intuition becomes more pronounced if it is a drone war, especially when not a single but more than dozens of drones move together.
For one side, the algorithm has to solve a cooperative game for two drones and a coordination game for a group of drones on my side. On top of that, in the presence of enemies, now the algorithm has to take into account enemy drones strategies. So, it becomes a double-sided coordination problem. And lastly, the game does not end in a single stage, if evasive movement works.
Exactly the same logic can be applied to AI units in video games like Football. With current AI, unless the algorithm has a pre-mapped options like Alpha-Go, it cannot dynamically update the optimal responses. The game theory augmented by computational science, therefore, is another challenge that will make current AI more close to real AI.
Join the Research: MSc AI/Data Science at SIAI
This project is one of the key research opportunities in the MSc AI/Data Science program for the 2025-2026 cycle. This project demands more than just enthusiasm for AI—it requires the ability to navigate complex, multi-layered problems where business reality meets mathematical precision.
If you are passionate about:
🔹 Applying cutting-edge machine learning techniques to real-world business challenges. 🔹 Exploring AI-driven approaches to performance evaluation. 🔹 Using graph theory, game theory (Shapley value), and NLP for corporate applications.
Then this could be the perfect research opportunity for you.
💡 Exceptional students who demonstrate strong analytical skills and a commitment to AI-driven research may be considered for scholarships and funding opportunities.
However, I want to be clear—this is not a program for those seeking an easy credential. The MSc AI/Data Science at SIAI is for students who:
✔️ Want to work on serious, high-impact AI research. ✔️ Are ready to challenge traditional methods with new AI-driven approaches. ✔️ Aspire to develop solutions that companies can implement in real-world settings.
I welcome smart, ambitious, and research-driven students to join me in pushing the boundaries of AI for business.
Not sure if a year work will be enough to build a fully robust, easily modifiable, and conceptually intuitive model, but application of the work-in-progress model will be periodically shared as a form of case studies.
Necessary knowledge
Game Theory
Network Theory
Machine Learning
Large Language Model
(Some level of) Panel data
Key concepts are discussed in PreMSc (or MBA AI), and deeper ones to come in MSc AI/Data Science.
Most AI-driven HR analytics focus on traditional models. We are developing an advanced, multi-stage contribution evaluation framework—something that could redefine how businesses measure and reward employees' true impact. This is not about minor improvements; this is about setting a new industry standard. Likely mind-set is also strongly emphasized.
If interested, feel free to ask questions in comments through GIAI Square.
Picture
Member for
1 year 2 months
Real name
Keith Lee
Bio
Keith Lee is a Professor of AI and Data Science at the Gordon School of Business, part of the Swiss Institute of Artificial Intelligence (SIAI), where he leads research and teaching on AI-driven finance and data science. He is also a Senior Research Fellow with the GIAI Council, advising on the institute’s global research and financial strategy, including initiatives in Asia and the Middle East.
3 types of 'Math Genius', 2 of which will be replaced by AI
Picture
Member for
1 year 1 month
Real name
David O'Neill
Bio
David O’Neill is a Professor of Finance and Data Analytics at the Gordon School of Business, SIAI. A Swiss-based researcher, his work explores the intersection of quantitative finance, AI, and educational innovation, particularly in designing executive-level curricula for AI-driven investment strategy. In addition to teaching, he manages the operational and financial oversight of SIAI’s education programs in Europe, contributing to the institute’s broader initiatives in hedge fund research and emerging market financial systems.
Published
Modified
Mathematical ability differs across cultures, with Western academia emphasizing abstraction over procedural speed
AI is automating routine calculations, making conceptual thinking more valuable than ever
Future professionals must focus on logical reasoning and model formulation to stay relevant
After years of teaching here at SIAI, we have witnessed a varying cultural differences in perception of experts in AI/Data Science in the western hemisphere and in Asia. What was pronounced the most was the concept of mathematics necessary in this particular field. Most Asian students blindly thought that calculation capacity and problem solving skills are emphasized in our curriculum, just by reading the phrases like 'The Most Rigorous MBA in the world'.
We don't.
And we finally understand where the confusion comes from. Here is our scientific analysis of the differences.
Education researchers often distinguish between procedural fluency (being able to execute mathematical procedures quickly and accurately) and conceptual understanding (grasping the underlying principles and structures of mathematics). Many studies indicate that East Asian education systems emphasize procedural fluency, while Western systems, particularly in higher education, prioritize conceptual depth.
Research Backing This View: Studies comparing math education in China, Japan, South Korea, and Western countries (such as the US and UK) consistently show that Asian students outperform in procedural tasks but may struggle with non-standard, open-ended problems requiring deeper conceptual thinking (Ma, 1999; Stigler & Hiebert, 1999).
So, we have formalized 3 types of 'Math Genius', and please note that only the last type is needed at SIAI.
Calculator
Problem Solver
Thinker
Let's go over the dichotomy from our definition of 'Math Genius'.
1.Calculator: Speed and Accuracy as Genius
Mathematical ability is often perceived differently across educational systems. In many East Asian countries, proficiency in mathematics is equated with speed and accuracy in calculations. A student who can quickly solve a quadratic equation or compute complex arithmetic is often considered a math genius. This perception aligns with research by Stigler and Hiebert (1999), which highlights that Asian students tend to excel in procedural fluency due to structured and rigorous mathematical training at an early stage.
However, in higher education, particularly in Western academic institutions, mathematical proficiency is defined differently. The emphasis shifts from speed to logical reasoning, abstract thinking, and the ability to construct mathematical models. Research in mathematics education (Ma, 1999; Li & Collins, 2021) shows that while Asian students tend to perform well in structured mathematical settings, they often face challenges when required to engage in open-ended problem-solving and theoretical abstraction.
2.Problem Solver: Procedural fluency as 'Math Genius'
At the high school level, the focus of mathematics education begins to shift from pure calculation to problem-solving. Advanced mathematics curricula require students to derive solutions from first principles, navigate multi-step logical reasoning, and understand abstract mathematical structures. This transition is critical for success in competitive university entrance exams, as seen in South Korea’s CSAT and similar standardized assessments in other countries.
This is why Asian students excel in competitive math Olympiads, which require both procedural skill and non-standard problem-solving.
As students enter university, particularly in STEM fields, the nature of mathematics evolves further. Research in international mathematics education (Li & Shavelson, 2001) suggests that students who rely primarily on procedural problem-solving may struggle when confronted with theoretical coursework that requires constructing formal proofs and engaging with abstract concepts. This distinction between procedural fluency and conceptual understanding is well-documented in the literature on cognitive development in mathematics (Tall, 2004).
Western academia sees calculation speed as "machine-like" rather than as a sign of intelligence is supported by psychological studies on how different cultures define intelligence.
Expert Perspective: In Western academia, a "math genius" is often equated with someone who can create new mathematical theories, prove complex theorems, or develop novel models—not just someone who is quick at calculations. This is evident in how Western math competitions, graduate exams, and research expectations focus on deep reasoning rather than speed.
Historical Context: The Western concept of a mathematical genius is shaped by figures like Gauss, Euler, and Gödel, who were not just quick calculators but pioneers in abstract reasoning.
3.Thinker: The Role of Mathematical Thinking in AI and Data Science Education
In applied fields such as AI and Data Science, mathematical proficiency takes on yet another dimension. While theoretical knowledge remains essential for foundational research, most practical applications of AI do not require deep engagement with mathematical proofs. Instead, students must understand the conditions under which mathematical models apply and be able to critically evaluate their limitations.
Given this reality, the MBA AI/Big Data program at SIAI has been strategically designed to align with industry needs while accommodating different mathematical backgrounds. Rather than focusing on formal proofs, the curriculum emphasizes:
Understanding Model Assumptions – Students are trained to recognize the conditions under which different AI models (e.g., neural networks, decision trees) are effective and where they may fail.
Applying Mathematics to Business Problems – Instead of proving theorems, the focus is on using mathematical reasoning to optimize decision-making in real-world scenarios.
Bridging Procedural Fluency with Conceptual Thinking – While problem-solving remains an essential skill, students are guided to transition towards abstract thinking where necessary, particularly in courses on machine learning interpretability and data-driven strategy.
This approach aligns with the findings of mathematics education researchers (Schoenfeld, 2007), who argue that effective mathematical training must be contextualized within the problems students are expected to solve in their professional careers.
Why This Matters for Asian Students in STEM Fields
Many Asian students who transition to Western universities for undergraduate or graduate studies in STEM fields often experience a sudden drop in their perceived mathematical ability. This is not because they lack intelligence, but because their definition of mathematical proficiency has been shaped differently.
Studies on international students in STEM (Li & Collins, 2021) show that Asian students often find proof-based courses, abstract algebra, and mathematical modeling more challenging compared to their Western peers, precisely because their training has emphasized computational efficiency rather than abstraction
Students who have excelled in rapid problem-solving often struggle with abstract mathematical thinking. They may find courses in theoretical physics, real analysis, or mathematical finance unexpectedly difficult because the emphasis shifts from computation to proof-based reasoning and conceptual applications.
This is particularly critical for aspiring data scientists. In real-world applications of data science and AI, the ability to logically build models, understand theoretical underpinnings, and translate abstract mathematical ideas into real-world applications is far more valuable than simply applying pre-existing formulas.
Case 1
Let's just come to an example. A Korean student at SIAI tried his dissertation on a set of data from shipping company's use of tools like containers, boxes, baskets, and folklifts. Unless the data is only for a few clients of the shipping company, it was expected that there will be a number of one-time clients whose use of tools will unlikely be repeated in out of sample data. The student, despite learning that RNN can only be applied to time series without non-stationary movements, was not able to link the learned math concept to RNN and the data. He suffered from gradient's divergence, and tried to control the parameters of RNN instead of 'cleaning' the data itself.
Case 2
Addtionally, many Asian students are too busy jumping on code lines rather than accessing the problem set's background description. In the introductory math and stat courses (STA501, STA502, STA503), we emphasize a lot about how important the data generating process (DGP) can be, like whether the e-commerce company's daily visitor data being from matured incumbents like Amazon or a start-up looking for next round funding. Like case 1, your application of RNN can be challenged depending on how actively the company is engaged in promotions. Little differences in question's setting is thoroughly designed by professors as the change requires an entirely different set of data scientific tools. Many Asian students struggle to understand why an Instrumental Variable (IV) has to be replaced just because the start-up's series-C funding is postponed, for instance. If the company does not need a short-term boost in website visitors, reference data points should remove exploding ups and downs for next month's projection, isn't it?
Cases like this occur a lot among Asian students whose course grade is high enough for us to trust their mastery in skills. And unfortunately, they end up poor performance at the dissertation stage.
Then, is it really a necessary skill? Isn't just an application of previous project's code lines good enough?
AI may soon replace first two types of 'Math Genius'
The rise of AI tools like ChatGPT and other advanced language models is further shifting the definition of mathematical proficiency. While traditional education has emphasized procedural fluency and structured problem-solving, AI can now perform these tasks instantly. Routine calculations, algebraic manipulations, and even structured problem-solving techniques are increasingly automated, reducing the necessity for individuals to master these skills manually.
As AI continues to evolve, it is likely that calculator-type mathematicians and even structured problem-solvers will find themselves increasingly displaced. These AI systems can solve equations, optimize parameters, and generate step-by-step solutions for a wide range of mathematical problems more efficiently than humans. This transformation raises a fundamental question: What kind of mathematical thinking remains irreplaceable?
One of the key limitations of AI in mathematics is its reliance on pattern matching. Despite their computational power, AI tools do not “understand” mathematics in the same way humans do. They recognize patterns in vast datasets and generate responses based on probabilistic relationships rather than true logical reasoning or deep abstraction. Mathematical creativity, proof construction, and conceptual modeling remain beyond the reach of AI, as these require forming genuinely novel insights rather than simply retrieving and recombining existing information.
For this reason, the focus of mathematical education should shift toward logical reasoning, model formulation, and critical evaluation of AI-generated outputs. While AI can provide solutions, human expertise is required to assess their correctness, interpret results, and apply them meaningfully within different contexts. In fields such as AI and Data Science, those who master abstract thinking and theoretical modeling will remain indispensable, while those who rely solely on procedural problem-solving may find their skills increasingly redundant.
Conclusion: Redefining Mathematical Proficiency for AI and Data Science
As mathematics continues to evolve as a discipline, educational institutions must adapt their teaching methodologies to prepare students for both theoretical and applied domains. Traditional views of mathematical ability—whether based on calculation speed or structured problem-solving—must be expanded to include logical reasoning, conceptual understanding, and model applicability.
For students entering AI and Data Science, the ability to think abstractly is crucial for research, but applied roles require a balance between problem-solving skills and an understanding of mathematical conditions. By designing curricula that acknowledge these distinctions, our institution ensures that graduates are equipped to excel in both academic and industry settings.
By aligning mathematical training with practical applications, educators can bridge the gap between traditional perceptions of math proficiency and the skills required for success in the modern AI-driven economy.
In short, SIAI teaches most unlikely replaceable data science tools in AI.
Picture
Member for
1 year 1 month
Real name
David O'Neill
Bio
David O’Neill is a Professor of Finance and Data Analytics at the Gordon School of Business, SIAI. A Swiss-based researcher, his work explores the intersection of quantitative finance, AI, and educational innovation, particularly in designing executive-level curricula for AI-driven investment strategy. In addition to teaching, he manages the operational and financial oversight of SIAI’s education programs in Europe, contributing to the institute’s broader initiatives in hedge fund research and emerging market financial systems.
Forecasting Demand for Public Bicycling in Seoul to Optimize Bike Repositioning at Rental Stations
Published
Sungsu Han*
* Swiss Institute of Artificial Intelligence, Chaltenbodenstrasse 26, 8834 Schindellegi, Schwyz, Switzerland
Abstract
This study aims to improve operational efficiency by optimizing the relocation of Seoul's public bicycles, “ Seoul Bike.” The highlights of the study include Analyzing the usage patterns of Seoul Bike: We analyzed the bicycle travel patterns between business and residential areas, focusing on rush hour.
Introduced the concept of spatiotemporal equilibrium: We proposed the concept of “equilibrium,” where borrowing and returning are balanced in a daily cycle, to provide a basis for relocation strategies.
Developed a data-driven prediction model: We used a SARIMAX model to predict usage by rental location, taking into account variables such as weather, time of day, and location. Optimized relocation strategy: Introduced the D-Index to identify borrowing stations that needed to be relocated, and clustering using the Louvain algorithm to establish efficient relocation zones.
Visualization tool development: We created a visualization tool to intuitively understand bicycle movement patterns and relocation strategies by time of day. This study suggests ways to increase the efficiency of public bicycle operations and improve user satisfaction through a data-driven scientific approach. The results of this study are expected to help Seoul Bike, Seoul reduce operating costs and improve service quality.
This paper will be organized as follows. For the sake of readability, I'm going to divide this article which may seem long and complicated with multiple stories into 5 section.
In the first section, I'll talk about why I became interested in Seoul Bike and the question “Where are the best places to ride Seoul Bike in Seoul?”. In section 2, I would like to talk about the environmental benefits of public bicycle rental in Seoul and what are the biggest problems with this hot item that can solve the health problems of Seoul office workers.
Through this, I will talk about what causes the public bicycle business to always be in the red and analyze the operating costs of the public bicycle industry to lay the groundwork for my thesis topic. In the third section, I will analyze the usage patterns of Seoul Bike from the perspective of users of Seoul Bike in Seoul to lay the groundwork for the core concept of this thesis, equilibrium, which will be discussed in section 4.
In the fourth section, I will discuss specific ideas on how to forecast the usage of each Seoul Bike rental center to solve the deficit of Seoul Bike in Seoul. Finally, we will talk about additional ideas that can maximize the utility of the previous ideas and describe the datasets that can be applied to them.
We'll also talk about the implications of Seoul Bike's work on the Han River bike path and conclude with a discussion of the implications for the future.
Keywords: Seoul Bike, User Patterns, Rental Stations
I am a researcher who commutes from Goyang City to my office near Magok Naru Station in Seoul. One day on my way to work, I rubbed my sleepy eyes as I stepped off the shuttle and was surprised to see hundreds of green bicycles.
I usually commute to work on the company shuttle, but I've recently gotten into cycling, and when the weather is nice, I even bike from home to work. The biggest reason I got into cycling is because of the positive image of Seoul Bike, the city's public bicycle program.
The question “Where did all those hundreds of bicycles at Magok intersection on my way to work come from?” [figure 1-1-1] stuck in my mind for a while, and I realized that I could combine my interest in bicycles with my thesis on public bicycle projects.
[Fig.1-1-1] Seoul Bike Station in Magok
1.2. Problem Statement
Seoul Bike, the city's public bicycle program, has grown rapidly since its introduction in 2015, with great support from Seoul citizens. As of 2023, more than 43,000 bicycles were in operation at more than 2,700 rental stations, and the cumulative number of subscribers exceeded 4 million, making it one of the city's most successful policies. This has had a variety of positive effects, including reducing traffic congestion, addressing air pollution, and improving the health of citizens.
But with this rapid growth has come a number of challenges. The city of Seoul runs a deficit of around 10 billion won every year. This is due to low user fees and high operating costs due to the nature of the public service.
There is a shortage or surplus of bicycles in certain time zones and neighborhoods. This causes inconvenience to users and reduces service efficiency. As of 2019, Seoul Bike had only 60 maintenance staff to manage 25,000 bikes, resulting in an increase in broken bikes and poor management of rental stations. Irresponsible use, vandalism, and loss of bikes are causing additional management costs.
1.3. Aims and Objectives
To address these issues, this study aims to develop a demand prediction model for each Seoul Bike rental station: Develop a model that accurately predicts the demand for each bike rental station by time of day and day of the week by utilizing time series analysis and machine learning algorithms. Based on the predicted demand, we optimize the number of bikes at each rental location to solve the problem of demand-supply imbalance.
Reduce unnecessary relocation costs, increase user satisfaction, and improve overall operational efficiency through accurate demand forecasting and optimized deployment. The developed model provides data-driven decision support for critical operational decisions, such as selecting new rental station locations and deciding to purchase additional bikes.
1.4. Summary of Contributions and Achievements
This study aims to improve operational efficiency by optimizing the relocation of Seoul's public bicycles, “Seoul Bike.” The highlights of the study include analyzing the usage patterns of Seoul Bike: We analyzed the bicycle travel patterns between business and residential areas, focusing on rush hour.
Introduced the concept of spatiotemporal equilibrium: We proposed the concept of “equilibrium,” where borrowing and returning are balanced in a daily cycle, to provide a basis for relocation strategies.
Developed a data-driven prediction model: We used a SARIMAX model to predict usage by rental location, taking into account variables such as weather, time of day, and location. Optimized relocation strategy: Introduced the D-Index to identify borrowing stations that needed to be relocated, and clustering using the Louvain algorithm to establish efficient relocation zones.
Visualization tool development: We created a visualization tool to intuitively understand bicycle movement patterns and relocation strategies by time of day.
This study suggests ways to increase the efficiency of public bicycle operations and improve user satisfaction through a data-driven scientific approach. The results of this study are expected to help Seoul Bike, Seoul reduce operating costs and improve service quality.
2. Solution Approach
2.1. Best Place to Ride a Seoul Bike
According to the Seoul 2022 Transport Usage Statistics Report, Gangseo-gu is the most populated bike traffic district in Seoul, and within Gangseo-gu, seven locations near the Magok business district ranked among the top.
This is probably due to the nature of the Magok business district. Magok is a large, recently developed business complex with a high concentration of office workers. As a new development, the area is well equipped with infrastructure such as cycle paths. Its proximity to the Han River bike path makes it convenient for both commuting and leisure use.
Bicycle commuting is becoming increasingly popular due to growing environmental and health concerns. On the other hand, Goyang City's public bike "Fifteen" was closed due to losses, showing that it is not easy to ensure profitability in the public bike business. In the case of Seoul Bike, the company runs a deficit of more than 10 billion won every year, so it needs to find ways to operate sustainably.
[Fig.2-1-1] Seoul Bike Station in Magok
2.2. The Cause of The Public Bike Projects' Deficit
The issue of deficits in the public bicycle business and the analysis of relocation costs is a very important topic.
Reducing the cost of relocating bicycles is likely to be the most effective way to reduce the deficit of the public bicycle project. Analyzing the usage patterns of users in Seoul Bike to optimize the relocation of bicycles will contribute to reducing the deficit.
Based on this analysis, it is important to explore efficient operational measures to increase the sustainability of the public bicycle business. Along with establishing specific strategies to reduce relocation costs, it is likely that a variety of approaches will be needed, including engaging users and diversifying revenue.
[Fig 2-2-1] Deficit status of public bicycle business Nubiza (Changwo): KRW 4.5 billion, Tashu (Daejeon): KRW 3.6 billion, Tarangae (Gwangju): KRW 1 billion, Eoulling (Sejong): KRW 0.6 billion, Seoul Bike (Seoul): Over KRW 10.3 billion[Fig 2-2-2] Operating cost structure of Go-yang City case The cost of relocating bikes is the largest component, accounting for around 30 per cent of total operating costs. Relocation costs of KRW 525 million, about 30% of total maintenance costs of KRW 1.78 billion. Relocation costs of KRW 525 million: On-site distribution - KRW 375 million, Relocation vehicle operating costs - KRW 150 million
2.3. Time of Day Usage Patterns of Seoul Bike Users
First, let's look at the usage patterns by time of day. On weekdays, the usage is concentrated at 7-8am and 5-6pm, which are commuting hours, suggesting that Seoul Bike is mainly used for commuting. However, on weekends, the usage is relatively evenly distributed from 12 noon to 6 pm, suggesting that it is used for leisure activities or going out. Next, let's look at the user characteristics. In terms of age, users are mainly in their 20s and 30s. This shows that Seoul Bike is becoming a popular means of transport for young people. In addition, the purpose of use is mainly commuting on weekdays and leisure activities on weekends.
[Fig.2-3-1] Seoul Bike Usage Rate by age group
Next, we look at the most important regional clusters. Regional clustering is characterized by a tendency for Seoul Bike to be concentrated in large business districts, areas with well-developed infrastructure such as bike paths, and areas with easy access to the Han River. In particular, it is more popular in areas with easy access to subway stations and workplaces. Representative areas include Magok District in Gangseo-gu, Songpa-gu (Lotte World Tower in Jamsil), Yeongdeungpo-gu (Yeouido business district), and Seongdong-gu .
2.4. Summary
Cycling has become a popular mode of transport for commuting and leisure activities, especially among young people, and tends to be concentrated in large business districts and areas with well-developed cycling infrastructure. These findings suggest that cycling is more than just a mode of transport, but is influencing urban transport systems and lifestyles. Taken together, these findings suggest that cycling policies need to take into account the characteristics of each neighbourhood and focus management and investment on areas with high usage. It is also necessary to consider changes in demand by time of day, with bicycle deployment needing to be adjusted to meet different demands during rush hour and on weekends.
3. Methodology
As we mentioned in Chapter 2, we were able to find out that among the newly developed business districts in Seoul, the business districts where bicycle paths are well-constructed and where access to the Han River bicycle path is easy are the ones where Seoul Bike usage is concentrated.
We will start the analysis by focusing on the super-large business districts that are the cause of the concentration of Seoul Bike usage.
So we have come up with a hypothesis. It is a concept called equilibrium. The idea is that if rentals and returns are equilibrium based on the rental station, there is no need for additional reallocation. To this end, the goal of this paper is to reduce the deficit by improving the efficiency of Seoul Bike reallocation by predicting bicycle rentals and returns.
3.1. Correlation of 24-hour weather data with rental demand
First, in order to confirm the correlation between the Seoul Bike rental data and the weather, pre-processing was performed to integrate Seoul Bike usage information and weather based on the time variable. In addition, in order to identify the phenomenon of rush hour and rush hour, the time was divided as follows.
The preprocessed data below is data from January to December 2022 at the rental office at Exit 2 of Magoknaru Station. From the leftmost column, the time variables are month/day rental time rentals and day of the week, and the weather variables are temperature, wind direction, wind speed, and accumulated precipitation. In addition, Seoul Bike usage information was added to the right column and preprocessed.
[Fig.3-1-1] Seoul Bike User Data + Weather Observation Data
Let's first look at the correlation between the preprocessed data and weather data based on the number of users per minute. First, when looking at the wind speed, the stronger the wind, the more the usage tends to decrease. The highest Seoul Bike usage was at the 3m/s level, which can be felt as a gentle breeze rather than no wind at all.
[Fig.3-1-2] Correlation with Seoul Bike User Data Weather Observation Data
Also, since the accumulated precipitation is inversely proportional to the number of users, it seems to match the general idea that people will not ride bicycles as much as they would use Seoul Bike when it rains. Lastly, the temperature was the highest when people think it is a little cool, around 15~17 degrees, but it was confirmed that the usage was low when it was too cold or too hot.
The correlation with temperature is expected to have seasonality as the Seoul Bike rental volume is correlated with the season. Therefore, we plan to conduct a time series analysis using SARIMAX using external weather variables for the remaining residuals after dividing them into seasonality and trend using STL Decomposition.
3.2. Seoul Bike Daily Return Pattern
In this part, we will analyze the return volume data of Seoul Bike rental stations for 4 years to check the seasonality of the return volume of Seoul Bike rental stations and perform STL Decomposition.
Before conducting a time series analysis to predict the demand for bicycles by Seoul Bike rental station, we drew the daily return volume of Seoul Bike behind Exit 5 of Magoknaru Station from 2019 to 2023. The reason why we did not select Exit 3 of Magoknaru Station, which has the largest number, is because it is a recently built station and there is no data for 19-20 years, so we selected the rental station behind Exit 5 of Magoknaru Station.
We drew a graph of the return and rental during commuting hours of the rental station behind Exit 5 of Magoknaru Station and the cumulative amount, which is the value of the return minus the rental. You can see that the number of bicycles increases as the years go by, and that there is seasonality with the lowest usage in winter and the highest usage in spring and fall.
[Fig.3-2-1] Correlation with Seoul Bike User Data Weather Observation Data
3.3. Modeling of Seoul bike cumulative volume during commuting hours
The number of rentals and returns is symmetrical overall, but you can see that it is biased over time. (There are more returns during commuting hours: returns > rentals)
[Fig.3-3-1] Return and Rent numbers in Go work Time
The cumulative amount representing the number of bicycles in the rental station is expressed as the difference between the number of returns and the number of rentals, and has a trend and seasonality. This means that the use of STL Decomposition, a pre-processing step for time series analysis, is necessary.
The following diagram is a graphical representation of the cumulative amount during go work Time in the business district.
[Fig.3-3-2] Return and Rent numbers in Go work Time
And when you look at the cumulative amount, you can see that the number of returns is a very large proportion compared to the number of rentals, and there is an excess of bicycles.
[Fig.3-3-3] Cumulative number of bicycles during Go work Time
This suggests that the bikes in the business area should be distributed to other areas, leaving only the minimum number of bikes needed in the early morning hours.
As expected from the graphs above, there is a significant concentration of bikes rented in the morning hours from residential districts to business districts.
This could be a reason to have more racks in business areas than in residential districts.
[Fig.3-3-4] Seoul Bike moving Pattern in Go work time
3.4. Bike Demand Forecasting by using SARIMAX Modeling
3.4.1. Return amount in go work time - STL Decomposition results
The return amount of the rental office behind Exit 5 of Magok Naru, which we looked at earlier, was decomposed into Seasonal Trend Loess (Local regression). This is the result of extracting the trend and periodicity included in the time series data through STL Decomposition with the period set to 7 days and using the Multiplicative Option.
[Fig.3-4-1] Return amount of Seoul Bike in Go work Time - STL Decomposition Results
We conducted time series analysis using SARIMAX on the residuals from the STL Decomposition discussed above.
We added weather data and time variables and used them as external variables. When we modeled the residuals from the STL Decomposition using SARIMAX, the results are as follows.
[Fig.3-4-2] Return residual of Seoul bike in Go work Time - SARIMAX Results
When the ADF Test was conducted separately, the results showed that stationary was secured, and when looking at the Ljung Box results after running SARIMAX, the P-Value was greater than 0.05, so there was no autocorrelation, and the Heteroskedasticity test results also showed that there was no heteroscedasticity because the P-Value was greater than 0.05. The results when using 90\% of the 4-year return volume data of the rental office behind Exit 5 of Magoknaru Station during rush hour as Train Data and predicting the remaining 10\% are as follows.
When the residuals with secured normality were predicted using the SARIMAX model with the cumulative precipitation and wind direction humidity return years as external variables, it was confirmed that heteroscedasticity was eliminated, and when the return amount was predicted using this modeling, it was confirmed that the R2 value was approximately 0.73.
[Fig.3-4-3] SARIMAX Modeling Block Diagram
3.5. Logic for improving the efficiency of bicycle relocation
[Fig.3-5-1] Idea based on the Seoul Bike Rental Station
So, the idea that we have thought about based on the Seoul Bike rental stations we have seen so far is the concept of equilibrium. If we group a day into 1 cycle based on commuting hours, it is a concept that a state of equilibrium of bicycle increase and decrease comes. If we summarize it, it can be explained as shown in the figure below.
In addition, the non-equilibrium state is a pattern that can be confirmed by limiting it to the commuting and returning hours of the business area. The non-equilibrium state 1 (commuting hours) where there is an excess of bicycles during the morning commuting hours when returns are the main thing, and the non-equilibrium state 2 (commuting hours) where there is a shortage of bicycles during the afternoon commuting hours when rentals are the main thing can be diagrammed as shown below.
Ultimately, the core logic of this paper is to predict the user factor and determine the optimal redistribution factor in order to resolve the temporal Seoul Bike non-equilibrium of commuting and returning hours.
[Fig.3-5-2] Idea based on the Seoul Bike Rental Station
To summarize the content so far, the areas where Seoul Bike reallocation is most necessary are areas with a large imbalance in bicycle rentals and returns in the super-large business districts where the most rentals and returns occur during commuting hours. (Example: Gang-seo LGSP)
Since people who ride Seoul Bike to work are more likely to ride Seoul Bike home, it would be more efficient to place them only in areas where demand is expected to be excessively high in the afternoon and a shortage is expected, rather than evenly reallocating bicycles that were crowded in the morning by utilizing the equilibrium state.
Also, since a shortage is more likely to cause dissatisfaction with Seoul Bike use than an excess, it is important to predict a shortage rather than an excess, which is the summary of the ideas derived so far.
3.6. Seoul Bike cost efficiency solution according to 1-Day Index Range
In order to create a logic for selecting the rental stations to be relocated, the 1 Day Seoul Bike equilibrium state Index can be selected as the ratio of the expected rental volume and the expected return volume, and the equilibrium state and non-equilibrium state can be divided.
[Fig.3-6-1] Logic for selecting the minimum number of rental stations
The wider the range of the Index 1 is selected, the more rental stations are excluded from the relocation target, which reduces the cost. In addition, the non-equilibrium state is divided into excess and shortage states, and the relocation is carried out mainly in the shortage state, so that the Seoul Bike bike relocation is carried out in a way that maximizes customer satisfaction.
In order to verify this logic for selecting the rental stations to be relocated to the Seoul Bike, we first looked at the distribution of the Seoul Bike Index and the cumulative amount. The distribution of the D Index and the cumulative number for June 2023 was expressed as a violin plot and histogram.
[Fig.3-6-2] D_Index & Cumulative Number of Seoul Bike for June 2023
The cumulative number, which is the return amount minus the rental amount, mostly had a ratio of 0, and the D Index, which is the ratio of the return amount to the rental amount, had a ratio of 1.
This can be seen as confirming a certain degree of equilibrium between the rental and return amounts of the Seoul Bike mentioned earlier. Therefore, I visualized the efficiency of Seoul Bike reallocation using the temporal equilibrium state through the D Index mentioned earlier.
The reduction in Seoul Bike reallocation cost using the D Index Range can be flexibly managed by flexibly adjusting the ratio of the Seoul Bike D Index as shown in the figure below.
[Fig.3-6-3] Decrease in relocation costs by utilizing D_Index Range
As we have seen above, efficient operation is possible by increasing the Index Range excluding relocation and reducing the relocation cost, and it is also possible to operate by flexibly adjusting the Index ratio according to the allocated cost of the Seoul City Seoul Bike budget.
The example presented below is visualized data when the Seoul City operation unit is divided using K-Means Clustering and the D Index Range is flexibly operated from 0.1 to 0.4.
This is the result of diagramming the entire Seoul City using the same method as the Gangseo-gu visualization data confirmed earlier.
In the case of the minimum Index Range excluding relocation target rental stations of 0.1, the ratio of excluded rental stations is 13.4\% and the ratio of target rental stations is 86\%, which is a high ratio. This case will be the most expensive and the relocation efficiency is bound to be low.
[Fig.3-6-4] The ratio of rental station count to relocation according to the D Index Range
In the case where the Index Range for the target rental stations is 0.4, which is the maximum, the ratio of rental stations excluded from the target rental station relocation is 66.6\%, and the ratio of rental stations targeted for relocation is low at 33.4\%.
In this case, the cost of relocation for Seoul bike will be greatly reduced, and the relocation efficiency can also be the highest.
If the ratio of rental stations relocated according to the D Index Range is operated with a focus on practicality, the Index ratio can be flexibly adjusted according to the Seoul Metropolitan Government's Seoul bike budget.
As we have seen above, the ratio of rental stations excluded and rental stations targeted has a trade-off tendency, and the more the ratio of rental stations excluded using the Seoul bike D Index is increased, the fewer bicycles will be targeted for relocation, which can reduce the relocation cost.
3.7. Seoul Bike operation plan idea using spatial equilibrium
Using the spatial equilibrium examined so far, we have come to think that clustering is necessary to minimize the distance of bicycle redistribution movement by using the distribution of total returns and rentals by district, and clustering that has a partial sum where business districts and residential districts are grouped as a pair.
Since there is a distribution of total returns and rentals by district, we thought about how about managing the redistribution districts by grouping them according to the overall increase and decrease.
Therefore, if the selection of districts according to the increase and decrease by district is selected so that the total daily cumulative number of bicycles within the district is close to 0, there is no need to exchange Seoul Bike between the set districts for redistribution, and the effect of reducing the distance of redistribution bicycles within the district area can be achieved, which leads to improved redistribution efficiency overall because the travel distance of redistribution trucks is reduced.
[Fig.3-7-1] The ratio of rental station bike count to relocation according to the D Index Range
3.7.1. Clustering application idea for implementing spatial equilibrium of Seoul Bike
To represent the spatial equilibrium as a mathematical vector, we can think of moving from a rental station node to a return station node as a link.
With this idea, we converted Seoul bike rental information into graph data and then came up with the idea of finding a community partition set with the most links.
[Fig.3-7-2] Network analysis idea for spatial equilibrium analysis of Seoul bikes
3.7.2. What is Network Community Detection?
The problem of dividing a graph into several clusters is called Community Detection.
While thinking about the algorithm to apply to the idea I thought of earlier, I thought about using a cluster detection algorithm that can be used for Seoul Bike movement network data.
The basic concept is that ‘groups with high connection density are tied together’, and the algorithm that calls a group that is tied together and more tightly tied (with high modularity) a Community is defined as Network Community Detection.
[Fig.3-7-4] Community partition clustering in the sense of spatial equilibrium
3.7.3. What is Modularity?
Here, we will learn more about the definition of Modularity, a concept that can quantify the concentration of Network Nodes.
First, the definition of Modularity quantifies the degree to which Nodes are concentrated based on Links in a network structure. The purpose is to measure the structure of a structurally divided network (graph), and the degree of modularity of the network can be expressed as a value of -1 to 1. A value of around 0.3 to 0.7 can be said to indicate a significant cluster.
It can be said to be an Index with a large value when there are many connections within a distinct group within the network and few connections between groups.
The definition of the Index is defined as the ratio of the density of [links within a community] and [links between communities].
For example, a network with high modularity has dense connections between nodes within a module, so the clustering ratio of the same area is high, and the connections between nodes in other modules are sparse, indicating clustering in other areas, which is expressed as a modularity value close to 0.
Modularity
Quantifies the degree to which nodes are clustered based on links in a network structure
Measures the structure of a structurally divided network (graph)
Indicates the degree of modularity of a network with a value of -1 to 1 (significant clustering of 0.3 to 0.7)
A measure that indicates the property that there are many connections within a distinct group in a network and few connections between groups
: Defined as the ratio of the density of [inter-community links] and [inter-community links]
* A network with high modularity has dense connections between nodes within a module (clustering in the same area) but sparse connections between nodes in different modules (clustering in different areas).
3.7.4. Introduction to Louvain Algorithm
Among the algorithms that utilize modularity, we decided to verify the idea through the Louvain algorithm, a representative algorithm. The Louvain algorithm is divided into Phase 1 and Phase 2.
The purpose of Phase 1 is to maximize modularity. The method is to measure modularity by placing a node in another adjacent community and determine the cluster so that modularity is maximized.
[Fig.3-7-5] Louvain Algorithm Phase 1
The purpose of Phase 2 is to simplify the network by maximizing modularity in Phase 1. The link weights that were connected between existing communities are merged into a single link, and the links between nodes in new communities are replaced with self-loops to simplify the network.
[Fig.3-7-6] Louvain Algorithm Phase 2
4. Results
4.1. Cumulative number of Seoul Bike rentals in Gangseo-gu according to D Index
The picture you see is a picture showing the cumulative number of Seoul Bike bicycles in Gangseo-gu by rental station location in June 2023, with a D Index of 0.95 to 1.05 and a range of 0.1.
Transparent circles are negative cumulative amounts, indicating places with a large number of rentals, and filled circles are places with a large number of returns. Also, the places with dots are rental stations excluded from the reallocation target.
Currently, only rental stations within the range of 0.1 are excluded, so there are many target rental stations, and the cost is higher than other cases with a large range. (Low reallocation efficiency)
When the range is as wide as 0.4, this is the case where there are the most rental stations excluded from relocation, and in this case, there are the fewest rental stations subject to relocation, so the cost of relocating Seoul Bike is greatly reduced, making it an efficient section.
[Fig.4-1-1] Gangseo-gu rental stations scheduled to relocate bikes (D Index : 0.1)[Fig.4-1-2] Gangseo-gu rental stations scheduled to relocate bikes (D Index : 0.2)[Fig.4-1-3] Gangseo-gu rental stations scheduled to relocate bikes (D Index : 0.3)[Fig.4-1-4] Gangseo-gu rental stations scheduled to relocate bikes (D Index : 0.4)
4.2. Cumulative number of Seoul Bike rentals in Seoul according to D Index
The picture you see is a picture showing the cumulative number of Seoul bikes by rental station location in Seoul as of June 2023. The D index is 0.95~1.05, and the range is 0.1.
Transparent circles indicate places with a large number of rentals with negative cumulative numbers, and filled circles indicate places with a large number of returns. Also, the places with dots are rental stations excluded from the redistribution target.
Currently, only rental stations within the 0.1 range are excluded, so the cost is higher than other cases with many target rental stations and a large range. (Low redistribution efficiency)
[Fig.4-2-1] Seoul rental stations scheduled to relocate bikes (D Index : 0.1)[Fig.4-2-2] Seoul rental stations scheduled to relocate bikes (D Index : 0.2)[Fig.4-2-3] Seoul rental stations scheduled to relocate bikes (D Index : 0.3)[Fig.4-2-4] Seoul rental stations scheduled to relocate bikes (D Index : 0.4)
In the case where the Index Range for the Rental Stations Excluding Relocation Targets is 0.4, the ratio of rental stations excluded from the relocation target for Seoul Bike is 66.6\%, and the ratio of rental stations subject to relocation is low at 33.4\%.
In this case, the cost of Seoul Bike relocation will be greatly reduced, and the relocation efficiency can also be the highest.
4.3. Results of spatial equilibrium implementation through application of Louvain algorithm
As explained in the idea above, the result of setting up Node and Edge using the rental and return data of Seoul Bike and applying the Louvain algorithm showed much better results than K-Means, which simply used the Euclidean coordinates of the rental station location.
When clustering, the cumulative average by region was used as an indicator of performance, and the closer it is to 0, the closer it is to the partial sum, which means that the bicycle is likely to move only within the clustering region.
K-Means is 21.19, while Louvain is 9.23. We were able to confirm that the cumulative average by region was reduced by almost half.
[Fig.4-3-1] Results of spatial equilibrium implementation through application of Louvain algorithm
And the K-Means algorithm shows clustering that ignores the Han River, while the Louvain algorithm clustering seems to reflect the geographical characteristics of Seoul better.
The results below are the results of the Seoul Bike clustering classified by the K-Means and Louvain algorithms, colored on the map of Seoul.
While K-Means shows results that cannot distinguish the Han River, the Louvain algorithm shows the boundary of the Han River much more clearly.
[Fig.4-3-2] Clustering results comparison : Louvain Vs K-Means[Fig.4-3-3] Louvain Result : Go work Time[Fig.4-3-4] Louvain Result : Off work Time
References
1. F. Chiariotti, C. Pielli, A. Zanella and M. Zorzi, "A dynamic approach to rebalancing bike-sharing systems", Sensors
2. M. Dell’Amico, E. Hadjicostantinou, M. Iori and S. Novellani, "The bike sharing rebalancing problem: Mathematical formulations and benchmark instances"
3. P. Yi, F. Huang and J. Peng, "A rebalancing strategy for the imbalance problem in bike-sharing systems"
4. C. M. de Chardon, G. Caruso and I. Thomas, "Bike-share rebalancing strategies patterns and purpose"
5. C. Zhang, L. Zhang, Y. Liu and X. Yang, "Short-term prediction of bike-sharing usage considering public transport: A lstm approach"
6. S. Ruffieux, N. Spycher, E. Mugellini and O. A. Khaled, "Real-time usage forecasting for bike-sharing systems: A study on random forest and convolutional neural network applicability"
Recent developments in the urban landscape—such as sub-urbanization, counter-urbanization, and re-urbanization—have given rise to complicated scenarios where administrative jurisdictions are newly formed and intersect with existing ones. This presents unique challenges to traditional models of tax competition. This paper examines the economic implications of such overlapping jurisdictions, particularly focusing on their impact on tax policy.
Overlapping jurisdictions are characterized by multiple local governments exercising different levels of authority over the same geographic area. These arrangements often merge as responses to address urban sprawl or provide specialized services tailored to local needs. However, they also create a intricate network of fiscal relationships that can lead to inefficiencies in resource allocation and public service delivery.
For instance, local governments in the U.S., including counties, cities, towns, and special districts, have varying degrees of taxing powers. Residents in certain areas might be subject to property taxes levied by their town, county, school district, and special districts (such as fire or library districts), all operating within the same geographic space. This multi-layered governance structure not only affects residents' tax burdens but also influences local governments' decision-making processes regarding tax rates and public service provision. The resulting fiscal landscape provides a rich setting for examining the dynamics of tax competition and cooperation among overlapping jurisdictions.
Traditional models of tax competition, such as Wilson [4] and Zodrow and Mieszkowski [5], typically assume clear demarcations between competing jurisdictions. However, these models do not adequately capture the dynamics of overlapping administrative divisions. In such settings, local governments must navigate not only horizontal competition with neighboring jurisdictions but also a form of vertical competition within the same geographic space.
This paper aims to extend the literature on tax competition by developing a theoretical framework that accounts for the unique characteristics of overlapping jurisdictions. Specifically, we address the following research questions:
How do overlapping administrative divisions affect the strategic tax-setting behavior of local governments?
What are the implications of such overlapping structures for the provision of public goods and services?
How does the presence of overlapping jurisdictions influence the welfare outcomes predicted by traditional tax competition models?
To address these questions, we develop a game-theoretic model that incorporates multiple layers of local government operating within the same geographic space. This approach allows us to analyze the strategic interactions between overlapping jurisdictions and derive insights into the resulting equilibrium tax rates and levels of public good provision.
Our analysis contributes to the existing literature in several ways. First, it provides a formal framework for understanding tax competition in the context of overlapping jurisdictions, which is increasingly relevant in modern urban governance. Second, it offers insights into the potential inefficiencies that arise from such administrative structures and suggests possible policy interventions to mitigate these issues. Finally, it extends the theoretical foundations of fiscal federalism to account for more complex governance arrangements.
Our study addresses the current realities of fiscal federalism in developed economies as well as provides valuable insights for countries where local governments are yet to achieve significant fiscal autonomy. The lessons drawn from this analysis can inform policy discussions on decentralization, local governance structures, and intergovernmental fiscal relations in various contexts.
The remainder of this paper is organized as follows: Section II reviews the relevant literature on tax competition and fiscal federalism. Section III presents our theoretical model and derives key equilibrium results. Section IV discusses the implications of our findings for public policy and urban governance. Section V concludes and suggests directions for future research.
II. Literature Review
Tax competition has been one of the central themes in public economics. Tiebout [1]'s seminal work on local public goods laid the foundation for this field, proposing a model of "voting with feet" where residents moving to jurisdictions offering their preferred combination of taxes and public services. Oates [2] further developed these ideas and presented the decentralization theorem which posits that, under certain conditions, decentralized provision of public goods is welfare-maximizing.
Works of Wilson [4] and Zodrow and Mieszkowski [5] developed the basic tax competition model, where jurisdictions compete for a mobile capital tax base. This model predicts inefficiently low tax rates and underprovision of public goods, which is often referred to as the "race to the bottom." Wildasin [6] further demonstrated that the Nash equilibrium in tax rates is generally inefficient by incorporating strategic interactions between jurisdictions.
Researchers began to consider more diverse institutional settings. Keen and Kotsogiannis [11] analyzed the interaction between vertical tax competition (between different levels of government) and horizontal tax competition (between governments at the same level). Their work demonstrated that in federal systems, the tax rates can be high or low depending on the relative strength of vertical and horizontal tax externalities, contrary to the ``race to the bottom'' prediction of earlier models.
Itaya, Okamura, and Yamaguchi [12] examined tax coordination in a repeated game setting with asymmetric regions. They find that the sustainability of tax coordination depends on the degree of asymmetry between regions and the type of coordination--partial or full. While asymmetries can complicate coordination efforts, the repeated nature of interactions can facilitate cooperation under certain conditions. Their work demonstrates that full tax coordination can be sustained for a wider range of parameters compared to partial coordination.
Building on this, Ogawa and Wang [14] incorporated fiscal equalization into the framework of asymmetric tax competition in a repeated game context. Their findings reveal that fiscal equalization can influence the sustainability of tax coordination, sometimes making it more difficult to maintain. The impact of equalization schemes on tax coordination is contingent on the degree of regional asymmetry and the specific parameters of the equalization policy.
The case of overlapping jurisdictions represents a frontier in tax competition research. While not extensively studied, some works have begun to address such cases. Hochman, Pines, and Thisse [9] developed a model of metropolitan governance with overlapping jurisdictions, showing how this can lead to inefficiencies in public good provision. Esteller-Mor´e and Sol´e-Oll´e [10] analyzed tax mimicking in a setting with overlapping tax bases, finding evidence of both horizontal and vertical interactions.
Game-theoretic approaches have been instrumental in advancing our understanding of tax competition dynamics. Wildasin [6] pioneered the use of game theory in tax competition, modeling jurisdictions as strategic players in a non-cooperative game. This approach demonstrated that the Nash equilibrium in tax rates is generally inefficient, providing a formal basis for the ``race to the bottom'' hypothesis. The work of Itaya, Okamura, and Yamaguchi [12] and Ogawa and Wang [14] further extended this game-theoretic approach to repeated games, offering insights into the possibilities for tax coordination over time.
While these game-theoretic approaches have significantly advanced our understanding of tax competition, they have largely failed to address the complexities of fully overlapping jurisdictions. Most models assume clear boundaries between competing jurisdictions, leaving a gap in our understanding of scenarios where multiple levels of government have taxing authority over the same geographic area.
The welfare implications of tax competition have been a subject of ongoing debate. While the "race to the bottom" hypothesis suggests negative welfare consequences, some scholars have argued for potential benefits. Brennan and Buchanan [3] proposed that tax competition could serve as a check on the excessive growth of government, a view that has found some support in subsequent empirical work (e.g., [13]).
Policy responses to tax competition have also been extensively studied. Proposals range from tax harmonization [7] to the implementation of corrective subsidies [8]. The effectiveness of these measures, particularly in complex settings with overlapping jurisdictions, remains an active area of research.
While the literature on tax competition has made significant strides in understanding the dynamics of fiscal interactions between jurisdictions, several areas warrant further investigation. The case of fully overlapping jurisdictions, in particular, presents a rich opportunity for both theoretical modeling and empirical analysis. This study aims to fill in this gap by accounting for overlapping jurisdictions in traditional game-theoretic models of tax competition.
III. Model
This study extends the tax competition models of Itaya, Okamura, and Yamaguchi [12] and Ogawa and Wang [14] by introducing an overlapping jurisdiction. Our approach is grounded in the Solow growth model, which provides a robust framework for analyzing long-term economic growth and capital accumulation. The Solow model's emphasis on capital accumulation and technological progress makes it suitable for our analysis of tax competition, as these factors influence jurisdictions' tax bases and policy decisions.
The Solow model's assumptions of diminishing returns to capital and constant returns to scale align well with our focus on regional differences in capital endowments and production technologies. Moreover, its simplicity allows for tractable extensions to multi-jurisdiction settings.
A. Setup
We consider a country divided into three regions: two asymmetric regions, $S$ and $L$, and an overlapping region, $O$, which equally overlaps with $S$ and $L$. All regions have independent authority to impose capital taxes. This setup allows us to examine the interactions between horizontal tax competition (between $S$ and $L$) and the unique dynamics introduced by the overlapping jurisdiction $O$. Let us further denote that the regions of $S$ and $L$ that do not overlap with $O$ are $SS$(Sub-$S$) and $SL$(Sub-$L$), respectively, while those that overlap with $O$ are $OS$ and $OL$ (see Figure 1).
FIGURE 1. VISUAL EXPLANATION OF THE HYPOTHETICAL COUNTRY
Here, we make the following key assumptions:
Population: Population is evenly spread across the country. Hence, regions $S$ and $L$ have equal populations. Furthermore, regions $SS$, $SL$, and $O$ have equal populations. This assumption, while strong, allows us to isolate the effects of capital endowment and technology differences.
Labor Supply and Individual Preferences: Residents inelastically supply one unit of labor to firms in their region and have identical preferences. Furthermore, they strive to maximize their utilities given their budget constraints. While this assumption simplifies labor market dynamics, it is reasonable in the short to medium term, especially in areas with limited inter-regional mobility.
Production: Firms in each region produce homogeneous consumer goods and maximize their profits. This assumption allows us to focus on capital allocation without the complications of product differentiation.
Capital Mobility: Capital is perfectly mobile across regions, reflecting the ease of capital movement in economies, especially within a single country.
Asymmetric Endowments and Technology: Regions $S$ and $L$ differ in capital endowments and production technologies. This assumption captures real-world regional disparities and is crucial for generating meaningful tax competition dynamics.
Public Goods Provision: Regions $S$ and $L$ provide generic public goods $G$, while region $O$ provides specific public goods $H$ to the extent that maximizes their representative resident's utilities. This reflects the often-observed division of responsibilities between different levels of government.
These assumptions, while simplifying the real world, allow us to focus on the core mechanisms of tax competition in overlapping jurisdictions. They provide a tractable framework for analyzing the strategic interactions between jurisdictions while capturing key elements of real-world complexity.
B. Production and Capital Allocation
Let $\bar{k}_i$ be the capital endowment per capita for regions $i$ and $\bar{k}$ be the capital endowments per capita of the national economy. For regions $S, L$ and $O$, it can be expressed as follows:
where $\varepsilon \in \left( 0,\ \bar{k} \right\rbrack$ represents asymmetric endowments between regions $S$ and $L$. $\bar{k}_O = \bar{k}$ follows from the assumption that the population is evenly dispersed across the country.
Let $L_i$ and $K_i$ be the labor and capital inputs for production in region $i$. It can be easily inferred that
where $A_i$ and $B_i > 2K_i / L_i$ represent labor and capital productivity coefficients, respectively. Although regions $S$ and $L$ differ in capital production technology, there is no difference in labor production technology, so $A_{L} = A_{S}$ while $B_{L} \neq \ B_{S}$. Note that this function exhibits constant returns to scale and diminishing returns to capital. Furthermore, we assume that sub-regions without overlaps ($SL$ and $SS$) have equivalent technology coefficients with their super-regions ($L$ and $S$). The technology parameter of the overlapping region is a weighted average of $B_S$ and $B_L$, where the weights are the proportion of capital invested from $S$ and $L$.
As mentioned above, capital allocation across regions is determined by profit-maximizing firms and the free movement of capital. Let $\tau_i$ be the effective tax rate for region $i$. Then, we can infer that the real wage rate $w_i$ and real interest rates $r_i$ are:
where $t_i = (1-\alpha_i)\tau_O$ for $i \in \{S, L\}$, $0$ for $i\in \{SS, SL\}$, $\tau_S$ for $i = OS$, and $\tau_L$ for $i = OL$.
The capital market equilibrium for the national economy is reached when the sum of capital demands is equal to the exogenously fixed total capital endowment: $K_S + K_L = 2l\bar{k}$. In equilibrium, the interest rates and capital demanded in each region are as follows: \begin{equation} \begin{aligned} &r^* = \frac{1}{2}\big(\left(B_S + B_L \right) - \left(\tau_S + \tau_L + (2-\alpha_S-\alpha_L)\tau_O\right)\big) - 2\bar{k} \\ &K_S^* = lk_S^* = l\bigg(\bar{k} + \frac{1}{4}\big( (\tau_L - \tau_S - (\alpha_L - \alpha_S)\tau_O ) - (B_L - B_S)\big) \bigg) \\ &K_L^* = lk_L^* = l\bigg(\bar{k} + \frac{1}{4}\big( (\tau_S - \tau_L + (\alpha_L - \alpha_S)\tau_O ) + (B_L - B_S)\big) \bigg) \\ &K_{SS}^* = \frac{2}{3}lk_{SS}^* = \frac{2l}{3}\bigg(\bar{k} + \frac{1}{4}\big( (\tau_L - \tau_S + (2 - \alpha_L - \alpha_S)\tau_O ) - (B_L - B_S)\big) \bigg) \\ &K_{SL}^* = \frac{2}{3}lk_{SL}^* =\frac{2l}{3}\bigg(\bar{k} + \frac{1}{4}\big( (\tau_S - \tau_L + (2 - \alpha_L - \alpha_S)\tau_O ) + (B_L - B_S)\big) \bigg) \\ &K_O^* = \frac{2}{3}lk_O^* = \frac{2l}{3}\left(\bar{k} - \frac{1}{2}(2-\alpha_S-\alpha_L)\tau_O\right) \end{aligned} \end{equation} We denote $B_L - B_S = \theta$, henceforth.
C. Government Objectives and Tax Rates
Given that individuals in the country have identical preferences and inelastically supply one unit of labor to the regional firms, we can infer that all inhabitants receive a common return on capital of $r^*$ eventually, and they use all income to consume private good $c_i$. Hence, the budget constraint for an individual residing in region $i \in \{S, L, O\}$ and the sum of individuals in region $i$ will be
\begin{equation} \begin{aligned} c_i &= w_i^* + r^*\bar{k}_i \\ C_i &= \begin{cases} l(w^*_i + r^*\bar{k}_i) & \text{ for } i \in \{S, L\}\\ \dfrac{2l}{3}(w^*_i + r^*\bar{k}) & \text{ for } i = O \end{cases} \end{aligned} \end{equation}
In addition, we have assumed that the overlapping district is a special district providing a special public good--for example education or health--that the other two districts do not provide. $S$ and $L$ provide their local public goods $G_i$. Then the total public goods provided in region $i$ can be expressed as:
\begin{equation} \begin{aligned} G_i &= \begin{cases} K_i^*\tau_i & \text{ for } i \in \{S, L\}\\ (1-\alpha_S)K_S^*\tau_S + (1-\alpha_L)K_L^*\tau_L & \text{ for } i = O \end{cases} \\ H_i &= \begin{cases} (1-\alpha_i)K_i^*\tau_O & \hspace{1.45in} \text{ for } i \in \{S, L\}\\ K_i^*\tau_i & \hspace{1.45in} \text{ for } i = O \end{cases} \end{aligned} \end{equation}
Accordingly, each government in region $i$ chooses $\tau_i$ such that maximizes the following social welfare function, which is represented as the sum of individual consumption and public good provision:
This objective function captures the trade-off faced by governments between attracting capital through lower tax rates and generating revenue for public goods provision. After solving equation (9), we obtain the reaction functions, i.e. the tax rates at the market equilibrium (see Appendix 1 for details):
The tax rates derived in the previous section represent the optimal response functions for each region. These functions encapsulate each region's best strategy given the strategies of other regions, as each jurisdiction aims to maximize its social welfare function. In essence, these functions delineate the most advantageous tax rate for each region, contingent upon the tax rates set by other regions.
The existence of a Nash equilibrium is guaranteed in our model, as the slopes of the reaction functions are less than unity, satisfying the contraction mapping principle. This ensures that the iterative process of best responses converges to a unique equilibrium point given $\alpha_L$ and $\alpha_S$. The one-shot Nash equilibrium tax rates are given by (see Appendix 2 for details):
These equilibrium tax rates reveal several important insights. First, the tax rates of regions $S$ and $L$ are influenced by the asymmetry in capital endowments ($\varepsilon$) and productivity ($\theta$), as well as the presence of the overlapping jurisdiction $O$. Second, the overlapping jurisdiction's tax rate is solely determined by the average capital endowment ($\bar{k}$) and the proportion of resources allocated from S and L ($\alpha_S$ and $\alpha_L$). Third, When $\alpha_L + \alpha_S = 4/3$, we have $\tau_O^N = 0$, which effectively reduces our model to a scenario without the overlapping jurisdiction.
The Nash equilibrium also yields equilibrium values for the interest rate and capital demanded in each region:
LEMMA 1 (Net Capital Position): The sign of $\Phi \equiv \varepsilon - \frac{\theta}{4}$ determines the net capital position of regions S and L. When $\Phi > 0$, L is a net capital exporter and S is a net capital importer, and vice versa when $\Phi < 0$. PROOF: From equation (12), we can see that: \begin{align*} K_L^N - K_S^N &= l\left(\left(\bar{k} + \frac{1}{2}\left(\varepsilon + \frac{\theta}{4} \right)\right) - \left(\bar{k} - \frac{1}{2}\left(\varepsilon + \frac{\theta}{4} \right)\right)\right) \\ &= l\left(\varepsilon + \frac{\theta}{4}\right) \end{align*} The sign of this difference is determined by $\varepsilon - \frac{\theta}{4} \equiv \Phi$.
LEMMA 2 (Overlapping Jurisdiction’s Effectiveness): The sign of $\Gamma \equiv \frac{3\left( \alpha_{L} + \alpha_{S} \right) - 4}{\left( 2 - \left( \alpha_{L} + \alpha_{S} \right) \right)\left( \alpha_{L} + \alpha_{S} \right)}$ determines the effective tax rate of O. Moreover, $\alpha_{L} + \alpha_{S}$ must be greater than 4/3 for O to provide a positive sum of special public good H. PROOF: From equation (11), we see that the sign of $\tau_O^N$ is determined by the sign of $\Gamma$. The numerator of $\Gamma$ is positive when $\alpha_{L} + \alpha_{S} > 4/3$, and the denominator is always positive for $\alpha_{L} + \alpha_{S} < 2$. Therefore, $\Gamma > 0$ (and consequently $\tau_O^N > 0$) when $\alpha_{L} + \alpha_{S} > 4/3$.
These lemmas provide crucial insights into the dynamics of our model. First, the introduction of the overlapping jurisdiction $O$ does not alter the net capital positions of $S$ and $L$ compared to a scenario without $O$. The capital flow between $S$ and $L$ is determined solely by the relative strengths of their capital endowments ($\varepsilon$) and productivity differences ($\theta$). In addition, the effectiveness of the overlapping jurisdiction in providing public goods is contingent on receiving a sufficient allocation of resources from $S$ and $L$.
These findings contribute to our understanding of tax competition in multi-layered jurisdictional settings and provide a foundation for analyzing the welfare implications of overlapping administrative structures.
IV. Simulations and Results
To better understand the implications of our theoretical model and address the research questions posed in the introduction, we conducted a series of simulations. These simulations allow us to visualize the non-linear relationships between key variables and provide insights into the strategic behavior of jurisdictions in our overlapping tax competition model.
A. Net Capital Positions and Tax Competition Dynamics
Our first simulation focuses on the net capital positions of regions $S$ and $L$, as determined by the parameter $\Phi \equiv \varepsilon - \theta/4$. Figure 2 illustrates how changes in $\Phi$ affect the capital demanded by each region.
As shown in Figure 2, when $\Phi > 0$, region $L$ becomes a net capital exporter, while region $S$ becomes a net capital importer. This result directly addresses our first research question about how overlapping administrative divisions affect strategic tax-setting behavior. The presence of the overlapping jurisdiction $O$ does not alter the net capital positions of $S$ and $L$ compared to a scenario without $O$.
However, it does influence their tax-setting strategies, as evidenced by the Nash equilibrium tax rates in equation (12). These equations show that $S$ and $L$ adjust their tax rates in response to the overlapping jurisdiction $O$ by factors of $\Gamma(1-\alpha_S)\bar{k}$ and $\Gamma(1-\alpha_L)\bar{k}$, respectively. This strategic adjustment demonstrates how the presence of an overlapping jurisdiction alters tax-setting behavior, even when it doesn't change net capital positions.
FIGURE 2. CAPITAL DEMANDED BY REGIONFIGURE 3. REPRESENTATIVE RESIDENTS’ UTILITY FROM PUBLIC GOODS
B. Public Good Provision and Welfare Implications
Then, we examine the utility derived from public goods by representative residents in each region. Figure 3 visualizes these utilities across different values of $\Phi$ and $\tau_O$.
Figure 3 reveals several important insights. First, the utility derived from public goods varies significantly across sub-regions ($SS$, $SL$, $OS$, $OL$), highlighting the complex welfare implications of overlapping jurisdictions. Second, the overlapping region $O$'s tax rate ($\tau_O$) has a substantial impact on the utility derived from public goods, especially in the overlapping sub-regions $OS$ and $OL$. Third, the relationship between $\Phi$ and public good utility is non-linear and differs across regions, suggesting that the welfare implications of tax competition are not uniform. These findings suggest that the presence of an overlapping jurisdiction can lead to heterogeneous welfare effects.
C. The Role of the Overlapping Jurisdiction
Our model and simulations highlight the crucial role played by the overlapping jurisdiction $O$. The tax rate of $O$ ($\tau_O^N = \Gamma\bar{k}$) is determined by the proportion of resources allocated from the primary economies ($\alpha_S$ and $\alpha_L$), or specifically $\Gamma$. This relationship reveals that the overlapping jurisdiction's ability to provide public goods ($H$) is contingent on receiving a sufficient proportion of resources from $S$ and $L$. Specifically, $\alpha_L + \alpha_S$ must exceed $4/3$ for $O$ to provide a positive sum of special public goods.
This finding has important implications for the design of multi-tiered governance systems. It suggests that overlapping jurisdictions need a critical mass of resource allocation to function effectively, which may inform decisions about the creation and empowerment of special-purpose districts or other overlapping administrative structures.
V. Conclusion
This study has examined the dynamics of tax competition in regions with overlapping tax jurisdictions, leveraging game theory to develop a theoretical framework for understanding these administrative structures. By constructing a simplified model and deriving Nash equilibrium conditions, we have identified several key insights that contribute to the existing literature on tax competition. Our analysis reveals that while the introduction of an overlapping jurisdiction does not alter the net capital positions of the primary regions, it leads to strategic adjustments in tax rates. This finding extends the traditional models of tax competition by incorporating the complexities of multi-tiered governance structures.
The effectiveness of the overlapping jurisdiction in providing public goods is found to be contingent on receiving a sufficient allocation of resources from the primary regions. Moreover, our simulations demonstrate that the presence of overlapping jurisdictions can lead to heterogeneous welfare effects across sub-regions, challenging the uniform predictions of traditional tax competition models and suggesting the need for more nuanced policy approaches.
While our study provides valuable insights, it is important to acknowledge its limitations. The use of a simplified model, while allowing for tractable analysis, inevitably omits some real-world complexities. Factors such as population mobility, diverse tax bases, and income disparities among residents were not incorporated into the model. Furthermore, our analysis is static, which may not capture the dynamic nature of tax competition and capital flows over time. The assumption of identical preferences for public goods across all residents may not reflect the heterogeneity of preferences in real-world settings. Additionally, the model assumes perfect information among all players, which may not hold in practice where information asymmetries can influence strategic decisions.
To address these limitations and further advance our understanding of tax competition in a wide array of administrative structures, several avenues for future research are proposed. Developing dynamic models that capture the evolution of tax competition over time, potentially using differential game theory approaches, could provide insights into the long-term implications of overlapping jurisdictions. Incorporating heterogeneous preferences for public goods among residents would allow for a more nuanced examination of how diverse citizen demands affect tax competition and public good provision in overlapping jurisdictions.
Empirical studies using data from regions with overlapping jurisdictions, such as special districts in the United States, could test the predictions of our theoretical model and provide valuable real-world validation. Extending the model to include various policy interventions, such as intergovernmental transfers or tax harmonization efforts, could help evaluate their effectiveness in mitigating potential inefficiencies. Incorporating insights from behavioral economics to account for bounded rationality and other cognitive factors may provide a more realistic representation of tax-setting behavior in complex jurisdictional settings.
In conclusion, this study provides a theoretical foundation for understanding tax competition in regions with overlapping jurisdictions. By highlighting the complex interactions between multiple layers of government, our findings contribute to the broader literature on fiscal federalism and public economics. As urbanization continues and governance structures become increasingly complex, the insights derived from this research can inform policy discussions on decentralization, local governance structures, and intergovernmental fiscal relations. Future work in this area has the potential to significantly enhance our understanding of modern urban governance and contribute to the development of more effective and equitable fiscal policies in multi-tiered administrative structures.
References
[1] C. M. Tiebout, “A pure theory of local expenditures,” Journal of Political Economy, vol. 64, no. 5, pp. 416–424, 1956. [2] W. E. Oates, Fiscal federalism. Harcourt Brace Jovanovich, 1972. [3] G. Brennan and J. M. Buchanan, The power to tax: Analytical foundations of a fiscal constitution. Cambridge University Press, 1980. [4] J. D. Wilson, “A theory of interregional tax competition,” Journal of Urban Economics, vol. 19, no. 3, pp. 296–315, 1986. [5] G. R. Zodrow and P. Mieszkowski, “Pigou, tiebout, property taxation, and the underprovision of local public goods,” Journal of Urban Economics, vol. 19, no. 3, pp. 356–370, 1986. [6] D. E. Wildasin, “Nash equilibria in models of fiscal competition,” Journal of Public Economics, vol. 35, no. 2, pp. 229–240, 1988. [7] R. Kanbur and M. Keen, “Jeux sans fronti`eres: Tax competition and tax coordination when countries differ in size,” American Economic Review, pp. 877–892, 1993. [8] J. A. DePater and G. M. Myers, “Strategic capital tax competition: A pecuniary externality and a corrective device,” Journal of Urban Economics, vol. 36, no. 1, pp. 66–78, 1994. [9] O. Hochman, D. Pines, and J.-F. Thisse, “On the optimality of local government: The effects of metropolitan spatial structure,” Journal of Economic Theory, vol. 65, no. 2, pp. 334–363, 1995. [10] A. Esteller-Mor´e and A. Sol´e-Oll´e, “Vertical income tax externalities and fiscal interdependence: Evidence from the us,” Regional Science and Urban Economics, vol. 31, no. 2-3, pp. 247–272, 2001. [11] M. Keen and C. Kotsogiannis, “Does federalism lead to excessively high taxes?” American Economic Review, vol. 92, no. 1, pp. 363–370, 2002. [12] J.-i. Itaya, M. Okamura, and C. Yamaguchi, “Are regional asymmetries detrimental to tax coordination in a repeated game setting?” Journal of Public Economics, vol. 92, no. 12, pp. 2403–2411, 2008. [13] L. P. Feld, G. Kirchg¨assner, and C. A. Schaltegger, “Decentralized taxation and the size of government: Evidence from swiss state and local governments,” Southern Economic Journal, vol. 77, no. 1, pp. 27–48, 2010. [14] H. Ogawa and W. Wang, “Asymmetric tax competition and fiscal equalization in a repeated game setting,” International Tax and Public Finance, vol. 23, no. 6, pp. 1035–1064, 2016.
APPENDIX 1 - DERIVING REACTION FUNCTIONS
Let us get the partial derivatives that are needed to get the first order condition of social utility function. Starting with the easier ones,
How is Korea’s Blood Supply Maintained? - Effects of the COVID-19 Pandemic, Blood Shortage Periods, and Promotions on Blood Supply Dynamics
Published
Donggyu Kim*
* Swiss Institute of Artificial Intelligence, Chaltenbodenstrasse 26, 8834 Schindellegi, Schwyz, Switzerland
Abstract
This study quantitatively assesses the effects of the COVID-19 pandemic, blood shortage periods, and promotional activities on blood supply and usage in Korea. Multiple linear regression analysis was conducted using daily blood supply, usage, and stock data from 2018 to 2023, incorporating various control variables. Findings revealed that blood supply decreased by 5.11% and blood usage decreased by 4.25% during the pandemic. During blood shortage periods, blood supply increased by 3.96%, while blood usage decreased by 1.98%. Although the signs of the estimated coefficients aligned with those of a previous study [1], their magnitudes differed. Promotional activities had positive effects on blood donation across all groups, but the magnitude of the impact varied by region and gender. Special promotions offering sports viewing tickets were particularly effective. This study illustrates the necessity of controlling exogenous variables to accurately measure their effects on blood supply and usage, which are influenced by various social factors. The findings underscore the importance of systematic promotion planning and suggest the need for tailored strategies across different regions and demographic groups to maintain a stable blood supply.
Blood transfusion is an essential treatment method used in emergencies, for certain diseases, and during surgeries. Blood required for transfusion cannot be substituted with other substances and has a short shelf life. Additionally, it experiences demand surges that are difficult to predict. Therefore, systematic management of blood stock is required. Blood stock is determined by blood usage and the number of blood donors. Thus, understanding the relationship between usage and supply, as well as the effect of promotions, is essential for effective blood stock management.
There is a paucity of quantitative research on blood supply dynamics during crises and the effects of promotional activities. This study aims to address this research gap. Previous studies in Korea on blood donation have primarily focused on qualitative analysis to identify motives for blood donation through surveys.[2][3][4][5] Kim (2015)[6] used multiple linear regression analysis to predict the number of donations for individual donors but used the personal information of each donor as explanatory variables and did not consider time series characteristics. For this reason, understanding the dynamics of the total number of donors was difficult. Kim (2023)[1] studied the impact of the COVID-19 epidemic on the number of blood donations but did not control exogenous variables and types of blood donation.
This study aims to quantify the effects of the COVID-19 pandemic and blood shortage periods on blood supply and usage. Additionally, it measures the quantitative effects of various promotions on the number of blood donors. To achieve these objectives, regression analysis was utilized with control variables, enabling more precise estimations than previous studies. Based on these findings, this paper proposes effective blood management methods.
2. Methodology
2.1. Research Subjects
According to the Blood Management Act[7], blood can be collected at medical institutions, the Korean Red Cross Blood Services, and the Hanmaeum Blood Center. According to the Annual Statistical Report[8], blood donations conducted by the Korean Red Cross accounted for 92% of all blood donations in 2022. This study uses blood donation data from the Korean Red Cross Blood Services, which accounts for the majority of the blood supply.
2.2. Data Sources
The data for the number of blood donors by location utilized in this study were obtained from the Annual Statistical Report on Blood Services.[8] The daily data on the number of blood donors, blood usage, blood stock, and promotion dates were provided by the Korean Red Cross Blood Services[9].
The data for the number of blood donors, blood usage, and blood stock used in the study cover the period from January 1, 2018, to July 31, 2023, and the promotion date data cover the period from January 1, 2021, to July 31, 2023. Temperature and precipitation data were obtained from the Korea Meteorological Administration’s Automated Synoptic Observing System (ASOS) [10].
2.3. Variable Definitions
In this study, the number of blood donors, or blood supply, is defined as the number of whole blood donations at the Korean Red Cross. The COVID-19 pandemic period is defined as the duration from January 20, 2020, (the first case in Korea) to March 1, 2022 (the end of the vaccine pass operation). Red blood cell product stock is defined as the sum of concentrated red blood cells, washed red blood cells, leukocyte-reduced red blood cells, and leukocyte-filtered red blood cell stock.
Blood usage is based on the quantity of red blood cell products supplied by the Korean Red Cross to medical institutions. The regions in the study are divided according to the jurisdictions of the blood centers under the Korean Red Cross Blood Service and do not necessarily coincide with Korea’s administrative districts. Data from the Seoul Central, Seoul Southern, and Seoul Eastern Blood Centers were integrated and used as Seoul.
Weather information for each region is based on measurements from the nearest weather observation station to the blood center, and the corresponding Observation point numbers for each region are listed in Table 1. Public holidays are based on the Special Day Information[11] from the Korea Astronomy and Space Science Institute.
Table 1. Observation Station number for each Blood Center
2.4. Variable Composition
Dependent Variable Plasma donations, 67% of which are used for pharmaceutical raw materials[8], can be imported due to their long shelf life of 1 year.[7] Also, in the case of platelet and multi-component blood donation, 95% of donors are male[8], and a decent number of days have no female blood donors, affecting the analysis. For these reasons, this study used the number of whole-blood donors as the target variable. Additionally, the amount of collected blood is determined by an individual’s physical condition[7], not by preference. Therefore, it is integrated within gender groups.
Explanatory Variables Considering the differences in operating hours of blood donation centers on weekdays, weekends, and holidays as shown in Figure 1, variables indicating the day of week and holiday were considered. Table 2 shows that there are differences between regions on blood donation. Considering this, region was used as a control variable.
Figure 1. Distribution of number of blood donors by day of the week and holiday
The annual seasonal effects are not controlled by the holidays variable alone. for this reason, Fourier terms were introduced as explanatory variables [12].
Where $\operatorname{doy}\text{(01-01-yyyy)} = 0,\dots, \operatorname{doy}\text{(12-31-yyyy)} = 364$
$j$ is a hyperparameter setting the number of Fourier terms added. $j = 1,\dots,6$ terms with optimal AIC[13] were added to the model.
According to Table 2, 70% of all blood donors visit blood donation centers to donate. Therefore, to control the influence of weather conditions[14] that affect the pedestrian volume and the number of blood donors, a precipitation variable was included. Meanwhile, the temperature variable, which has a strong relationship with the season, was already controlled by Fourier terms and was found to be insignificant, so it was excluded from the explanatory variables.
The Public-Private Joint Blood Supply Crisis Response Manual[15] specifies blood usage control measures during crises (Table 4). To reflect the effect of crises in the model, a blood shortage day variable was created and introduced as a proxy for the crisis stage.
A blood shortage day is defined as a day when the daily blood stock is three times less than the average daily usage in the previous year, as well as the following 7 days. The mathematical expression is as follows:
According to Figure 2, population movement decreased during the COVID-19 pandemic, and blood donation was not allowed for a certain period after COVID recovery or vaccination.[7] Bae (2021)[16] showed factors that affected the decrease in blood donors during this period. Kim (2023)[1] also showed that there was a decrease in blood donation during this period. To control the effect of the pandemic, COVID-19 pandemic period dummy variables were introduced.
Figure 2. Daily Population Movement
2.5. Research Method
To analyze the effects of promotions according to region and gender, this study conducted multiple regression analyses. The OLS model was chosen because it clearly shows the linear relationship between variables, and the results are highly interpretable.
Control variables were used to derive accurate estimates. The model’s accuracy was confirmed through the estimated effect of the COVID-19 and blood shortage period variables, consistent with the prior studies.
The proposed model for the number of blood donors is expressed as an OLS model, as shown in Equation (1). The model for blood usage is also defined using the same explanatory variables.
The regression results presented in Tables 6 and 7 reveal patterns in blood supply and usage dynamics. The day of the week significantly influences both supply and demand, also holidays have a substantial negative impact. The high R-squared value (0.902) of the blood usage model suggests that it accounts for most of the variability in blood usage.
While the relatively lower R-squared value (0.657) of the blood supply model indicates that it is influenced by various random social effects.
3.1. Impact of the COVID-19 Pandemic
Table 7 shows that blood usage decreased by 4.25% during the COVID-19 pandemic period. This includes not only the man-made decrease in the supply due to the blood shortage but also the impact on the demand, where surgeries were reduced due to COVID- 19. Table 6 shows that blood supply also decreased by 5.11% during the same period.
3.2. Impact of Blood Shortage Periods
During blood shortage periods, blood usage decreased by 1.98% (Table 7), while blood supply increased by 3.96% (Table 6) due to conservation efforts and increased donations. This result reflects the efforts made by medical institutions to adjust blood usage and the impact of blood donation promotion campaigns in response to blood shortage situations.
The signs of these estimates are consistent with a previous study [1], providing evidence that the model is well-identified.
3.3. Effects of Promotions
The Korean Red Cross employs promotional methods such as additional giveaways and sending blood donation request messages to address blood shortages. Among these methods, the additional giveaway promotion was conducted uniformly across all regions over a long period, rather than as a one-time event. For this reason, the effect of this promotion was primarily analyzed. Park (2018)[18] showed the impact of promotions on blood donors, but it was based on a survey and had limitations in that quantitative changes could not be measured.
The effect of the additional giveaway promotion on the number of blood donors was confirmed by using promotion days as a dummy variable while controlling the exogenous factors considered earlier. To control the trend effect that could occur due to the clustering of promotion days (Figure 3), the entire period was divided into quarters, and the effect of promotions was measured within each quarter. To prevent outliers that could occur due to data imbalance within the quarter, only periods where the ratio of promotion days within the period ranged from 10% to 90% were used for the analysis. The Seoul, Gyeonggi, and Incheon regions were excluded from the analysis as the promotions were always carried out, making comparison impossible. Due to the nature of the dependent variable being affected by various social factors, the effect of promotions showed variance, but the mean of the distribution was estimated to be positive for all groups (Figure 4, Table 8).
Figure 3. Promotion Dates by Region
In addition to the additional giveaway promotion, the Korean Red Cross conducts various special promotions. Special promotions include all promotions other than the additional giveaway promotion. As special promotions are conducted for short periods, trend effects cannot be eliminated by simply using dummy variables. Therefore, the net effect on the number of blood donors during special promotion periods was measured by comparing the number of donors during the promotion period with two weeks before and
after the promotion. Among various special promotions, the sports viewing ticket give- away promotion showed a high performance in several regions: Gangwon (basketball), Gwangju/Jeonnam (baseball), Ulsan (baseball), and Jeju (soccer) Table 9.
Table 9. Performance of sports viewing ticket giveaway promotions by region
4. Conclusion
Previous studies had the limitation of not being able to quantitatively measure changes in blood stock during the COVID-19 pandemic and blood shortage situations. Furthermore, the effect of blood donation promotions on the number of blood donors in Korea has not been studied.
This study quantitatively analyzed changes in supply and usage during the pandemic and blood shortage situations, as well as the impact of various promotions, using exogenous variables, including time series variables, as control variables. According to the findings of this study, a stable blood supply can be achieved by improving the low promotion response in the Ulsan-si and Jeju-do regions and implementing sports viewing ticket giveaway promotions nationwide.
This study was conducted using short-term regional grouped data due to constraints in data collection. Given that blood donation centers within a region are not homogeneous and the characteristics of blood donors change over time, future research utilizing long-term individual blood donation center data and promotion data could significantly enhance the rigor and granularity of the analysis.
References
[1] Eunhee Kim. Impact of the covid-19 pandemic on blood donation. 2023.
[2] JunSeok Yang. The relationship between attitude on blood donation and altruism of blood donors in gwangju-jeonnam area. 2019.
[3] Jihye Yang. The factor of undergraduate student’s blood donation. 2013.
[4] Eui Yeong Shin. A study on the motivations of the committed blood donors. 2021.
[5] Dong Han Lee. Segmenting blood donors by motivations and strategies for retaining the donors in each segment. 2013.
[6] Shin Kim. A study on prediction of blood donation number using multiple linear regression analysis. 2015.
[7] Republic of Korea. Blood management act, 2023.
[8] Korean Red Cross Blood Services. 2022 annual statistical report on blood services, 2022.
[9] Korean Red Cross Blood Services. Daily data for the number of blood donors, blood usage, blood stock, and promotion dates, 2023.
[11] Korea Astronomy and Space Science Institute. Special day information, 2023.
[12] Peter C Young, Diego J Pedregal, and Wlodek Tych. Dynamic harmonic regression. Journal of forecasting, 18:369–394, 1999.
[13] H Akaike. Information theory as an extension of the maximum likelihood principle. A ́ in: Petrov, bn and csaki, f. In Second International Symposium on InformationTheory. Akademiai Kiado, Budapest, pp. 276A ́281, 1973.
[14] Su mi Lee and Sungjo Hong. The effect of weather and season on pedestrian volume in urban space. Journal of the Korea Academia-Industrial cooperation Society, 20:56–65, 2019.
[15] Republic of Korea. Framework act on the management of disasters and safety, 2023.
[16] Hye Jin Bae, Byong Sun Ahn, Mi Ae Youn, and Don Young Park. Survey on blood donation recognition and korean red cross’ response during covid-19 pandemic. TheKorean Journal of Blood Transfusion, 32:191–200, 2021.
[17] Statistics Korea. Sk telecom mobile data, 2020.
[18] Seongmin Park. Effects of blood donation events on the donors’ intentions of visit in ulsan. 2018.
Data Scientific Intuition that defines Good vs. Bad scientists
Picture
Member for
1 year 2 months
Real name
Keith Lee
Bio
Keith Lee is a Professor of AI and Data Science at the Gordon School of Business, part of the Swiss Institute of Artificial Intelligence (SIAI), where he leads research and teaching on AI-driven finance and data science. He is also a Senior Research Fellow with the GIAI Council, advising on the institute’s global research and financial strategy, including initiatives in Asia and the Middle East.
Published
Modified
Many amateur data scientists have little respect to math/stat behind all computational models Math/stat contains the modelers' logic and intuition to real world data Good data scientists are ones with excellent intuition
We get that they are 'wannabe' data scientists with passion, motivation, and dream with self-confidence that they are the top 1%. But the reality is harsh. So far, less than 5% applicants have been able to pass the admission exam to MSc AI/Data Science's longer version. Almost never we have applicants who are ready to do the shorter one. Most, in fact, almost all students should compromise their dream and accept the reality. The fact that the admision exam is the first two courses of the AI MBA, lowest tier program, already bring students to senses that over a half of applicants usually disappear before and after the exam. Some students choose to retake the exam in the following year, but mostly end up with the same score. Then, they either criticize the school in very creative ways or walk away with frustrated faces. I am sorry for keeping such high integrity of the school.
Sourece: ChatGPT
Data Scientific Intuition that matters the most
The school focuses on two things in its education. First, we want students to understand the thought processes of data science modelers. Support Vector Machine (SVM), for example, reflects the idea that fitting can be more generalized if a separating hyperplane is bounded with inequalities, instead of fixed conditions. If one can understand that the hyperplane itself is already a generalization, it can be much easier to see through why SVM was introduced as an alternative to linear form fitting and what are the applicable cases in real life data science exercises. The very nature of this process is embedded in the school's motto, 'Rerum Cognoscere Causas' ((Felix, qui potuit rerum cognoscere causas - Wikipedia)), meaning a person pursuing the fundamental causes.
The second focus of the school is to help students where and how to apply data science tools to solve real life puzzles. We call this process as the building data scientific instuition. Often, math equations in the textbooks and code lines in one's program console screens do not have any meaning, unless it is combined in a way to solve a particular problem in a peculiar context with a specific object. Unlike many amateur data scientists' belief, coding libraries have not democratized data science to untrained students. In fact, the codes copied by the amateurs are evident examples of rookie failures that data science tools need must deeper background knowledge in statistics than simple code libraries.
Our admission exam is designed to weed out the dreamers or amateurs. After years of trials and errors, we have decided to give a full lecture of elementary math/stat course to all applicants so that we can not only offer them a fair chance but also give them a warning as realistic as our coursework. Previous schooling from other schools may help them, but the exam help us to see if one has potential to develop 'Rerum Cognoscere Causas' and data scientific intuition.
Intution does not come from hard study alone
When I first raised my voice for the importance of data scientific intution, I had had severe conflicts with amateur engineers. They thought copying one's code lines from a class (or a github page) and applying it to other places will make them as good as high paid data scientists. They thought these are nothing more than programming for websites, apps, and/or any other basic programming exercises. These amateurs never understand why you need to do 2nd-stage-least-square (2SLS) regression to remove measurement error effects for a particular data set in a specific time range, just as an example. They just load data from SQL server, add it to code library, and change input variables, time ranges, and computer resources, hoping that one combination out of many can help them to find what their bosses want (or what they can claim they did something cool). Without understanding the nature of data process, which we call 'data generating process' (DGP), their trials and errors are nothing more than higher correlation hunting like untrained sociologists do in their junk researches.
Instead of blaming one code library worse performing than other ones, true data scientists look for embedded DGP and try to build a model following intuitive logic. Every step of the model requires concreate arguments reflecting how the data was constructed and sometimes require data cleaning by variable re-structuring, carving out endogeneity with 2SLS, and/or countless model revisions.
It has been witnessed by years of education that we can help students to memorize all the necessary steps for each textbook case, but not that many students were able to extend the understanding to ones own research. In fact, the potential is well visible in the admission exam or in the early stage of the coursework. Promising students always ask why and what if. Why SVM's functional shape has $1/C$ which may limit the range of $C$ in his/her model, and what if his/her data sets with zero truncation ends up with close to 0 separating hyperplane? Once the student can see how to match equations with real cases, they can upgrade imaginative thought processes to model building logic. For other students, I am sorry but I cannot recall successful students without that ability. High grades in simple memory tests can convince us that they study hard, but lack of intuition make them no better than a textbook. With the experience, we design all our exams to measure how intuitive students are.
Source= Reddit
Intuition that frees a data scientist
In my Machine Learning class for tree models, I always emphasize that a variable with multiple disconnected effective ranges in trees has a different spanned space from linear/non-linear regressions. One variable that is important in a tree space, for example, may not display strong tendency in linear vector spaces. A drug that is only effective to certain age/gender groups (say 5~15, 60~ for male, 20~45 female) can be a good example. Linear regression hardly will capture the same efffective range. After the class, most students understand that relying on Variable Importances of tree models may conflict with p-value type variable selections in regression-based models. But only students with intuition find a way to combine both models that they find the effective range of variables from the tree and redesign the regression model with 0/1 signal variables to separate the effective range.
The extend of these types of thought process is hardly visible from ordinary and disqualified students. Ordinary ones may have capacity to discern what is good, but they often have hard time to apply new findings to one's own. Disqualified students do not even see why that was a neat trick to the better exploitation of DGP.
What's surprising is that previous math/stat education mattered the least. It was more about how logical they are, how hard-working they are, and how intuitive they are. Many students come with the first two, but hardly the third. We help them to build the third muscle, while strenghtening the first. (No one but you can help the second.)
The re-trying students ending up with the same grades in the admission exam are largely because they fail to embody the intuition. It may take years to develop the third muscle. Some students are smart enough to see the value of intuition almost right away. Others may never find that. For failing students, as much as we feel sorry for them, we think that their undergraduate education did not help them to build the muscle, and they were unable to build it by themselves.
The less chanllenging tier programs are designed in a way to help the unlucky ones, if they want to make up the missing pieces from their undergraduate coursework. Blue pills only make you live in fake reality. We just hope our red pill to help you find the bitter but rewarding reality.
Picture
Member for
1 year 2 months
Real name
Keith Lee
Bio
Keith Lee is a Professor of AI and Data Science at the Gordon School of Business, part of the Swiss Institute of Artificial Intelligence (SIAI), where he leads research and teaching on AI-driven finance and data science. He is also a Senior Research Fellow with the GIAI Council, advising on the institute’s global research and financial strategy, including initiatives in Asia and the Middle East.
AI Pessimism, just another correction of exorbitant optimism
Picture
Member for
1 year 1 month
Real name
Ethan McGowan
Bio
Ethan McGowan is a Professor of Financial Technology and Legal Analytics at the Gordon School of Business, SIAI. Originally from the United Kingdom, he works at the frontier of AI applications in financial regulation and institutional strategy, advising on governance and legal frameworks for next-generation investment vehicles. McGowan plays a key role in SIAI’s expansion into global finance hubs, including oversight of the institute’s initiatives in the Middle East and its emerging hedge fund operations.
Published
Modified
AI talks turned the table and become more pessimistic It is just another correction of exorbitant optimism and realisation of AI's current capabilities AI can only help us to replace jobs in low noise data Jobs needing to find new patterns and from high noise data industry, mostly paid more, will not be replaceable by current AI
There have been pessimistic talks about the future of AI recently that have created sudden drops in BigTech firms' stock prices. In all of a sudden, all pessimistic talks from Investors, experts, and academics in reputed institutions are re-visited and re-evaluated. They claim that ROI (Return on Investment) for AI is too low, AI products are too over-priced, and economic impact by AI is minimal. In fact, many of us have raised our voices for years with the exactly same warnings. 'AI is not a magic wand'. 'It is just correlation but not causality / intelligence'. 'Don't be overly enthusiastic about what a simple automation algorithm can do'.
As an institution with AI in our name, we often receive emails from a bunch of 'dreamers' that they wonder if we can make a predictive algorithm that can foretell stock price movements with 99.99% accuracy. If we could do that, why do you think we would share the algorithm with you? We should probably keep it for secret and make billions of dollars just for ourselves. As much as the famous expression by Milton Friedman, a Nobel economist, there is no such thing as a free lunch. If we have a perfect predictability and it is widely public, then the prediction is no longer a prediction. If everyone knows the stock A's price goes up, then everyone would buy the stock A, until it reaches to the predicted value. Knowing that, the price will jump to the predicted value, almost instantly. In other words, the future becomes today, and no one gets benefited.
AI = God? AI = A machine for pattern matching
A lot of enthusiasts have exorbitant optimism that AI can overwhelm human cognitivie capacity and soon become god-like feature. Well, the current forms of AI, be it Machine Learning, Deep Learning, and Generative AI, are no more than a machine for pattern matching. You touch a hot pot, you get a burn. It is painful experience, but you learn that you should not touch when it is hot. The worse the pain, the more careful you become. Hopefully it does not make your skin irrecoverable. The exact same pattern works for what they call AI. If you apply the learning processes dynamically, that's where Generative AI comes. The system is constantly adding more patterns into the database.
Though the extensive size of patterns does have great potential, it does not mean that the machine has cognitive capacity to understand the pattern's causality and/or to find new breakthrough patterns from list of patterns in the database. As long as it is nothing more than a pattern matching system, it never will.
To give you an example, can it be used what words you are expected to answer in a class that has been repeated for thousand times? Definitely. Then, can you use the same machine to predict the stock price? Aren't the stock market repeating the same behavior over a century? Well, unfortunately it is not, thus you can't be benefited by the same machine for financial investments.
Two types of data - Low noise vs. High noise
On and near the Wall Street, you can sometimes meet an excessively confident hedge fund manager with claims on near perfect foresight for financial market movements. Some of them have outstanding track records, and surprisingly persuasive. In New York Times archive back in 1940s, or even as early as 1910s, you can see people with similar claims were eventually sued by investors, arrested due to false claims, and/or just disappeared from the street within a few years. If they were that good, why then they lost money and got sued/arrested?
There are two types of data. One set of data that you can see from machine (or highly controlled environment) is called 'Low-noise' data. It has high predictability. Even in cases where embedded patterns are invisible by bare eyes, you either need more analytic brain or a machine to test all possibilities within the possible sets. For the game of Go, the brain was Se-dol Lee and the machine was Alpha-Go. The game needs to test 19x19 possible sets with around 300 possible steps. Even if your brain is not as good as Se-dol Lee, as long as your computer can find the winning patterns, you can win. This is what has been witnessed.
The other set of data comes from largely uncontrolled environment. There potentially is a pattern, but it is not the single impetus that drives every motion of the space. There are thousands, if not millions, of patterns that the driver is not observable. This is where randomness is needed for modeling, and it is unfortunately impossible to predict accurate move, because the driver is not observable. We call this set of data 'High-noise'. The stock market is the very example of such. There are millions of unknown, unexpectable, and at least unmeasurable influences that disable any analyst or machine to predict with accuracy level upto 100%. This is why financial models are not researched for predictability but used only to backtest financial derivatives for reasonable pricing.
Natural language process (NLP) is one example of low noise. Our language follows a certain set of rules (or patterns), which are called grammar. Unless you are uneducated or intentionally out of grammar (or make mistakes), people generally follow grammar. Weather is mostly low noise, but it has high noise components. Sometimes typhoons are unpredictable, or less predictable. Stock market? Be my guest. There have been 4 Nobel Prizes given to financial economists by year 2023, and all of them are based on the belief that stock markets follow random processes, be it Gaussian, Poisson, and/or any other unknown random distributions. (Just in case, if a process follows any known distribution, that means it is probabilistic, which means it is random.)
Pessimism / Photo by Mizuno K
Potential benefits of AI
We as an institution hardly believe current forms of AI will make any significant changes in businesses and our life in short term. The best we can expect is automation of mundane tasks. Like laundary machine in early 20th century. ChatGPT already has shown us a path. Soon, CS operators will largely be replaced by LLM based chatbots. US companies actively outsourced the function from India for the past a few decades, thanks to cheaper international connectivity via internet. It will still remain, but human actions will be needed way less than before. In fact, we already get machine generated answers from a number of international services. If we complain about a program's malfunction on a WordPress plugin, for instance, established services email us machine answers first. For a few cases, it actually is enough. The practice will become more popular to less-established services as it becomes easier and cheaper to implement.
Teamed up with EduTimes, we also are working on a research to replace 'Copy Boys/Girls'. Journalists that we know from large news magazines are not always running on the street to find new and fascinating stories. In fact, most of them read other newspapers and rewrite the contents as if they were the original sources. Although it is not an important job, it is still needed for the newspaper to run. They need to keep up the current events, accoring to the EduTimes journalists from other renouned newspapers. The copy team is usually paid the least and seen a death sentence as a journalist. What makes the job more sympathetic on top of the least respect, it will soon be replaced by LLM based copywriters.
In fact, any job that generates patterned contents without much of cognitivie functions will gradually be replaced.
What about automotive driving? Is it a low-noise pattern job or a high-noise complicated cognitive job? Well, although Elon Musk claims high possibility of Lv. 4 auto-driving within next a few years, we don't believe so. None of us at GIAI have seen any game theorists have solved multi-agent ($n$>2) Bayesian belief game with imperfect information and unknown agent types by computer so that the automotive driving algorithm can predict what other drivers on the road will do. Without the right prediction of others on the fast moving vehicles, it is hard to tell if your AI will help you successfully avoid other crazy drivers. The driving job for those eventful cases needs 'instinct', which requires another set of bodily function different from cognitive intelligence. The best that the current algorithm can do is to tighten it up to perfection for a single car, which already needs to go over a lot of mathematical, mechanical, organisational, legal, and commercial (and many more) challenges.
Don't they know all that? Aren't the Wall Street investors self-confident, egocentric, but ultra smart that they already know all the limitations of AI? We believe so. At least we hope so. Then, why do they pay attention to the discontentful pessimism now, and create heavy drops in tech stock prices?
Guess the Wall Street hates to see Silicon Valley to be paid too much. American East often think the West too unrealistic and floating in the air. OpenAI's next round funding may surprise us in a totally opposite direction.
Picture
Member for
1 year 1 month
Real name
Ethan McGowan
Bio
Ethan McGowan is a Professor of Financial Technology and Legal Analytics at the Gordon School of Business, SIAI. Originally from the United Kingdom, he works at the frontier of AI applications in financial regulation and institutional strategy, advising on governance and legal frameworks for next-generation investment vehicles. McGowan plays a key role in SIAI’s expansion into global finance hubs, including oversight of the institute’s initiatives in the Middle East and its emerging hedge fund operations.