Solving Problems with Algorithms and AI
Main contact
Timeline
-
October 20, 2025Experience start
-
November 2, 2025Sprint 1: Project Brief & Requirements Analysis
-
November 9, 2025Sprint 2: Open-Source LLM Swap
-
November 23, 2025Sprint 3: Validation & Evaluation
-
December 2, 2025Sprint 4: Presentation & Submission
-
December 2, 2025Experience end
Timeline
-
October 20, 2025Experience start
-
November 2, 2025Sprint 1: Project Brief & Requirements Analysis
Deliverable: Project Initiation Document
The deliverable is a Project Initiation Document (PID), a formal report submitted to the instruction team and the company stakeholders. This document will serve as a foundational blueprint for your project and must include the following three components:
-
Framework Deployment Confirmation: A screenshot demonstrating the successful deployment and operational status of the provided backend and frontend starter framework. -
LLM Strategy Plan: A comprehensive plan for the selection of an appropriate open-source Large Language Model (LLM). This plan should justify your choice based on project needs (e.g., performance, licensing, size). -
Key Term Alignment Design: A detailed design proposal outlining the methodology for aligning the company's specified key business terms with the structured data extracted from the relevant forms.
Purpose: Foundational Alignment and Technical Readiness
This milestone is critical to establishing a clear and correct starting point for the project. It serves two primary purposes:
-
Strategic Alignment: To ensure that students immediately develop a solid and accurate understanding of the company’s specific business needs and requirements for term alignment. -
Technical Foundation: To confirm the technical readiness of the student by verifying their ability to successfully deploy and utilize the starter framework developed by a previous high-performing master's student.
-
-
November 9, 2025Sprint 2: Open-Source LLM Swap
Tasks: LLM Selection and Framework Integration
The following tasks must be completed by each individual team member:
-
Open-Source LLM Selection: Select two distinct open-source Large Language Models (LLMs) from the provided list. The selection must be based on suitability for the member's designated computing environment (personal machine or university cluster).In teams of two or more members, individual LLM selections may overlap. However, the team must collectively select a minimum of one more LLM than the number of team members ($\text{K}+1$ unique LLMs for $\text{K}$ members). -
Framework Adaptation: Modify the existing starter framework to replace the placeholder commercial LLM API calls with local, on-device inference using the two selected open-source LLMs. -
UI Verification: Successfully execute and demonstrate the functionality of the revised web User Interface (UI) for each integrated open-source LLM separately. -
Data Processing and Extraction: Individually upload a designated sample data form through the revised web UI for each of the two integrated LLMs, ensuring the successful extraction of structured data. -
Output Capture: Save the resulting structured data output from each individual execution as a separate JSON file.
Deliverable: Source Code and Comparative Extraction Results
The formal deliverable includes both the technical assets and the resulting structured data:
-
Revised Source Code: The complete, functional source code for the starter framework, modified to successfully run inference with the two selected open-source LLMs. -
Comparative JSON Results: A total of six (6) structured JSON files per team member capturing the output of the data extraction process. (Assuming one sample form is provided, the total is two (2) JSON files per member—one for each selected LLM).
Purpose: Core Skill Development and Functional Validation
The primary goal of this milestone is two-fold:
-
Technical Proficiency: To develop practical, hands-on experience in local Large Language Model (LLM) deployment and the process of integrating open-source models into an existing application framework. -
Functional Validation: To functionally validate the revised starter framework's ability to perform structured data extraction using different open-source models, thereby providing a preliminary comparative analysis of their respective performance on a common task.
-
-
November 23, 2025Sprint 3: Validation & Evaluation
Tasks: Iterative Refinement and Benchmarking
The core activities for this sprint involve assessing initial performance, optimizing the model strategy, and rigorously documenting all experimental data:
-
Digitalization Result Assessment: Conduct a thorough evaluation and qualitative analysis of the structured data output (the initial digitalization results) generated in the previous sprint. -
LLM Strategy Refinement: Based on the assessment, refine the open-source LLM selection or tuning strategy. This may involve exploring new models, optimizing existing model hyperparameters, or improving prompt engineering techniques. -
Experimentation and Documentation: Rigorously record all experiments and results for every LLM used by the individual or the team. This documentation must include the specific model version, any prompt templates used, and the raw extraction output (JSON files) or the calculated evaluation metrics.
Deliverable: Optimized System and Comprehensive Evaluation Report
The formal deliverables will demonstrate both the functional improvement of the system and the analytical rigor of the team:
-
Refined Source Code: The optimized source code of the starter framework, incorporating all necessary changes for the refined LLM strategy (e.g., new models, prompt engineering, or post-processing logic). -
Comprehensive Evaluation Report: A detailed report containing the quantitative evaluation results. This report must include the calculated performance metrics (which will be specified by the instruction team) and clearly demonstrate the method used to align the extracted data schema tree with the company's designated key business terms.
Purpose: Performance Enhancement and Advanced Skill Acquisition
The primary goal of this milestone is to elevate technical proficiency beyond basic integration:
To improve AI system development skills by moving into the optimization phase. This involves exploring advanced techniques such as prompt engineering optimization, comparative benchmarking across diverse open-source LLMs, and formalizing the data alignment process to meet functional performance requirements.
-
-
December 2, 2025Sprint 4: Presentation & Submission
Deliver final report, code, and presentation to company.
Reflect on outcomes, lessons learned, and possible extensions.
-
December 2, 2025Experience end
Experience scope
Categories
Software development Machine learning Artificial intelligenceSkills
computational thinking algorithms algorithm design problem solving artificial intelligence big data large language modeling machine learning methods computer systems artificial intelligence systemsOur graduate learners bring strong skills in algorithm design, data structures, and computational problem-solving. They are capable of analyzing complex challenges, developing efficient solutions, and applying modern approaches—including AI and big data techniques—to real-world problems. Companies can expect motivated learners who can deliver computational thinking-oriented insights, prototypes, and optimized solutions to support decision-making and innovation.
Learners
Team size may not exceed 4
Expected Outcomes for Learners:
Learners will strengthen their ability to analyze complex problems, design and implement efficient algorithms, and apply modern data structures, databases, and AI techniques. They will gain hands-on experience working with real-world data and industry challenges, improving both technical and professional skills.
Deliverables for the Employer:
Employers can expect clear documentation of problem analysis, algorithm design, and solution implementation, along with data-driven insights, computational thinking, prototypes, or visualizations tailored to the project. Final deliverables may include a written report, an application form, code, and a presentation summarizing findings and recommendations.
Project timeline
-
October 20, 2025Experience start
-
November 2, 2025Sprint 1: Project Brief & Requirements Analysis
-
November 9, 2025Sprint 2: Open-Source LLM Swap
-
November 23, 2025Sprint 3: Validation & Evaluation
-
December 2, 2025Sprint 4: Presentation & Submission
-
December 2, 2025Experience end
Timeline
-
October 20, 2025Experience start
-
November 2, 2025Sprint 1: Project Brief & Requirements Analysis
Deliverable: Project Initiation Document
The deliverable is a Project Initiation Document (PID), a formal report submitted to the instruction team and the company stakeholders. This document will serve as a foundational blueprint for your project and must include the following three components:
-
Framework Deployment Confirmation: A screenshot demonstrating the successful deployment and operational status of the provided backend and frontend starter framework. -
LLM Strategy Plan: A comprehensive plan for the selection of an appropriate open-source Large Language Model (LLM). This plan should justify your choice based on project needs (e.g., performance, licensing, size). -
Key Term Alignment Design: A detailed design proposal outlining the methodology for aligning the company's specified key business terms with the structured data extracted from the relevant forms.
Purpose: Foundational Alignment and Technical Readiness
This milestone is critical to establishing a clear and correct starting point for the project. It serves two primary purposes:
-
Strategic Alignment: To ensure that students immediately develop a solid and accurate understanding of the company’s specific business needs and requirements for term alignment. -
Technical Foundation: To confirm the technical readiness of the student by verifying their ability to successfully deploy and utilize the starter framework developed by a previous high-performing master's student.
-
-
November 9, 2025Sprint 2: Open-Source LLM Swap
Tasks: LLM Selection and Framework Integration
The following tasks must be completed by each individual team member:
-
Open-Source LLM Selection: Select two distinct open-source Large Language Models (LLMs) from the provided list. The selection must be based on suitability for the member's designated computing environment (personal machine or university cluster).In teams of two or more members, individual LLM selections may overlap. However, the team must collectively select a minimum of one more LLM than the number of team members ($\text{K}+1$ unique LLMs for $\text{K}$ members). -
Framework Adaptation: Modify the existing starter framework to replace the placeholder commercial LLM API calls with local, on-device inference using the two selected open-source LLMs. -
UI Verification: Successfully execute and demonstrate the functionality of the revised web User Interface (UI) for each integrated open-source LLM separately. -
Data Processing and Extraction: Individually upload a designated sample data form through the revised web UI for each of the two integrated LLMs, ensuring the successful extraction of structured data. -
Output Capture: Save the resulting structured data output from each individual execution as a separate JSON file.
Deliverable: Source Code and Comparative Extraction Results
The formal deliverable includes both the technical assets and the resulting structured data:
-
Revised Source Code: The complete, functional source code for the starter framework, modified to successfully run inference with the two selected open-source LLMs. -
Comparative JSON Results: A total of six (6) structured JSON files per team member capturing the output of the data extraction process. (Assuming one sample form is provided, the total is two (2) JSON files per member—one for each selected LLM).
Purpose: Core Skill Development and Functional Validation
The primary goal of this milestone is two-fold:
-
Technical Proficiency: To develop practical, hands-on experience in local Large Language Model (LLM) deployment and the process of integrating open-source models into an existing application framework. -
Functional Validation: To functionally validate the revised starter framework's ability to perform structured data extraction using different open-source models, thereby providing a preliminary comparative analysis of their respective performance on a common task.
-
-
November 23, 2025Sprint 3: Validation & Evaluation
Tasks: Iterative Refinement and Benchmarking
The core activities for this sprint involve assessing initial performance, optimizing the model strategy, and rigorously documenting all experimental data:
-
Digitalization Result Assessment: Conduct a thorough evaluation and qualitative analysis of the structured data output (the initial digitalization results) generated in the previous sprint. -
LLM Strategy Refinement: Based on the assessment, refine the open-source LLM selection or tuning strategy. This may involve exploring new models, optimizing existing model hyperparameters, or improving prompt engineering techniques. -
Experimentation and Documentation: Rigorously record all experiments and results for every LLM used by the individual or the team. This documentation must include the specific model version, any prompt templates used, and the raw extraction output (JSON files) or the calculated evaluation metrics.
Deliverable: Optimized System and Comprehensive Evaluation Report
The formal deliverables will demonstrate both the functional improvement of the system and the analytical rigor of the team:
-
Refined Source Code: The optimized source code of the starter framework, incorporating all necessary changes for the refined LLM strategy (e.g., new models, prompt engineering, or post-processing logic). -
Comprehensive Evaluation Report: A detailed report containing the quantitative evaluation results. This report must include the calculated performance metrics (which will be specified by the instruction team) and clearly demonstrate the method used to align the extracted data schema tree with the company's designated key business terms.
Purpose: Performance Enhancement and Advanced Skill Acquisition
The primary goal of this milestone is to elevate technical proficiency beyond basic integration:
To improve AI system development skills by moving into the optimization phase. This involves exploring advanced techniques such as prompt engineering optimization, comparative benchmarking across diverse open-source LLMs, and formalizing the data alignment process to meet functional performance requirements.
-
-
December 2, 2025Sprint 4: Presentation & Submission
Deliver final report, code, and presentation to company.
Reflect on outcomes, lessons learned, and possible extensions.
-
December 2, 2025Experience end
Project examples
Data Optimization: Designing algorithms to optimize logistics, scheduling, or resource allocation.
Big Data Analysis: Applying algorithms to extract insights from large or complex datasets.
AI-Enhanced Solutions: Implementing machine learning or AI techniques for prediction, classification, or recommendation.
Search & Matching Systems: Building efficient search, ranking, or recommendation algorithms for real-world applications.
Network & Graph Analysis: Applying graph algorithms to social networks, transportation systems, or supply chains.
Data Structure Design: Developing customized data structures for performance-critical applications.
Process Automation: Creating algorithmic workflows that improve efficiency or reduce manual effort.
Visualization & Reporting Tools: Turning raw data into meaningful dashboards and visual insights.
Additional company criteria
Companies must answer the following questions to submit a match request to this experience:
-
Q1 - Text short
-
Q2 - Text short
-
Q3 - Text short
-
Q4 - Text short
-
Q5 - Text short
Main contact
Timeline
-
October 20, 2025Experience start
-
November 2, 2025Sprint 1: Project Brief & Requirements Analysis
-
November 9, 2025Sprint 2: Open-Source LLM Swap
-
November 23, 2025Sprint 3: Validation & Evaluation
-
December 2, 2025Sprint 4: Presentation & Submission
-
December 2, 2025Experience end
Timeline
-
October 20, 2025Experience start
-
November 2, 2025Sprint 1: Project Brief & Requirements Analysis
Deliverable: Project Initiation Document
The deliverable is a Project Initiation Document (PID), a formal report submitted to the instruction team and the company stakeholders. This document will serve as a foundational blueprint for your project and must include the following three components:
-
Framework Deployment Confirmation: A screenshot demonstrating the successful deployment and operational status of the provided backend and frontend starter framework. -
LLM Strategy Plan: A comprehensive plan for the selection of an appropriate open-source Large Language Model (LLM). This plan should justify your choice based on project needs (e.g., performance, licensing, size). -
Key Term Alignment Design: A detailed design proposal outlining the methodology for aligning the company's specified key business terms with the structured data extracted from the relevant forms.
Purpose: Foundational Alignment and Technical Readiness
This milestone is critical to establishing a clear and correct starting point for the project. It serves two primary purposes:
-
Strategic Alignment: To ensure that students immediately develop a solid and accurate understanding of the company’s specific business needs and requirements for term alignment. -
Technical Foundation: To confirm the technical readiness of the student by verifying their ability to successfully deploy and utilize the starter framework developed by a previous high-performing master's student.
-
-
November 9, 2025Sprint 2: Open-Source LLM Swap
Tasks: LLM Selection and Framework Integration
The following tasks must be completed by each individual team member:
-
Open-Source LLM Selection: Select two distinct open-source Large Language Models (LLMs) from the provided list. The selection must be based on suitability for the member's designated computing environment (personal machine or university cluster).In teams of two or more members, individual LLM selections may overlap. However, the team must collectively select a minimum of one more LLM than the number of team members ($\text{K}+1$ unique LLMs for $\text{K}$ members). -
Framework Adaptation: Modify the existing starter framework to replace the placeholder commercial LLM API calls with local, on-device inference using the two selected open-source LLMs. -
UI Verification: Successfully execute and demonstrate the functionality of the revised web User Interface (UI) for each integrated open-source LLM separately. -
Data Processing and Extraction: Individually upload a designated sample data form through the revised web UI for each of the two integrated LLMs, ensuring the successful extraction of structured data. -
Output Capture: Save the resulting structured data output from each individual execution as a separate JSON file.
Deliverable: Source Code and Comparative Extraction Results
The formal deliverable includes both the technical assets and the resulting structured data:
-
Revised Source Code: The complete, functional source code for the starter framework, modified to successfully run inference with the two selected open-source LLMs. -
Comparative JSON Results: A total of six (6) structured JSON files per team member capturing the output of the data extraction process. (Assuming one sample form is provided, the total is two (2) JSON files per member—one for each selected LLM).
Purpose: Core Skill Development and Functional Validation
The primary goal of this milestone is two-fold:
-
Technical Proficiency: To develop practical, hands-on experience in local Large Language Model (LLM) deployment and the process of integrating open-source models into an existing application framework. -
Functional Validation: To functionally validate the revised starter framework's ability to perform structured data extraction using different open-source models, thereby providing a preliminary comparative analysis of their respective performance on a common task.
-
-
November 23, 2025Sprint 3: Validation & Evaluation
Tasks: Iterative Refinement and Benchmarking
The core activities for this sprint involve assessing initial performance, optimizing the model strategy, and rigorously documenting all experimental data:
-
Digitalization Result Assessment: Conduct a thorough evaluation and qualitative analysis of the structured data output (the initial digitalization results) generated in the previous sprint. -
LLM Strategy Refinement: Based on the assessment, refine the open-source LLM selection or tuning strategy. This may involve exploring new models, optimizing existing model hyperparameters, or improving prompt engineering techniques. -
Experimentation and Documentation: Rigorously record all experiments and results for every LLM used by the individual or the team. This documentation must include the specific model version, any prompt templates used, and the raw extraction output (JSON files) or the calculated evaluation metrics.
Deliverable: Optimized System and Comprehensive Evaluation Report
The formal deliverables will demonstrate both the functional improvement of the system and the analytical rigor of the team:
-
Refined Source Code: The optimized source code of the starter framework, incorporating all necessary changes for the refined LLM strategy (e.g., new models, prompt engineering, or post-processing logic). -
Comprehensive Evaluation Report: A detailed report containing the quantitative evaluation results. This report must include the calculated performance metrics (which will be specified by the instruction team) and clearly demonstrate the method used to align the extracted data schema tree with the company's designated key business terms.
Purpose: Performance Enhancement and Advanced Skill Acquisition
The primary goal of this milestone is to elevate technical proficiency beyond basic integration:
To improve AI system development skills by moving into the optimization phase. This involves exploring advanced techniques such as prompt engineering optimization, comparative benchmarking across diverse open-source LLMs, and formalizing the data alignment process to meet functional performance requirements.
-
-
December 2, 2025Sprint 4: Presentation & Submission
Deliver final report, code, and presentation to company.
Reflect on outcomes, lessons learned, and possible extensions.
-
December 2, 2025Experience end