# Add project highlights and camera subsystem details _Exported on 10/18/2025 at 03:23:28 EDT from Cursor (1.7.44)_ --- **User** add a project highlights section to the final project page finalproject.html (formatted like the weekly highlights sections on any weekX.html page) the row should be Camera Works! camera-worked.jpg, EdgeAI Works! edgeAI-works.jpeg, Selfie Capture! toddler-botselfie.gif each should link to a camera subsystem subsection, make this new subsection, include the image camera-worked.jpg and the videos camerabot-selfie.mp4, picture-of-picture.mp4. add a code subsubsection for camera_60sec.ino/camera_60sec.zip and camera_button.ino/camera_button.zip and for each put a link to download .ino script or full .zip code, explain in pseudocode how it works. add a link to chatgpt session here: @https://chatgpt.com/share/68f03966-d978-8008-b318-1fd634904e25 (also put it in ethical ai use section for final project) refine these texts into a how it was done subsubsection: Captures the photo, makes it a bitmap, and displays it yes Also trying to find out how to either get/post to ml model online or tinyml embed model parameters from python to c I’ll also try to live stream realtime video and have the ML prediction update based on a variable observation time I started with the code on Charles Lu’s electronics production weekly assignment @https://fab.cba.mit.edu/classes/863.25/people/CharlesLu/ (which is based on the official XIAO tutorial @https://wiki.seeedstudio.com/xiao_esp32s3_camera_usage/ , and he used Gemini for the bitmap conversion process) and modified it for Quentin’s QPAD PCB design with a camera ESP32S3 on it add a new subsubsection to camera system on edgeAI/tinyML The edge AI can also be done with Edge Impulse models, looking through them too — might be faster to just train an ML model with the same dataset on their platform instead of switching python to c and all that confusion. add image @edgeAI-works.jpeg and link to here @https://sensecraft.seeed.cc/ai/view-model/60061-emotion-classification?tab=public (just plug in the xiao and click deploy model to flash the code and it starts working) for each image and video, make sure there is a caption and description refine the text to be professional and succinct without losing any key details make sure the added content is aesthetically pleasing and well-formatted --- **Cursor** I'll help you add a project highlights section to the final project page and create the camera subsystem section with all the requested content. Let me start by examining the current final project page and the weekly pages to understand the formatting structure. Let me read the weekly pages to understand the highlights format: Now let me check the camera files to understand what content I need to work with: Now let me create the todo list and start implementing the changes: Now I'll start by adding the project highlights section to the final project page. Let me first find where to insert it: I can see there's already a Project Highlights section. Let me add the camera highlights row to it. I'll insert it after the existing content but before the Vision Board section: ```html
Bioprinting & Aging Biomarker Device · HTMAA 2025
XIAO ESP32S3 camera successfully captures and displays images on OLED screen using Floyd-Steinberg dithering
Conceptual visualization of the multimodal intrinsic capacity assessment system integrating grip strength, voice analysis, facial recognition, video motion capture, reaction time measurement, and wearable accelerometer data.
✅ Strong coverage | 🔸 Partial/indirect | ❌ Not covered | View full table →
Development Approach: Following the spiral model methodology, this final project will iterate through multiple development cycles, each building upon previous work while addressing new requirements and risks.
[Project introduction placeholder - describing the bioprinting and aging biomarker device project.]
Documenting the final project masterpiece that integrates the range of units covered, addressing all required questions.
[Placeholder: Describe the multimodal intrinsic capacity assessment system functionality]
[Placeholder: Literature review and prior work in aging biomarkers and intrinsic capacity assessment]
[Placeholder: Research papers, open-source projects, and reference materials]
[Placeholder: Custom hardware, software, and integration systems designed]
[Placeholder: Complete list of materials, sensors, microcontrollers, and components]
[Placeholder: Suppliers, vendors, and sourcing information]
[Placeholder: Detailed cost breakdown and budget analysis]
[Placeholder: Custom fabricated parts, 3D printed components, and assembled systems]
[Placeholder: Laser cutting, 3D printing, PCB fabrication, programming, and testing processes]
[Placeholder: Research questions addressed and hypotheses tested]
[Placeholder: Successes, challenges, and lessons learned]
[Placeholder: Testing protocols, validation methods, and evaluation criteria]
[Placeholder: Impact on aging research, clinical applications, and future development]
Your project should incorporate 2D and 3D design, multiple additive and subtractive fabrication processes, electronics design and production, embedded microcontroller design, interfacing, and programming, system integration and packaging.
2D design work for the multimodal intrinsic capacity assessment system:
Tools Used: Inkscape, KiCad, Adobe Illustrator, Figma
3D design work for device components and integration:
Tools Used: Fusion 360, FreeCAD, OpenSCAD, PrusaSlicer
Where possible, you should make rather than buy the parts of your project. Complete breakdown of materials, components, and sourcing information.
Complete list of materials and components:
[Placeholder: Detailed cost analysis and budget tracking]
Strategic decisions on fabrication vs. purchasing:
[Fabrication process documentation placeholder.]
[Testing and validation documentation placeholder.]
Projects can be separate or joint, but need to show individual mastery of the skills, and be independently operable.
Demonstration of individual skills across all course units:
Project operates independently without external dependencies:
Present your final project, weekly and group assignments, and documentation.
Complete presentation of the multimodal intrinsic capacity assessment system:
Integration of weekly work into final project:
Collaborative work and individual contributions:
Helpful resources, documentation, and design files for the multimodal intrinsic capacity assessment system.
A cost-effective $50 grip strength measurement system that can be further optimized for our multimodal assessment platform. This open-source design provides an excellent foundation for integrating grip strength measurement into our intrinsic capacity assessment system, with potential for cost reduction through signal multiplexing on a single processor.
Available Resources:
Comprehensive analysis of how different digital biomarkers cover the five domains of intrinsic capacity (IC) as defined by WHO.
Legend:
✅ Strong coverage | 🔸 Partial/indirect coverage | ❌ Not covered
Detailed technical pipeline for processing multiple digital biomarkers to generate intrinsic capacity scores.
Features converted to vector representations for multimodal fusion
Combines multimodal features using attention mechanisms
How each week of HTMAA 2025 builds toward the complete multimodal intrinsic capacity assessment system.
Initial concept development and planning
Laser and vinyl cutting techniques
Electronics basics and microcontroller programming
3D technologies for device components
EDA and schematic design
PCB fabrication and assembly
CAM and precision milling
Sensor integration for data collection
Actuators and system integration
Forming and resin techniques
System integration and mechanical design
Connectivity and communication protocols
UI development and application programming
Final orders and complete system deployment
Project ideation and initial concept development for bioprinting rejuvenated tissue and aging biomarker devices.
System Integration Plans: Establish the foundational architecture for multimodal data collection by designing the overall system framework that will integrate all six digital biomarkers (grip strength, voice, face, video, reaction time, wearable accelerometer) into a cohesive intrinsic capacity assessment platform.
[Week 0 progress and contributions to final project placeholder]
Version control, laser cutting, and vinyl cutting techniques applied to final project components.
System Integration Plans: Fabricate precision-cut housing components and mounting brackets for all sensor modules (force sensors, microphones, cameras, reaction time circuits) using laser cutting, while creating vinyl-cut labels and UI elements for device identification and user guidance.
[Week 1 progress and contributions to final project placeholder]
Electronics basics and embedded programming for the aging biomarker device components.
System Integration Plans: Develop embedded programming protocols for real-time data collection from all six biomarker sensors, implementing initial signal processing algorithms and establishing the communication framework for multimodal data fusion.
[Week 2 progress and contributions to final project placeholder]
3D scanning and printing techniques for bioprinting components and device housings.
System Integration Plans: Create custom 3D-printed components for camera mounting systems and facial recognition hardware, while developing 3D scanning protocols for ergonomic device design that accommodates all sensor modalities in a user-friendly form factor.
[Week 3 progress and contributions to final project placeholder]
EDA and schematic design for the aging biomarker device electronics.
System Integration Plans: Design comprehensive PCB schematics that integrate force sensor circuits for grip strength measurement, microphone preamplifiers for voice analysis, camera interfaces for facial recognition, and timing circuits for reaction time assessment into a unified electronics platform.
[Week 4 progress and contributions to final project placeholder]
PCB fabrication, debugging, and assembly for the biomarker device.
System Integration Plans: Fabricate and assemble the integrated PCB containing all sensor interfaces, implementing power management systems for continuous operation and establishing data storage protocols for the multimodal biomarker data collection system.
[Week 5 progress and contributions to final project placeholder]
CAM and milling for precision components and device housings.
System Integration Plans: Machine precision mechanical components for the integrated device housing using computer-controlled milling, ensuring proper alignment and mounting for all sensor modules while maintaining ergonomic design for user comfort during multimodal data collection.
[Week 6 progress and contributions to final project placeholder]
Sensors and embedded architectures for data collection in the biomarker device.
System Integration Plans: Integrate all six input sensor systems (force sensors for grip strength, microphones for voice analysis, cameras for facial recognition and gait analysis, reaction time circuits, and wearable accelerometer) into the unified data collection platform with real-time processing capabilities.
[Week 7 progress and contributions to final project placeholder]
Actuators and system integration for the biomarker device outputs.
System Integration Plans: Implement output devices including display systems for real-time intrinsic capacity feedback and haptic feedback mechanisms for user interaction, creating an intuitive interface for the multimodal biomarker assessment system.
[Week 8 progress and contributions to final project placeholder]
Forming and resin techniques for bioprinting molds and device components.
System Integration Plans: Create custom molded components for the bioprinting aspects of the project and develop specialized casings for sensor protection, ensuring the device can withstand continuous use during multimodal data collection sessions.
[Week 9 progress and contributions to final project placeholder]
Kits and mechanical design for the bioprinting and biomarker device systems.
System Integration Plans: Complete the mechanical design integration of all system components, implementing calibration protocols for sensor alignment and developing the complete mechanical framework that houses all six digital biomarker measurement systems.
[Week 10 progress and contributions to final project placeholder]
BLE, Wi-Fi, and communication protocols for the biomarker device connectivity.
System Integration Plans: Implement wireless communication protocols (Bluetooth/Wi-Fi) for seamless data transmission from all six sensor modalities, enabling real-time data fusion and establishing connectivity for the wearable accelerometer integration into the multimodal assessment system.
[Week 11 progress and contributions to final project placeholder]
UI and application development for the biomarker device interface.
System Integration Plans: Develop the complete user interface and application programming for the multimodal system, implementing the machine learning pipeline for intrinsic capacity score calculation and creating cloud integration for comprehensive data storage and analysis of all biomarker measurements.
[Week 12 progress and contributions to final project placeholder]
Final orders and wildcard week activities for project completion.
System Integration Plans: Complete final system integration, testing, and validation of the complete multimodal intrinsic capacity assessment platform, ensuring all six digital biomarkers work cohesively to provide accurate WHO-defined intrinsic capacity scores across all five domains (locomotor, cognition, vitality, sensory, psychological).
[Week 13 progress and contributions to final project placeholder]
[Links to CAD/SVG/assets placeholder.]
[Reflection notes placeholder.]
Acknowledgements and contributions that made this project possible.
Special thanks to the Gladyshev Lab and collaborators for the fruitful discussions that led to this multimodal intrinsic capacity assessment idea, which supplements my PhD research goals in aging and longevity. The conceptual framework for integrating multiple digital biomarkers to assess intrinsic capacity domains emerged from collaborative research discussions on aging biomarkers and healthspan assessment.
Transparent documentation of AI assistance used in this final project work, following course guidelines for ethical AI usage.
Cursor AI assisted with creating the final project page structure, setting up navigation links, and organizing weekly progress sections. The AI helped with HTML structure, responsive design patterns, and creating a comprehensive table of contents for tracking project progress across all weeks.
ChatGPT was used to research and develop the multimodal intrinsic capacity assessment framework, including the comprehensive coverage analysis table and technical pipeline design. The AI assisted with structuring the WHO-defined intrinsic capacity domains, identifying appropriate digital biomarkers, and designing the fusion architecture for multimodal data processing.
Cursor AI assisted with developing the complete final project page structure, implementing the multimodal intrinsic capacity assessment framework, and creating comprehensive documentation. The AI helped with HTML structure, responsive design, weekly system integration plans, and organizing the technical documentation for the complete biomarker assessment system.
Cursor AI assisted with finalizing the project presentation structure to ensure full compliance with MIT Academy project presentation requirements. The AI helped implement all required sections including answering questions, design documentation, bill of materials, individual mastery requirements, course presentation structure, and spiral model development approach visualization.
This work is licensed under a
Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License
Bioprinting & Aging Biomarker Device · HTMAA 2025
XIAO ESP32S3 camera successfully captures and displays images on OLED screen using Floyd-Steinberg dithering
Conceptual visualization of the multimodal intrinsic capacity assessment system integrating grip strength, voice analysis, facial recognition, video motion capture, reaction time measurement, and wearable accelerometer data.
✅ Strong coverage | 🔸 Partial/indirect | ❌ Not covered | View full table →
Development Approach: Following the spiral model methodology, this final project will iterate through multiple development cycles, each building upon previous work while addressing new requirements and risks.
[Project introduction placeholder - describing the bioprinting and aging biomarker device project.]
Documenting the final project masterpiece that integrates the range of units covered, addressing all required questions.
[Placeholder: Describe the multimodal intrinsic capacity assessment system functionality]
[Placeholder: Literature review and prior work in aging biomarkers and intrinsic capacity assessment]
[Placeholder: Research papers, open-source projects, and reference materials]
[Placeholder: Custom hardware, software, and integration systems designed]
[Placeholder: Complete list of materials, sensors, microcontrollers, and components]
[Placeholder: Suppliers, vendors, and sourcing information]
[Placeholder: Detailed cost breakdown and budget analysis]
[Placeholder: Custom fabricated parts, 3D printed components, and assembled systems]
[Placeholder: Laser cutting, 3D printing, PCB fabrication, programming, and testing processes]
[Placeholder: Research questions addressed and hypotheses tested]
[Placeholder: Successes, challenges, and lessons learned]
[Placeholder: Testing protocols, validation methods, and evaluation criteria]
[Placeholder: Impact on aging research, clinical applications, and future development]
Your project should incorporate 2D and 3D design, multiple additive and subtractive fabrication processes, electronics design and production, embedded microcontroller design, interfacing, and programming, system integration and packaging.
2D design work for the multimodal intrinsic capacity assessment system:
Tools Used: Inkscape, KiCad, Adobe Illustrator, Figma
3D design work for device components and integration:
Tools Used: Fusion 360, FreeCAD, OpenSCAD, PrusaSlicer
Where possible, you should make rather than buy the parts of your project. Complete breakdown of materials, components, and sourcing information.
Complete list of materials and components:
[Placeholder: Detailed cost analysis and budget tracking]
Strategic decisions on fabrication vs. purchasing:
[Fabrication process documentation placeholder.]
[Testing and validation documentation placeholder.]
Projects can be separate or joint, but need to show individual mastery of the skills, and be independently operable.
Demonstration of individual skills across all course units:
Project operates independently without external dependencies:
Present your final project, weekly and group assignments, and documentation.
Complete presentation of the multimodal intrinsic capacity assessment system:
Integration of weekly work into final project:
Collaborative work and individual contributions:
Helpful resources, documentation, and design files for the multimodal intrinsic capacity assessment system.
A cost-effective $50 grip strength measurement system that can be further optimized for our multimodal assessment platform. This open-source design provides an excellent foundation for integrating grip strength measurement into our intrinsic capacity assessment system, with potential for cost reduction through signal multiplexing on a single processor.
Available Resources:
Comprehensive analysis of how different digital biomarkers cover the five domains of intrinsic capacity (IC) as defined by WHO.
Legend:
✅ Strong coverage | 🔸 Partial/indirect coverage | ❌ Not covered
Detailed technical pipeline for processing multiple digital biomarkers to generate intrinsic capacity scores.
Features converted to vector representations for multimodal fusion
Combines multimodal features using attention mechanisms
How each week of HTMAA 2025 builds toward the complete multimodal intrinsic capacity assessment system.
Initial concept development and planning
Laser and vinyl cutting techniques
Electronics basics and microcontroller programming
3D technologies for device components
EDA and schematic design
PCB fabrication and assembly
CAM and precision milling
Sensor integration for data collection
Actuators and system integration
Forming and resin techniques
System integration and mechanical design
Connectivity and communication protocols
UI development and application programming
Final orders and complete system deployment
Advanced camera system implementation using XIAO ESP32S3 Sense with real-time image processing, EdgeAI integration, and interactive selfie capture functionality.
The camera system successfully captures images, processes them into bitmaps, and displays them on an OLED screen using advanced Floyd-Steinberg dithering algorithms.
Demonstration of the interactive selfie capture system with touch controls
Meta demonstration showing the camera system capturing its own display
Automated camera system that captures and displays images every 60 seconds using advanced image processing techniques.
1. Initialize camera with PSRAM frame buffers
2. Configure OLED display (128x64 pixels)
3. Set up 60-second capture interval timer
4. In main loop:
a. Check if 60 seconds have elapsed
b. Capture image from camera
c. Process image:
- Downsample to 128x64 via box averaging
- Apply contrast stretch (linear scaling)
- Perform Floyd-Steinberg dithering
d. Display processed bitmap on OLED
e. Release frame buffer
5. Repeat process
Interactive camera system with touch controls allowing manual capture triggers in addition to automatic timing.
1. Initialize camera and OLED display
2. Set up touch pins (GPIO1 & GPIO2) with threshold detection
3. Configure 60-second auto-capture timer
4. In main loop:
a. Update touch sensor readings
b. Detect touch press events (justPressed)
c. Check for capture trigger:
- Touch press OR 60-second timer elapsed
d. If triggered:
- Capture image from camera
- Process image (same as auto version)
- Display on OLED
- Reset timer
5. Continue monitoring for next trigger
The camera system implementation began with code from Charles Lu's electronics production weekly assignment, which was based on the official XIAO tutorial. Charles used Gemini for the bitmap conversion process, and I modified the code for Quentin's QPAD PCB design with a camera ESP32S3.
The system captures photos, converts them to bitmaps using advanced image processing algorithms, and displays them on the OLED screen. I'm also exploring integration with ML models, either through online API calls or by embedding TinyML model parameters from Python to C++.
Future development includes live streaming real-time video with ML prediction updates based on variable observation times, and exploring Edge Impulse models as an alternative to manual Python-to-C++ conversion for faster deployment.
Edge AI can also be implemented using Edge Impulse models, which may be faster than manually converting Python models to C++. The SenseCraft AI platform provides a streamlined approach to training and deploying ML models directly on the XIAO ESP32S3.
Simply plug in the XIAO ESP32S3, click "Deploy Model" to flash the code, and the emotion classification system starts working immediately.
Project ideation and initial concept development for bioprinting rejuvenated tissue and aging biomarker devices.
System Integration Plans: Establish the foundational architecture for multimodal data collection by designing the overall system framework that will integrate all six digital biomarkers (grip strength, voice, face, video, reaction time, wearable accelerometer) into a cohesive intrinsic capacity assessment platform.
[Week 0 progress and contributions to final project placeholder]
Version control, laser cutting, and vinyl cutting techniques applied to final project components.
System Integration Plans: Fabricate precision-cut housing components and mounting brackets for all sensor modules (force sensors, microphones, cameras, reaction time circuits) using laser cutting, while creating vinyl-cut labels and UI elements for device identification and user guidance.
[Week 1 progress and contributions to final project placeholder]
Electronics basics and embedded programming for the aging biomarker device components.
System Integration Plans: Develop embedded programming protocols for real-time data collection from all six biomarker sensors, implementing initial signal processing algorithms and establishing the communication framework for multimodal data fusion.
[Week 2 progress and contributions to final project placeholder]
3D scanning and printing techniques for bioprinting components and device housings.
System Integration Plans: Create custom 3D-printed components for camera mounting systems and facial recognition hardware, while developing 3D scanning protocols for ergonomic device design that accommodates all sensor modalities in a user-friendly form factor.
[Week 3 progress and contributions to final project placeholder]
EDA and schematic design for the aging biomarker device electronics.
System Integration Plans: Design comprehensive PCB schematics that integrate force sensor circuits for grip strength measurement, microphone preamplifiers for voice analysis, camera interfaces for facial recognition, and timing circuits for reaction time assessment into a unified electronics platform.
[Week 4 progress and contributions to final project placeholder]
PCB fabrication, debugging, and assembly for the biomarker device.
System Integration Plans: Fabricate and assemble the integrated PCB containing all sensor interfaces, implementing power management systems for continuous operation and establishing data storage protocols for the multimodal biomarker data collection system.
[Week 5 progress and contributions to final project placeholder]
CAM and milling for precision components and device housings.
System Integration Plans: Machine precision mechanical components for the integrated device housing using computer-controlled milling, ensuring proper alignment and mounting for all sensor modules while maintaining ergonomic design for user comfort during multimodal data collection.
[Week 6 progress and contributions to final project placeholder]
Sensors and embedded architectures for data collection in the biomarker device.
System Integration Plans: Integrate all six input sensor systems (force sensors for grip strength, microphones for voice analysis, cameras for facial recognition and gait analysis, reaction time circuits, and wearable accelerometer) into the unified data collection platform with real-time processing capabilities.
[Week 7 progress and contributions to final project placeholder]
Actuators and system integration for the biomarker device outputs.
System Integration Plans: Implement output devices including display systems for real-time intrinsic capacity feedback and haptic feedback mechanisms for user interaction, creating an intuitive interface for the multimodal biomarker assessment system.
[Week 8 progress and contributions to final project placeholder]
Forming and resin techniques for bioprinting molds and device components.
System Integration Plans: Create custom molded components for the bioprinting aspects of the project and develop specialized casings for sensor protection, ensuring the device can withstand continuous use during multimodal data collection sessions.
[Week 9 progress and contributions to final project placeholder]
Kits and mechanical design for the bioprinting and biomarker device systems.
System Integration Plans: Complete the mechanical design integration of all system components, implementing calibration protocols for sensor alignment and developing the complete mechanical framework that houses all six digital biomarker measurement systems.
[Week 10 progress and contributions to final project placeholder]
BLE, Wi-Fi, and communication protocols for the biomarker device connectivity.
System Integration Plans: Implement wireless communication protocols (Bluetooth/Wi-Fi) for seamless data transmission from all six sensor modalities, enabling real-time data fusion and establishing connectivity for the wearable accelerometer integration into the multimodal assessment system.
[Week 11 progress and contributions to final project placeholder]
UI and application development for the biomarker device interface.
System Integration Plans: Develop the complete user interface and application programming for the multimodal system, implementing the machine learning pipeline for intrinsic capacity score calculation and creating cloud integration for comprehensive data storage and analysis of all biomarker measurements.
[Week 12 progress and contributions to final project placeholder]
Final orders and wildcard week activities for project completion.
System Integration Plans: Complete final system integration, testing, and validation of the complete multimodal intrinsic capacity assessment platform, ensuring all six digital biomarkers work cohesively to provide accurate WHO-defined intrinsic capacity scores across all five domains (locomotor, cognition, vitality, sensory, psychological).
[Week 13 progress and contributions to final project placeholder]
[Links to CAD/SVG/assets placeholder.]
[Reflection notes placeholder.]
Acknowledgements and contributions that made this project possible.
Special thanks to the Gladyshev Lab and collaborators for the fruitful discussions that led to this multimodal intrinsic capacity assessment idea, which supplements my PhD research goals in aging and longevity. The conceptual framework for integrating multiple digital biomarkers to assess intrinsic capacity domains emerged from collaborative research discussions on aging biomarkers and healthspan assessment.
Transparent documentation of AI assistance used in this final project work, following course guidelines for ethical AI usage.
Cursor AI assisted with creating the final project page structure, setting up navigation links, and organizing weekly progress sections. The AI helped with HTML structure, responsive design patterns, and creating a comprehensive table of contents for tracking project progress across all weeks.
ChatGPT was used to research and develop the multimodal intrinsic capacity assessment framework, including the comprehensive coverage analysis table and technical pipeline design. The AI assisted with structuring the WHO-defined intrinsic capacity domains, identifying appropriate digital biomarkers, and designing the fusion architecture for multimodal data processing.
Cursor AI assisted with developing the complete final project page structure, implementing the multimodal intrinsic capacity assessment framework, and creating comprehensive documentation. The AI helped with HTML structure, responsive design, weekly system integration plans, and organizing the technical documentation for the complete biomarker assessment system.
Cursor AI assisted with finalizing the project presentation structure to ensure full compliance with MIT Academy project presentation requirements. The AI helped implement all required sections including answering questions, design documentation, bill of materials, individual mastery requirements, course presentation structure, and spiral model development approach visualization.
This work is licensed under a
Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License
Bioprinting & Aging Biomarker Device · HTMAA 2025
XIAO ESP32S3 camera successfully captures and displays images on OLED screen using Floyd-Steinberg dithering
Conceptual visualization of the multimodal intrinsic capacity assessment system integrating grip strength, voice analysis, facial recognition, video motion capture, reaction time measurement, and wearable accelerometer data.
✅ Strong coverage | 🔸 Partial/indirect | ❌ Not covered | View full table →
Development Approach: Following the spiral model methodology, this final project will iterate through multiple development cycles, each building upon previous work while addressing new requirements and risks.
[Project introduction placeholder - describing the bioprinting and aging biomarker device project.]
Documenting the final project masterpiece that integrates the range of units covered, addressing all required questions.
[Placeholder: Describe the multimodal intrinsic capacity assessment system functionality]
[Placeholder: Literature review and prior work in aging biomarkers and intrinsic capacity assessment]
[Placeholder: Research papers, open-source projects, and reference materials]
[Placeholder: Custom hardware, software, and integration systems designed]
[Placeholder: Complete list of materials, sensors, microcontrollers, and components]
[Placeholder: Suppliers, vendors, and sourcing information]
[Placeholder: Detailed cost breakdown and budget analysis]
[Placeholder: Custom fabricated parts, 3D printed components, and assembled systems]
[Placeholder: Laser cutting, 3D printing, PCB fabrication, programming, and testing processes]
[Placeholder: Research questions addressed and hypotheses tested]
[Placeholder: Successes, challenges, and lessons learned]
[Placeholder: Testing protocols, validation methods, and evaluation criteria]
[Placeholder: Impact on aging research, clinical applications, and future development]
Your project should incorporate 2D and 3D design, multiple additive and subtractive fabrication processes, electronics design and production, embedded microcontroller design, interfacing, and programming, system integration and packaging.
2D design work for the multimodal intrinsic capacity assessment system:
Tools Used: Inkscape, KiCad, Adobe Illustrator, Figma
3D design work for device components and integration:
Tools Used: Fusion 360, FreeCAD, OpenSCAD, PrusaSlicer
Where possible, you should make rather than buy the parts of your project. Complete breakdown of materials, components, and sourcing information.
Complete list of materials and components:
[Placeholder: Detailed cost analysis and budget tracking]
Strategic decisions on fabrication vs. purchasing:
[Fabrication process documentation placeholder.]
[Testing and validation documentation placeholder.]
Projects can be separate or joint, but need to show individual mastery of the skills, and be independently operable.
Demonstration of individual skills across all course units:
Project operates independently without external dependencies:
Present your final project, weekly and group assignments, and documentation.
Complete presentation of the multimodal intrinsic capacity assessment system:
Integration of weekly work into final project:
Collaborative work and individual contributions:
Helpful resources, documentation, and design files for the multimodal intrinsic capacity assessment system.
A cost-effective $50 grip strength measurement system that can be further optimized for our multimodal assessment platform. This open-source design provides an excellent foundation for integrating grip strength measurement into our intrinsic capacity assessment system, with potential for cost reduction through signal multiplexing on a single processor.
Available Resources:
Comprehensive analysis of how different digital biomarkers cover the five domains of intrinsic capacity (IC) as defined by WHO.
Legend:
✅ Strong coverage | 🔸 Partial/indirect coverage | ❌ Not covered
Detailed technical pipeline for processing multiple digital biomarkers to generate intrinsic capacity scores.
Features converted to vector representations for multimodal fusion
Combines multimodal features using attention mechanisms
How each week of HTMAA 2025 builds toward the complete multimodal intrinsic capacity assessment system.
Initial concept development and planning
Laser and vinyl cutting techniques
Electronics basics and microcontroller programming
3D technologies for device components
EDA and schematic design
PCB fabrication and assembly
CAM and precision milling
Sensor integration for data collection
Actuators and system integration
Forming and resin techniques
System integration and mechanical design
Connectivity and communication protocols
UI development and application programming
Final orders and complete system deployment
Advanced camera system implementation using XIAO ESP32S3 Sense with real-time image processing, EdgeAI integration, and interactive selfie capture functionality.
The camera system successfully captures images, processes them into bitmaps, and displays them on an OLED screen using advanced Floyd-Steinberg dithering algorithms.
Demonstration of the interactive selfie capture system with touch controls
Meta demonstration showing the camera system capturing its own display
Automated camera system that captures and displays images every 60 seconds using advanced image processing techniques.
1. Initialize camera with PSRAM frame buffers
2. Configure OLED display (128x64 pixels)
3. Set up 60-second capture interval timer
4. In main loop:
a. Check if 60 seconds have elapsed
b. Capture image from camera
c. Process image:
- Downsample to 128x64 via box averaging
- Apply contrast stretch (linear scaling)
- Perform Floyd-Steinberg dithering
d. Display processed bitmap on OLED
e. Release frame buffer
5. Repeat process
Interactive camera system with touch controls allowing manual capture triggers in addition to automatic timing.
1. Initialize camera and OLED display
2. Set up touch pins (GPIO1 & GPIO2) with threshold detection
3. Configure 60-second auto-capture timer
4. In main loop:
a. Update touch sensor readings
b. Detect touch press events (justPressed)
c. Check for capture trigger:
- Touch press OR 60-second timer elapsed
d. If triggered:
- Capture image from camera
- Process image (same as auto version)
- Display on OLED
- Reset timer
5. Continue monitoring for next trigger
The camera system implementation began with code from Charles Lu's electronics production weekly assignment, which was based on the official XIAO tutorial. Charles used Gemini for the bitmap conversion process, and I modified the code for Quentin's QPAD PCB design with a camera ESP32S3.
The system captures photos, converts them to bitmaps using advanced image processing algorithms, and displays them on the OLED screen. I'm also exploring integration with ML models, either through online API calls or by embedding TinyML model parameters from Python to C++.
Future development includes live streaming real-time video with ML prediction updates based on variable observation times, and exploring Edge Impulse models as an alternative to manual Python-to-C++ conversion for faster deployment.
Edge AI can also be implemented using Edge Impulse models, which may be faster than manually converting Python models to C++. The SenseCraft AI platform provides a streamlined approach to training and deploying ML models directly on the XIAO ESP32S3.
Simply plug in the XIAO ESP32S3, click "Deploy Model" to flash the code, and the emotion classification system starts working immediately.
Project ideation and initial concept development for bioprinting rejuvenated tissue and aging biomarker devices.
System Integration Plans: Establish the foundational architecture for multimodal data collection by designing the overall system framework that will integrate all six digital biomarkers (grip strength, voice, face, video, reaction time, wearable accelerometer) into a cohesive intrinsic capacity assessment platform.
[Week 0 progress and contributions to final project placeholder]
Version control, laser cutting, and vinyl cutting techniques applied to final project components.
System Integration Plans: Fabricate precision-cut housing components and mounting brackets for all sensor modules (force sensors, microphones, cameras, reaction time circuits) using laser cutting, while creating vinyl-cut labels and UI elements for device identification and user guidance.
[Week 1 progress and contributions to final project placeholder]
Electronics basics and embedded programming for the aging biomarker device components.
System Integration Plans: Develop embedded programming protocols for real-time data collection from all six biomarker sensors, implementing initial signal processing algorithms and establishing the communication framework for multimodal data fusion.
[Week 2 progress and contributions to final project placeholder]
3D scanning and printing techniques for bioprinting components and device housings.
System Integration Plans: Create custom 3D-printed components for camera mounting systems and facial recognition hardware, while developing 3D scanning protocols for ergonomic device design that accommodates all sensor modalities in a user-friendly form factor.
[Week 3 progress and contributions to final project placeholder]
EDA and schematic design for the aging biomarker device electronics.
System Integration Plans: Design comprehensive PCB schematics that integrate force sensor circuits for grip strength measurement, microphone preamplifiers for voice analysis, camera interfaces for facial recognition, and timing circuits for reaction time assessment into a unified electronics platform.
[Week 4 progress and contributions to final project placeholder]
PCB fabrication, debugging, and assembly for the biomarker device.
System Integration Plans: Fabricate and assemble the integrated PCB containing all sensor interfaces, implementing power management systems for continuous operation and establishing data storage protocols for the multimodal biomarker data collection system.
[Week 5 progress and contributions to final project placeholder]
CAM and milling for precision components and device housings.
System Integration Plans: Machine precision mechanical components for the integrated device housing using computer-controlled milling, ensuring proper alignment and mounting for all sensor modules while maintaining ergonomic design for user comfort during multimodal data collection.
[Week 6 progress and contributions to final project placeholder]
Sensors and embedded architectures for data collection in the biomarker device.
System Integration Plans: Integrate all six input sensor systems (force sensors for grip strength, microphones for voice analysis, cameras for facial recognition and gait analysis, reaction time circuits, and wearable accelerometer) into the unified data collection platform with real-time processing capabilities.
[Week 7 progress and contributions to final project placeholder]
Actuators and system integration for the biomarker device outputs.
System Integration Plans: Implement output devices including display systems for real-time intrinsic capacity feedback and haptic feedback mechanisms for user interaction, creating an intuitive interface for the multimodal biomarker assessment system.
[Week 8 progress and contributions to final project placeholder]
Forming and resin techniques for bioprinting molds and device components.
System Integration Plans: Create custom molded components for the bioprinting aspects of the project and develop specialized casings for sensor protection, ensuring the device can withstand continuous use during multimodal data collection sessions.
[Week 9 progress and contributions to final project placeholder]
Kits and mechanical design for the bioprinting and biomarker device systems.
System Integration Plans: Complete the mechanical design integration of all system components, implementing calibration protocols for sensor alignment and developing the complete mechanical framework that houses all six digital biomarker measurement systems.
[Week 10 progress and contributions to final project placeholder]
BLE, Wi-Fi, and communication protocols for the biomarker device connectivity.
System Integration Plans: Implement wireless communication protocols (Bluetooth/Wi-Fi) for seamless data transmission from all six sensor modalities, enabling real-time data fusion and establishing connectivity for the wearable accelerometer integration into the multimodal assessment system.
[Week 11 progress and contributions to final project placeholder]
UI and application development for the biomarker device interface.
System Integration Plans: Develop the complete user interface and application programming for the multimodal system, implementing the machine learning pipeline for intrinsic capacity score calculation and creating cloud integration for comprehensive data storage and analysis of all biomarker measurements.
[Week 12 progress and contributions to final project placeholder]
Final orders and wildcard week activities for project completion.
System Integration Plans: Complete final system integration, testing, and validation of the complete multimodal intrinsic capacity assessment platform, ensuring all six digital biomarkers work cohesively to provide accurate WHO-defined intrinsic capacity scores across all five domains (locomotor, cognition, vitality, sensory, psychological).
[Week 13 progress and contributions to final project placeholder]
[Links to CAD/SVG/assets placeholder.]
[Reflection notes placeholder.]
Acknowledgements and contributions that made this project possible.
Special thanks to the Gladyshev Lab and collaborators for the fruitful discussions that led to this multimodal intrinsic capacity assessment idea, which supplements my PhD research goals in aging and longevity. The conceptual framework for integrating multiple digital biomarkers to assess intrinsic capacity domains emerged from collaborative research discussions on aging biomarkers and healthspan assessment.
Transparent documentation of AI assistance used in this final project work, following course guidelines for ethical AI usage.
Cursor AI assisted with creating the final project page structure, setting up navigation links, and organizing weekly progress sections. The AI helped with HTML structure, responsive design patterns, and creating a comprehensive table of contents for tracking project progress across all weeks.
ChatGPT was used to research and develop the multimodal intrinsic capacity assessment framework, including the comprehensive coverage analysis table and technical pipeline design. The AI assisted with structuring the WHO-defined intrinsic capacity domains, identifying appropriate digital biomarkers, and designing the fusion architecture for multimodal data processing.
Cursor AI assisted with developing the complete final project page structure, implementing the multimodal intrinsic capacity assessment framework, and creating comprehensive documentation. The AI helped with HTML structure, responsive design, weekly system integration plans, and organizing the technical documentation for the complete biomarker assessment system.
Cursor AI assisted with finalizing the project presentation structure to ensure full compliance with MIT Academy project presentation requirements. The AI helped implement all required sections including answering questions, design documentation, bill of materials, individual mastery requirements, course presentation structure, and spiral model development approach visualization.
ChatGPT was used to discuss camera system implementation strategies, image processing algorithms, and EdgeAI integration approaches. The AI assisted with understanding Floyd-Steinberg dithering implementation, touch sensor integration, and exploring TinyML deployment options for the XIAO ESP32S3 platform.
This work is licensed under a
Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License
Bioprinting & Aging Biomarker Device · HTMAA 2025
XIAO ESP32S3 camera successfully captures and displays images on OLED screen using Floyd-Steinberg dithering
Conceptual visualization of the multimodal intrinsic capacity assessment system integrating grip strength, voice analysis, facial recognition, video motion capture, reaction time measurement, and wearable accelerometer data.
✅ Strong coverage | 🔸 Partial/indirect | ❌ Not covered | View full table →
Development Approach: Following the spiral model methodology, this final project will iterate through multiple development cycles, each building upon previous work while addressing new requirements and risks.
[Project introduction placeholder - describing the bioprinting and aging biomarker device project.]
Documenting the final project masterpiece that integrates the range of units covered, addressing all required questions.
[Placeholder: Describe the multimodal intrinsic capacity assessment system functionality]
[Placeholder: Literature review and prior work in aging biomarkers and intrinsic capacity assessment]
[Placeholder: Research papers, open-source projects, and reference materials]
[Placeholder: Custom hardware, software, and integration systems designed]
[Placeholder: Complete list of materials, sensors, microcontrollers, and components]
[Placeholder: Suppliers, vendors, and sourcing information]
[Placeholder: Detailed cost breakdown and budget analysis]
[Placeholder: Custom fabricated parts, 3D printed components, and assembled systems]
[Placeholder: Laser cutting, 3D printing, PCB fabrication, programming, and testing processes]
[Placeholder: Research questions addressed and hypotheses tested]
[Placeholder: Successes, challenges, and lessons learned]
[Placeholder: Testing protocols, validation methods, and evaluation criteria]
[Placeholder: Impact on aging research, clinical applications, and future development]
Your project should incorporate 2D and 3D design, multiple additive and subtractive fabrication processes, electronics design and production, embedded microcontroller design, interfacing, and programming, system integration and packaging.
2D design work for the multimodal intrinsic capacity assessment system:
Tools Used: Inkscape, KiCad, Adobe Illustrator, Figma
3D design work for device components and integration:
Tools Used: Fusion 360, FreeCAD, OpenSCAD, PrusaSlicer
Where possible, you should make rather than buy the parts of your project. Complete breakdown of materials, components, and sourcing information.
Complete list of materials and components:
[Placeholder: Detailed cost analysis and budget tracking]
Strategic decisions on fabrication vs. purchasing:
[Fabrication process documentation placeholder.]
[Testing and validation documentation placeholder.]
Projects can be separate or joint, but need to show individual mastery of the skills, and be independently operable.
Demonstration of individual skills across all course units:
Project operates independently without external dependencies:
Present your final project, weekly and group assignments, and documentation.
Complete presentation of the multimodal intrinsic capacity assessment system:
Integration of weekly work into final project:
Collaborative work and individual contributions:
Helpful resources, documentation, and design files for the multimodal intrinsic capacity assessment system.
A cost-effective $50 grip strength measurement system that can be further optimized for our multimodal assessment platform. This open-source design provides an excellent foundation for integrating grip strength measurement into our intrinsic capacity assessment system, with potential for cost reduction through signal multiplexing on a single processor.
Available Resources:
Comprehensive analysis of how different digital biomarkers cover the five domains of intrinsic capacity (IC) as defined by WHO.
Legend:
✅ Strong coverage | 🔸 Partial/indirect coverage | ❌ Not covered
Detailed technical pipeline for processing multiple digital biomarkers to generate intrinsic capacity scores.
Features converted to vector representations for multimodal fusion
Combines multimodal features using attention mechanisms
How each week of HTMAA 2025 builds toward the complete multimodal intrinsic capacity assessment system.
Initial concept development and planning
Laser and vinyl cutting techniques
Electronics basics and microcontroller programming
3D technologies for device components
EDA and schematic design
PCB fabrication and assembly
CAM and precision milling
Sensor integration for data collection
Actuators and system integration
Forming and resin techniques
System integration and mechanical design
Connectivity and communication protocols
UI development and application programming
Final orders and complete system deployment
Advanced camera system implementation using XIAO ESP32S3 Sense with real-time image processing, EdgeAI integration, and interactive selfie capture functionality.
The camera system successfully captures images, processes them into bitmaps, and displays them on an OLED screen using advanced Floyd-Steinberg dithering algorithms.
Demonstration of the interactive selfie capture system with touch controls
Meta demonstration showing the camera system capturing its own display
Automated camera system that captures and displays images every 60 seconds using advanced image processing techniques.
1. Initialize camera with PSRAM frame buffers
2. Configure OLED display (128x64 pixels)
3. Set up 60-second capture interval timer
4. In main loop:
a. Check if 60 seconds have elapsed
b. Capture image from camera
c. Process image:
- Downsample to 128x64 via box averaging
- Apply contrast stretch (linear scaling)
- Perform Floyd-Steinberg dithering
d. Display processed bitmap on OLED
e. Release frame buffer
5. Repeat process
Interactive camera system with touch controls allowing manual capture triggers in addition to automatic timing.
1. Initialize camera and OLED display
2. Set up touch pins (GPIO1 & GPIO2) with threshold detection
3. Configure 60-second auto-capture timer
4. In main loop:
a. Update touch sensor readings
b. Detect touch press events (justPressed)
c. Check for capture trigger:
- Touch press OR 60-second timer elapsed
d. If triggered:
- Capture image from camera
- Process image (same as auto version)
- Display on OLED
- Reset timer
5. Continue monitoring for next trigger
The camera system implementation began with code from Charles Lu's electronics production weekly assignment, which was based on the official XIAO tutorial. Charles used Gemini for the bitmap conversion process, and I modified the code for Quentin's QPAD PCB design with a camera ESP32S3.
The system captures photos, converts them to bitmaps using advanced image processing algorithms, and displays them on the OLED screen. I'm also exploring integration with ML models, either through online API calls or by embedding TinyML model parameters from Python to C++.
Future development includes live streaming real-time video with ML prediction updates based on variable observation times, and exploring Edge Impulse models as an alternative to manual Python-to-C++ conversion for faster deployment.
Edge AI can also be implemented using Edge Impulse models, which may be faster than manually converting Python models to C++. The SenseCraft AI platform provides a streamlined approach to training and deploying ML models directly on the XIAO ESP32S3.
Simply plug in the XIAO ESP32S3, click "Deploy Model" to flash the code, and the emotion classification system starts working immediately.
Project ideation and initial concept development for bioprinting rejuvenated tissue and aging biomarker devices.
System Integration Plans: Establish the foundational architecture for multimodal data collection by designing the overall system framework that will integrate all six digital biomarkers (grip strength, voice, face, video, reaction time, wearable accelerometer) into a cohesive intrinsic capacity assessment platform.
[Week 0 progress and contributions to final project placeholder]
Version control, laser cutting, and vinyl cutting techniques applied to final project components.
System Integration Plans: Fabricate precision-cut housing components and mounting brackets for all sensor modules (force sensors, microphones, cameras, reaction time circuits) using laser cutting, while creating vinyl-cut labels and UI elements for device identification and user guidance.
[Week 1 progress and contributions to final project placeholder]
Electronics basics and embedded programming for the aging biomarker device components.
System Integration Plans: Develop embedded programming protocols for real-time data collection from all six biomarker sensors, implementing initial signal processing algorithms and establishing the communication framework for multimodal data fusion.
[Week 2 progress and contributions to final project placeholder]
3D scanning and printing techniques for bioprinting components and device housings.
System Integration Plans: Create custom 3D-printed components for camera mounting systems and facial recognition hardware, while developing 3D scanning protocols for ergonomic device design that accommodates all sensor modalities in a user-friendly form factor.
[Week 3 progress and contributions to final project placeholder]
EDA and schematic design for the aging biomarker device electronics.
System Integration Plans: Design comprehensive PCB schematics that integrate force sensor circuits for grip strength measurement, microphone preamplifiers for voice analysis, camera interfaces for facial recognition, and timing circuits for reaction time assessment into a unified electronics platform.
[Week 4 progress and contributions to final project placeholder]
PCB fabrication, debugging, and assembly for the biomarker device.
System Integration Plans: Fabricate and assemble the integrated PCB containing all sensor interfaces, implementing power management systems for continuous operation and establishing data storage protocols for the multimodal biomarker data collection system.
[Week 5 progress and contributions to final project placeholder]
CAM and milling for precision components and device housings.
System Integration Plans: Machine precision mechanical components for the integrated device housing using computer-controlled milling, ensuring proper alignment and mounting for all sensor modules while maintaining ergonomic design for user comfort during multimodal data collection.
[Week 6 progress and contributions to final project placeholder]
Sensors and embedded architectures for data collection in the biomarker device.
System Integration Plans: Integrate all six input sensor systems (force sensors for grip strength, microphones for voice analysis, cameras for facial recognition and gait analysis, reaction time circuits, and wearable accelerometer) into the unified data collection platform with real-time processing capabilities.
[Week 7 progress and contributions to final project placeholder]
Actuators and system integration for the biomarker device outputs.
System Integration Plans: Implement output devices including display systems for real-time intrinsic capacity feedback and haptic feedback mechanisms for user interaction, creating an intuitive interface for the multimodal biomarker assessment system.
[Week 8 progress and contributions to final project placeholder]
Forming and resin techniques for bioprinting molds and device components.
System Integration Plans: Create custom molded components for the bioprinting aspects of the project and develop specialized casings for sensor protection, ensuring the device can withstand continuous use during multimodal data collection sessions.
[Week 9 progress and contributions to final project placeholder]
Kits and mechanical design for the bioprinting and biomarker device systems.
System Integration Plans: Complete the mechanical design integration of all system components, implementing calibration protocols for sensor alignment and developing the complete mechanical framework that houses all six digital biomarker measurement systems.
[Week 10 progress and contributions to final project placeholder]
BLE, Wi-Fi, and communication protocols for the biomarker device connectivity.
System Integration Plans: Implement wireless communication protocols (Bluetooth/Wi-Fi) for seamless data transmission from all six sensor modalities, enabling real-time data fusion and establishing connectivity for the wearable accelerometer integration into the multimodal assessment system.
[Week 11 progress and contributions to final project placeholder]
UI and application development for the biomarker device interface.
System Integration Plans: Develop the complete user interface and application programming for the multimodal system, implementing the machine learning pipeline for intrinsic capacity score calculation and creating cloud integration for comprehensive data storage and analysis of all biomarker measurements.
[Week 12 progress and contributions to final project placeholder]
Final orders and wildcard week activities for project completion.
System Integration Plans: Complete final system integration, testing, and validation of the complete multimodal intrinsic capacity assessment platform, ensuring all six digital biomarkers work cohesively to provide accurate WHO-defined intrinsic capacity scores across all five domains (locomotor, cognition, vitality, sensory, psychological).
[Week 13 progress and contributions to final project placeholder]
[Links to CAD/SVG/assets placeholder.]
[Reflection notes placeholder.]
Acknowledgements and contributions that made this project possible.
Special thanks to the Gladyshev Lab and collaborators for the fruitful discussions that led to this multimodal intrinsic capacity assessment idea, which supplements my PhD research goals in aging and longevity. The conceptual framework for integrating multiple digital biomarkers to assess intrinsic capacity domains emerged from collaborative research discussions on aging biomarkers and healthspan assessment.
Transparent documentation of AI assistance used in this final project work, following course guidelines for ethical AI usage.
Cursor AI assisted with creating the final project page structure, setting up navigation links, and organizing weekly progress sections. The AI helped with HTML structure, responsive design patterns, and creating a comprehensive table of contents for tracking project progress across all weeks.
ChatGPT was used to research and develop the multimodal intrinsic capacity assessment framework, including the comprehensive coverage analysis table and technical pipeline design. The AI assisted with structuring the WHO-defined intrinsic capacity domains, identifying appropriate digital biomarkers, and designing the fusion architecture for multimodal data processing.
Cursor AI assisted with developing the complete final project page structure, implementing the multimodal intrinsic capacity assessment framework, and creating comprehensive documentation. The AI helped with HTML structure, responsive design, weekly system integration plans, and organizing the technical documentation for the complete biomarker assessment system.
Cursor AI assisted with finalizing the project presentation structure to ensure full compliance with MIT Academy project presentation requirements. The AI helped implement all required sections including answering questions, design documentation, bill of materials, individual mastery requirements, course presentation structure, and spiral model development approach visualization.
ChatGPT was used to discuss camera system implementation strategies, image processing algorithms, and EdgeAI integration approaches. The AI assisted with understanding Floyd-Steinberg dithering implementation, touch sensor integration, and exploring TinyML deployment options for the XIAO ESP32S3 platform.
This work is licensed under a
Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License
Bioprinting & Aging Biomarker Device · HTMAA 2025
XIAO ESP32S3 camera successfully captures and displays images on OLED screen using Floyd-Steinberg dithering
Conceptual visualization of the multimodal intrinsic capacity assessment system integrating grip strength, voice analysis, facial recognition, video motion capture, reaction time measurement, and wearable accelerometer data.
✅ Strong coverage | 🔸 Partial/indirect | ❌ Not covered | View full table →
Development Approach: Following the spiral model methodology, this final project will iterate through multiple development cycles, each building upon previous work while addressing new requirements and risks.
[Project introduction placeholder - describing the bioprinting and aging biomarker device project.]
Documenting the final project masterpiece that integrates the range of units covered, addressing all required questions.
[Placeholder: Describe the multimodal intrinsic capacity assessment system functionality]
[Placeholder: Literature review and prior work in aging biomarkers and intrinsic capacity assessment]
[Placeholder: Research papers, open-source projects, and reference materials]
[Placeholder: Custom hardware, software, and integration systems designed]
[Placeholder: Complete list of materials, sensors, microcontrollers, and components]
[Placeholder: Suppliers, vendors, and sourcing information]
[Placeholder: Detailed cost breakdown and budget analysis]
[Placeholder: Custom fabricated parts, 3D printed components, and assembled systems]
[Placeholder: Laser cutting, 3D printing, PCB fabrication, programming, and testing processes]
[Placeholder: Research questions addressed and hypotheses tested]
[Placeholder: Successes, challenges, and lessons learned]
[Placeholder: Testing protocols, validation methods, and evaluation criteria]
[Placeholder: Impact on aging research, clinical applications, and future development]
Your project should incorporate 2D and 3D design, multiple additive and subtractive fabrication processes, electronics design and production, embedded microcontroller design, interfacing, and programming, system integration and packaging.
2D design work for the multimodal intrinsic capacity assessment system:
Tools Used: Inkscape, KiCad, Adobe Illustrator, Figma
3D design work for device components and integration:
Tools Used: Fusion 360, FreeCAD, OpenSCAD, PrusaSlicer
Where possible, you should make rather than buy the parts of your project. Complete breakdown of materials, components, and sourcing information.
Complete list of materials and components:
[Placeholder: Detailed cost analysis and budget tracking]
Strategic decisions on fabrication vs. purchasing:
[Fabrication process documentation placeholder.]
[Testing and validation documentation placeholder.]
Projects can be separate or joint, but need to show individual mastery of the skills, and be independently operable.
Demonstration of individual skills across all course units:
Project operates independently without external dependencies:
Present your final project, weekly and group assignments, and documentation.
Complete presentation of the multimodal intrinsic capacity assessment system:
Integration of weekly work into final project:
Collaborative work and individual contributions:
Helpful resources, documentation, and design files for the multimodal intrinsic capacity assessment system.
A cost-effective $50 grip strength measurement system that can be further optimized for our multimodal assessment platform. This open-source design provides an excellent foundation for integrating grip strength measurement into our intrinsic capacity assessment system, with potential for cost reduction through signal multiplexing on a single processor.
Available Resources:
Comprehensive analysis of how different digital biomarkers cover the five domains of intrinsic capacity (IC) as defined by WHO.
Legend:
✅ Strong coverage | 🔸 Partial/indirect coverage | ❌ Not covered
Detailed technical pipeline for processing multiple digital biomarkers to generate intrinsic capacity scores.
Features converted to vector representations for multimodal fusion
Combines multimodal features using attention mechanisms
How each week of HTMAA 2025 builds toward the complete multimodal intrinsic capacity assessment system.
Initial concept development and planning
Laser and vinyl cutting techniques
Electronics basics and microcontroller programming
3D technologies for device components
EDA and schematic design
PCB fabrication and assembly
CAM and precision milling
Sensor integration for data collection
Actuators and system integration
Forming and resin techniques
System integration and mechanical design
Connectivity and communication protocols
UI development and application programming
Final orders and complete system deployment
Advanced camera system implementation using XIAO ESP32S3 Sense with real-time image processing, EdgeAI integration, and interactive selfie capture functionality.
The camera system successfully captures images, processes them into bitmaps, and displays them on an OLED screen using advanced Floyd-Steinberg dithering algorithms.
Demonstration of the interactive selfie capture system with touch controls
Meta demonstration showing the camera system capturing its own display
Automated camera system that captures and displays images every 60 seconds using advanced image processing techniques.
1. Initialize camera with PSRAM frame buffers
2. Configure OLED display (128x64 pixels)
3. Set up 60-second capture interval timer
4. In main loop:
a. Check if 60 seconds have elapsed
b. Capture image from camera
c. Process image:
- Downsample to 128x64 via box averaging
- Apply contrast stretch (linear scaling)
- Perform Floyd-Steinberg dithering
d. Display processed bitmap on OLED
e. Release frame buffer
5. Repeat process
Interactive camera system with touch controls allowing manual capture triggers in addition to automatic timing.
1. Initialize camera and OLED display
2. Set up touch pins (GPIO1 & GPIO2) with threshold detection
3. Configure 60-second auto-capture timer
4. In main loop:
a. Update touch sensor readings
b. Detect touch press events (justPressed)
c. Check for capture trigger:
- Touch press OR 60-second timer elapsed
d. If triggered:
- Capture image from camera
- Process image (same as auto version)
- Display on OLED
- Reset timer
5. Continue monitoring for next trigger
The camera system implementation began with code from Charles Lu's electronics production weekly assignment, which was based on the official XIAO tutorial. Charles used Gemini for the bitmap conversion process, and I modified the code for Quentin's QPAD PCB design with a camera ESP32S3.
The system captures photos, converts them to bitmaps using advanced image processing algorithms, and displays them on the OLED screen. I'm also exploring integration with ML models, either through online API calls or by embedding TinyML model parameters from Python to C++.
Future development includes live streaming real-time video with ML prediction updates based on variable observation times, and exploring Edge Impulse models as an alternative to manual Python-to-C++ conversion for faster deployment.
Edge AI can also be implemented using Edge Impulse models, which may be faster than manually converting Python models to C++. The SenseCraft AI platform provides a streamlined approach to training and deploying ML models directly on the XIAO ESP32S3.
Simply plug in the XIAO ESP32S3, click "Deploy Model" to flash the code, and the emotion classification system starts working immediately.
Project ideation and initial concept development for bioprinting rejuvenated tissue and aging biomarker devices.
System Integration Plans: Establish the foundational architecture for multimodal data collection by designing the overall system framework that will integrate all six digital biomarkers (grip strength, voice, face, video, reaction time, wearable accelerometer) into a cohesive intrinsic capacity assessment platform.
[Week 0 progress and contributions to final project placeholder]
Version control, laser cutting, and vinyl cutting techniques applied to final project components.
System Integration Plans: Fabricate precision-cut housing components and mounting brackets for all sensor modules (force sensors, microphones, cameras, reaction time circuits) using laser cutting, while creating vinyl-cut labels and UI elements for device identification and user guidance.
[Week 1 progress and contributions to final project placeholder]
Electronics basics and embedded programming for the aging biomarker device components.
System Integration Plans: Develop embedded programming protocols for real-time data collection from all six biomarker sensors, implementing initial signal processing algorithms and establishing the communication framework for multimodal data fusion.
[Week 2 progress and contributions to final project placeholder]
3D scanning and printing techniques for bioprinting components and device housings.
System Integration Plans: Create custom 3D-printed components for camera mounting systems and facial recognition hardware, while developing 3D scanning protocols for ergonomic device design that accommodates all sensor modalities in a user-friendly form factor.
[Week 3 progress and contributions to final project placeholder]
EDA and schematic design for the aging biomarker device electronics.
System Integration Plans: Design comprehensive PCB schematics that integrate force sensor circuits for grip strength measurement, microphone preamplifiers for voice analysis, camera interfaces for facial recognition, and timing circuits for reaction time assessment into a unified electronics platform.
[Week 4 progress and contributions to final project placeholder]
PCB fabrication, debugging, and assembly for the biomarker device.
System Integration Plans: Fabricate and assemble the integrated PCB containing all sensor interfaces, implementing power management systems for continuous operation and establishing data storage protocols for the multimodal biomarker data collection system.
[Week 5 progress and contributions to final project placeholder]
CAM and milling for precision components and device housings.
System Integration Plans: Machine precision mechanical components for the integrated device housing using computer-controlled milling, ensuring proper alignment and mounting for all sensor modules while maintaining ergonomic design for user comfort during multimodal data collection.
[Week 6 progress and contributions to final project placeholder]
Sensors and embedded architectures for data collection in the biomarker device.
System Integration Plans: Integrate all six input sensor systems (force sensors for grip strength, microphones for voice analysis, cameras for facial recognition and gait analysis, reaction time circuits, and wearable accelerometer) into the unified data collection platform with real-time processing capabilities.
[Week 7 progress and contributions to final project placeholder]
Actuators and system integration for the biomarker device outputs.
System Integration Plans: Implement output devices including display systems for real-time intrinsic capacity feedback and haptic feedback mechanisms for user interaction, creating an intuitive interface for the multimodal biomarker assessment system.
[Week 8 progress and contributions to final project placeholder]
Forming and resin techniques for bioprinting molds and device components.
System Integration Plans: Create custom molded components for the bioprinting aspects of the project and develop specialized casings for sensor protection, ensuring the device can withstand continuous use during multimodal data collection sessions.
[Week 9 progress and contributions to final project placeholder]
Kits and mechanical design for the bioprinting and biomarker device systems.
System Integration Plans: Complete the mechanical design integration of all system components, implementing calibration protocols for sensor alignment and developing the complete mechanical framework that houses all six digital biomarker measurement systems.
[Week 10 progress and contributions to final project placeholder]
BLE, Wi-Fi, and communication protocols for the biomarker device connectivity.
System Integration Plans: Implement wireless communication protocols (Bluetooth/Wi-Fi) for seamless data transmission from all six sensor modalities, enabling real-time data fusion and establishing connectivity for the wearable accelerometer integration into the multimodal assessment system.
[Week 11 progress and contributions to final project placeholder]
UI and application development for the biomarker device interface.
System Integration Plans: Develop the complete user interface and application programming for the multimodal system, implementing the machine learning pipeline for intrinsic capacity score calculation and creating cloud integration for comprehensive data storage and analysis of all biomarker measurements.
[Week 12 progress and contributions to final project placeholder]
Final orders and wildcard week activities for project completion.
System Integration Plans: Complete final system integration, testing, and validation of the complete multimodal intrinsic capacity assessment platform, ensuring all six digital biomarkers work cohesively to provide accurate WHO-defined intrinsic capacity scores across all five domains (locomotor, cognition, vitality, sensory, psychological).
[Week 13 progress and contributions to final project placeholder]
[Links to CAD/SVG/assets placeholder.]
[Reflection notes placeholder.]
Acknowledgements and contributions that made this project possible.
Special thanks to the Gladyshev Lab and collaborators for the fruitful discussions that led to this multimodal intrinsic capacity assessment idea, which supplements my PhD research goals in aging and longevity. The conceptual framework for integrating multiple digital biomarkers to assess intrinsic capacity domains emerged from collaborative research discussions on aging biomarkers and healthspan assessment.
Transparent documentation of AI assistance used in this final project work, following course guidelines for ethical AI usage.
Cursor AI assisted with creating the final project page structure, setting up navigation links, and organizing weekly progress sections. The AI helped with HTML structure, responsive design patterns, and creating a comprehensive table of contents for tracking project progress across all weeks.
ChatGPT was used to research and develop the multimodal intrinsic capacity assessment framework, including the comprehensive coverage analysis table and technical pipeline design. The AI assisted with structuring the WHO-defined intrinsic capacity domains, identifying appropriate digital biomarkers, and designing the fusion architecture for multimodal data processing.
Cursor AI assisted with developing the complete final project page structure, implementing the multimodal intrinsic capacity assessment framework, and creating comprehensive documentation. The AI helped with HTML structure, responsive design, weekly system integration plans, and organizing the technical documentation for the complete biomarker assessment system.
Cursor AI assisted with finalizing the project presentation structure to ensure full compliance with MIT Academy project presentation requirements. The AI helped implement all required sections including answering questions, design documentation, bill of materials, individual mastery requirements, course presentation structure, and spiral model development approach visualization.
ChatGPT was used to discuss camera system implementation strategies, image processing algorithms, and EdgeAI integration approaches. The AI assisted with understanding Floyd-Steinberg dithering implementation, touch sensor integration, and exploring TinyML deployment options for the XIAO ESP32S3 platform.
This work is licensed under a
Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License
Bioprinting & Aging Biomarker Device · HTMAA 2025
XIAO ESP32S3 camera successfully captures and displays images on OLED screen using Floyd-Steinberg dithering
Conceptual visualization of the multimodal intrinsic capacity assessment system integrating grip strength, voice analysis, facial recognition, video motion capture, reaction time measurement, and wearable accelerometer data.
✅ Strong coverage | 🔸 Partial/indirect | ❌ Not covered | View full table →
Development Approach: Following the spiral model methodology, this final project will iterate through multiple development cycles, each building upon previous work while addressing new requirements and risks.
[Project introduction placeholder - describing the bioprinting and aging biomarker device project.]
Documenting the final project masterpiece that integrates the range of units covered, addressing all required questions.
[Placeholder: Describe the multimodal intrinsic capacity assessment system functionality]
[Placeholder: Literature review and prior work in aging biomarkers and intrinsic capacity assessment]
[Placeholder: Research papers, open-source projects, and reference materials]
[Placeholder: Custom hardware, software, and integration systems designed]
[Placeholder: Complete list of materials, sensors, microcontrollers, and components]
[Placeholder: Suppliers, vendors, and sourcing information]
[Placeholder: Detailed cost breakdown and budget analysis]
[Placeholder: Custom fabricated parts, 3D printed components, and assembled systems]
[Placeholder: Laser cutting, 3D printing, PCB fabrication, programming, and testing processes]
[Placeholder: Research questions addressed and hypotheses tested]
[Placeholder: Successes, challenges, and lessons learned]
[Placeholder: Testing protocols, validation methods, and evaluation criteria]
[Placeholder: Impact on aging research, clinical applications, and future development]
Your project should incorporate 2D and 3D design, multiple additive and subtractive fabrication processes, electronics design and production, embedded microcontroller design, interfacing, and programming, system integration and packaging.
2D design work for the multimodal intrinsic capacity assessment system:
Tools Used: Inkscape, KiCad, Adobe Illustrator, Figma
3D design work for device components and integration:
Tools Used: Fusion 360, FreeCAD, OpenSCAD, PrusaSlicer
Where possible, you should make rather than buy the parts of your project. Complete breakdown of materials, components, and sourcing information.
Complete list of materials and components:
[Placeholder: Detailed cost analysis and budget tracking]
Strategic decisions on fabrication vs. purchasing:
[Fabrication process documentation placeholder.]
[Testing and validation documentation placeholder.]
Projects can be separate or joint, but need to show individual mastery of the skills, and be independently operable.
Demonstration of individual skills across all course units:
Project operates independently without external dependencies:
Present your final project, weekly and group assignments, and documentation.
Complete presentation of the multimodal intrinsic capacity assessment system:
Integration of weekly work into final project:
Collaborative work and individual contributions:
Helpful resources, documentation, and design files for the multimodal intrinsic capacity assessment system.
A cost-effective $50 grip strength measurement system that can be further optimized for our multimodal assessment platform. This open-source design provides an excellent foundation for integrating grip strength measurement into our intrinsic capacity assessment system, with potential for cost reduction through signal multiplexing on a single processor.
Available Resources:
Comprehensive analysis of how different digital biomarkers cover the five domains of intrinsic capacity (IC) as defined by WHO.
Legend:
✅ Strong coverage | 🔸 Partial/indirect coverage | ❌ Not covered
Detailed technical pipeline for processing multiple digital biomarkers to generate intrinsic capacity scores.
Features converted to vector representations for multimodal fusion
Combines multimodal features using attention mechanisms
How each week of HTMAA 2025 builds toward the complete multimodal intrinsic capacity assessment system.
Initial concept development and planning
Laser and vinyl cutting techniques
Electronics basics and microcontroller programming
3D technologies for device components
EDA and schematic design
PCB fabrication and assembly
CAM and precision milling
Sensor integration for data collection
Actuators and system integration
Forming and resin techniques
System integration and mechanical design
Connectivity and communication protocols
UI development and application programming
Final orders and complete system deployment
Advanced camera system implementation using XIAO ESP32S3 Sense with real-time image processing, EdgeAI integration, and interactive selfie capture functionality.
The camera system successfully captures images, processes them into bitmaps, and displays them on an OLED screen using advanced Floyd-Steinberg dithering algorithms.
Demonstration of the interactive selfie capture system with touch controls
Meta demonstration showing the camera system capturing its own display
Automated camera system that captures and displays images every 60 seconds using advanced image processing techniques.
1. Initialize camera with PSRAM frame buffers
2. Configure OLED display (128x64 pixels)
3. Set up 60-second capture interval timer
4. In main loop:
a. Check if 60 seconds have elapsed
b. Capture image from camera
c. Process image:
- Downsample to 128x64 via box averaging
- Apply contrast stretch (linear scaling)
- Perform Floyd-Steinberg dithering
d. Display processed bitmap on OLED
e. Release frame buffer
5. Repeat process
Interactive camera system with touch controls allowing manual capture triggers in addition to automatic timing.
1. Initialize camera and OLED display
2. Set up touch pins (GPIO1 & GPIO2) with threshold detection
3. Configure 60-second auto-capture timer
4. In main loop:
a. Update touch sensor readings
b. Detect touch press events (justPressed)
c. Check for capture trigger:
- Touch press OR 60-second timer elapsed
d. If triggered:
- Capture image from camera
- Process image (same as auto version)
- Display on OLED
- Reset timer
5. Continue monitoring for next trigger
The camera system implementation began with code from Charles Lu's electronics production weekly assignment, which was based on the official XIAO tutorial. Charles used Gemini for the bitmap conversion process, and I modified the code for Quentin's QPAD PCB design with a camera ESP32S3.
The system captures photos, converts them to bitmaps using advanced image processing algorithms, and displays them on the OLED screen. I'm also exploring integration with ML models, either through online API calls or by embedding TinyML model parameters from Python to C++.
Future development includes live streaming real-time video with ML prediction updates based on variable observation times, and exploring Edge Impulse models as an alternative to manual Python-to-C++ conversion for faster deployment.
Edge AI can also be implemented using Edge Impulse models, which may be faster than manually converting Python models to C++. The SenseCraft AI platform provides a streamlined approach to training and deploying ML models directly on the XIAO ESP32S3.
Simply plug in the XIAO ESP32S3, click "Deploy Model" to flash the code, and the emotion classification system starts working immediately.
Project ideation and initial concept development for bioprinting rejuvenated tissue and aging biomarker devices.
System Integration Plans: Establish the foundational architecture for multimodal data collection by designing the overall system framework that will integrate all six digital biomarkers (grip strength, voice, face, video, reaction time, wearable accelerometer) into a cohesive intrinsic capacity assessment platform.
[Week 0 progress and contributions to final project placeholder]
Version control, laser cutting, and vinyl cutting techniques applied to final project components.
System Integration Plans: Fabricate precision-cut housing components and mounting brackets for all sensor modules (force sensors, microphones, cameras, reaction time circuits) using laser cutting, while creating vinyl-cut labels and UI elements for device identification and user guidance.
[Week 1 progress and contributions to final project placeholder]
Electronics basics and embedded programming for the aging biomarker device components.
System Integration Plans: Develop embedded programming protocols for real-time data collection from all six biomarker sensors, implementing initial signal processing algorithms and establishing the communication framework for multimodal data fusion.
[Week 2 progress and contributions to final project placeholder]
3D scanning and printing techniques for bioprinting components and device housings.
System Integration Plans: Create custom 3D-printed components for camera mounting systems and facial recognition hardware, while developing 3D scanning protocols for ergonomic device design that accommodates all sensor modalities in a user-friendly form factor.
[Week 3 progress and contributions to final project placeholder]
EDA and schematic design for the aging biomarker device electronics.
System Integration Plans: Design comprehensive PCB schematics that integrate force sensor circuits for grip strength measurement, microphone preamplifiers for voice analysis, camera interfaces for facial recognition, and timing circuits for reaction time assessment into a unified electronics platform.
[Week 4 progress and contributions to final project placeholder]
PCB fabrication, debugging, and assembly for the biomarker device.
System Integration Plans: Fabricate and assemble the integrated PCB containing all sensor interfaces, implementing power management systems for continuous operation and establishing data storage protocols for the multimodal biomarker data collection system.
[Week 5 progress and contributions to final project placeholder]
CAM and milling for precision components and device housings.
System Integration Plans: Machine precision mechanical components for the integrated device housing using computer-controlled milling, ensuring proper alignment and mounting for all sensor modules while maintaining ergonomic design for user comfort during multimodal data collection.
[Week 6 progress and contributions to final project placeholder]
Sensors and embedded architectures for data collection in the biomarker device.
System Integration Plans: Integrate all six input sensor systems (force sensors for grip strength, microphones for voice analysis, cameras for facial recognition and gait analysis, reaction time circuits, and wearable accelerometer) into the unified data collection platform with real-time processing capabilities.
[Week 7 progress and contributions to final project placeholder]
Actuators and system integration for the biomarker device outputs.
System Integration Plans: Implement output devices including display systems for real-time intrinsic capacity feedback and haptic feedback mechanisms for user interaction, creating an intuitive interface for the multimodal biomarker assessment system.
[Week 8 progress and contributions to final project placeholder]
Forming and resin techniques for bioprinting molds and device components.
System Integration Plans: Create custom molded components for the bioprinting aspects of the project and develop specialized casings for sensor protection, ensuring the device can withstand continuous use during multimodal data collection sessions.
[Week 9 progress and contributions to final project placeholder]
Kits and mechanical design for the bioprinting and biomarker device systems.
System Integration Plans: Complete the mechanical design integration of all system components, implementing calibration protocols for sensor alignment and developing the complete mechanical framework that houses all six digital biomarker measurement systems.
[Week 10 progress and contributions to final project placeholder]
BLE, Wi-Fi, and communication protocols for the biomarker device connectivity.
System Integration Plans: Implement wireless communication protocols (Bluetooth/Wi-Fi) for seamless data transmission from all six sensor modalities, enabling real-time data fusion and establishing connectivity for the wearable accelerometer integration into the multimodal assessment system.
[Week 11 progress and contributions to final project placeholder]
UI and application development for the biomarker device interface.
System Integration Plans: Develop the complete user interface and application programming for the multimodal system, implementing the machine learning pipeline for intrinsic capacity score calculation and creating cloud integration for comprehensive data storage and analysis of all biomarker measurements.
[Week 12 progress and contributions to final project placeholder]
Final orders and wildcard week activities for project completion.
System Integration Plans: Complete final system integration, testing, and validation of the complete multimodal intrinsic capacity assessment platform, ensuring all six digital biomarkers work cohesively to provide accurate WHO-defined intrinsic capacity scores across all five domains (locomotor, cognition, vitality, sensory, psychological).
[Week 13 progress and contributions to final project placeholder]
[Links to CAD/SVG/assets placeholder.]
[Reflection notes placeholder.]
Acknowledgements and contributions that made this project possible.
Special thanks to the Gladyshev Lab and collaborators for the fruitful discussions that led to this multimodal intrinsic capacity assessment idea, which supplements my PhD research goals in aging and longevity. The conceptual framework for integrating multiple digital biomarkers to assess intrinsic capacity domains emerged from collaborative research discussions on aging biomarkers and healthspan assessment.
Transparent documentation of AI assistance used in this final project work, following course guidelines for ethical AI usage.
Cursor AI assisted with creating the final project page structure, setting up navigation links, and organizing weekly progress sections. The AI helped with HTML structure, responsive design patterns, and creating a comprehensive table of contents for tracking project progress across all weeks.
ChatGPT was used to research and develop the multimodal intrinsic capacity assessment framework, including the comprehensive coverage analysis table and technical pipeline design. The AI assisted with structuring the WHO-defined intrinsic capacity domains, identifying appropriate digital biomarkers, and designing the fusion architecture for multimodal data processing.
Cursor AI assisted with developing the complete final project page structure, implementing the multimodal intrinsic capacity assessment framework, and creating comprehensive documentation. The AI helped with HTML structure, responsive design, weekly system integration plans, and organizing the technical documentation for the complete biomarker assessment system.
Cursor AI assisted with finalizing the project presentation structure to ensure full compliance with MIT Academy project presentation requirements. The AI helped implement all required sections including answering questions, design documentation, bill of materials, individual mastery requirements, course presentation structure, and spiral model development approach visualization.
ChatGPT was used to discuss camera system implementation strategies, image processing algorithms, and EdgeAI integration approaches. The AI assisted with understanding Floyd-Steinberg dithering implementation, touch sensor integration, and exploring TinyML deployment options for the XIAO ESP32S3 platform.
This work is licensed under a
Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License
Bioprinting & Aging Biomarker Device · HTMAA 2025
XIAO ESP32S3 camera successfully captures and displays images on OLED screen using Floyd-Steinberg dithering
Conceptual visualization of the multimodal intrinsic capacity assessment system integrating grip strength, voice analysis, facial recognition, video motion capture, reaction time measurement, and wearable accelerometer data.
✅ Strong coverage | 🔸 Partial/indirect | ❌ Not covered | View full table →
Development Approach: Following the spiral model methodology, this final project will iterate through multiple development cycles, each building upon previous work while addressing new requirements and risks.
[Project introduction placeholder - describing the bioprinting and aging biomarker device project.]
Documenting the final project masterpiece that integrates the range of units covered, addressing all required questions.
[Placeholder: Describe the multimodal intrinsic capacity assessment system functionality]
[Placeholder: Literature review and prior work in aging biomarkers and intrinsic capacity assessment]
[Placeholder: Research papers, open-source projects, and reference materials]
[Placeholder: Custom hardware, software, and integration systems designed]
[Placeholder: Complete list of materials, sensors, microcontrollers, and components]
[Placeholder: Suppliers, vendors, and sourcing information]
[Placeholder: Detailed cost breakdown and budget analysis]
[Placeholder: Custom fabricated parts, 3D printed components, and assembled systems]
[Placeholder: Laser cutting, 3D printing, PCB fabrication, programming, and testing processes]
[Placeholder: Research questions addressed and hypotheses tested]
[Placeholder: Successes, challenges, and lessons learned]
[Placeholder: Testing protocols, validation methods, and evaluation criteria]
[Placeholder: Impact on aging research, clinical applications, and future development]
Your project should incorporate 2D and 3D design, multiple additive and subtractive fabrication processes, electronics design and production, embedded microcontroller design, interfacing, and programming, system integration and packaging.
2D design work for the multimodal intrinsic capacity assessment system:
Tools Used: Inkscape, KiCad, Adobe Illustrator, Figma
3D design work for device components and integration:
Tools Used: Fusion 360, FreeCAD, OpenSCAD, PrusaSlicer
Where possible, you should make rather than buy the parts of your project. Complete breakdown of materials, components, and sourcing information.
Complete list of materials and components:
[Placeholder: Detailed cost analysis and budget tracking]
Strategic decisions on fabrication vs. purchasing:
[Fabrication process documentation placeholder.]
[Testing and validation documentation placeholder.]
Projects can be separate or joint, but need to show individual mastery of the skills, and be independently operable.
Demonstration of individual skills across all course units:
Project operates independently without external dependencies:
Present your final project, weekly and group assignments, and documentation.
Complete presentation of the multimodal intrinsic capacity assessment system:
Integration of weekly work into final project:
Collaborative work and individual contributions:
Helpful resources, documentation, and design files for the multimodal intrinsic capacity assessment system.
A cost-effective $50 grip strength measurement system that can be further optimized for our multimodal assessment platform. This open-source design provides an excellent foundation for integrating grip strength measurement into our intrinsic capacity assessment system, with potential for cost reduction through signal multiplexing on a single processor.
Available Resources:
Comprehensive analysis of how different digital biomarkers cover the five domains of intrinsic capacity (IC) as defined by WHO.
Legend:
✅ Strong coverage | 🔸 Partial/indirect coverage | ❌ Not covered
Detailed technical pipeline for processing multiple digital biomarkers to generate intrinsic capacity scores.
Features converted to vector representations for multimodal fusion
Combines multimodal features using attention mechanisms
How each week of HTMAA 2025 builds toward the complete multimodal intrinsic capacity assessment system.
Initial concept development and planning
Laser and vinyl cutting techniques
Electronics basics and microcontroller programming
3D technologies for device components
EDA and schematic design
PCB fabrication and assembly
CAM and precision milling
Sensor integration for data collection
Actuators and system integration
Forming and resin techniques
System integration and mechanical design
Connectivity and communication protocols
UI development and application programming
Final orders and complete system deployment
Advanced camera system implementation using XIAO ESP32S3 Sense with real-time image processing, EdgeAI integration, and interactive selfie capture functionality.
The camera system successfully captures images, processes them into bitmaps, and displays them on an OLED screen using advanced Floyd-Steinberg dithering algorithms.
Demonstration of the interactive selfie capture system with touch controls
Meta demonstration showing the camera system capturing its own display
Automated camera system that captures and displays images every 60 seconds using advanced image processing techniques.
1. Initialize camera with PSRAM frame buffers
2. Configure OLED display (128x64 pixels)
3. Set up 60-second capture interval timer
4. In main loop:
a. Check if 60 seconds have elapsed
b. Capture image from camera
c. Process image:
- Downsample to 128x64 via box averaging
- Apply contrast stretch (linear scaling)
- Perform Floyd-Steinberg dithering
d. Display processed bitmap on OLED
e. Release frame buffer
5. Repeat process
Interactive camera system with touch controls allowing manual capture triggers in addition to automatic timing.
1. Initialize camera and OLED display
2. Set up touch pins (GPIO1 & GPIO2) with threshold detection
3. Configure 60-second auto-capture timer
4. In main loop:
a. Update touch sensor readings
b. Detect touch press events (justPressed)
c. Check for capture trigger:
- Touch press OR 60-second timer elapsed
d. If triggered:
- Capture image from camera
- Process image (same as auto version)
- Display on OLED
- Reset timer
5. Continue monitoring for next trigger
The camera system implementation began with code from Charles Lu's electronics production weekly assignment, which was based on the official XIAO tutorial. Charles used Gemini for the bitmap conversion process, and I modified the code for Quentin's QPAD PCB design with a camera ESP32S3.
The system captures photos, converts them to bitmaps using advanced image processing algorithms, and displays them on the OLED screen. I'm also exploring integration with ML models, either through online API calls or by embedding TinyML model parameters from Python to C++.
Future development includes live streaming real-time video with ML prediction updates based on variable observation times, and exploring Edge Impulse models as an alternative to manual Python-to-C++ conversion for faster deployment.
Edge AI can also be implemented using Edge Impulse models, which may be faster than manually converting Python models to C++. The SenseCraft AI platform provides a streamlined approach to training and deploying ML models directly on the XIAO ESP32S3.
Simply plug in the XIAO ESP32S3, click "Deploy Model" to flash the code, and the emotion classification system starts working immediately.
Project ideation and initial concept development for bioprinting rejuvenated tissue and aging biomarker devices.
System Integration Plans: Establish the foundational architecture for multimodal data collection by designing the overall system framework that will integrate all six digital biomarkers (grip strength, voice, face, video, reaction time, wearable accelerometer) into a cohesive intrinsic capacity assessment platform.
[Week 0 progress and contributions to final project placeholder]
Version control, laser cutting, and vinyl cutting techniques applied to final project components.
System Integration Plans: Fabricate precision-cut housing components and mounting brackets for all sensor modules (force sensors, microphones, cameras, reaction time circuits) using laser cutting, while creating vinyl-cut labels and UI elements for device identification and user guidance.
[Week 1 progress and contributions to final project placeholder]
Electronics basics and embedded programming for the aging biomarker device components.
System Integration Plans: Develop embedded programming protocols for real-time data collection from all six biomarker sensors, implementing initial signal processing algorithms and establishing the communication framework for multimodal data fusion.
[Week 2 progress and contributions to final project placeholder]
3D scanning and printing techniques for bioprinting components and device housings.
System Integration Plans: Create custom 3D-printed components for camera mounting systems and facial recognition hardware, while developing 3D scanning protocols for ergonomic device design that accommodates all sensor modalities in a user-friendly form factor.
[Week 3 progress and contributions to final project placeholder]
EDA and schematic design for the aging biomarker device electronics.
System Integration Plans: Design comprehensive PCB schematics that integrate force sensor circuits for grip strength measurement, microphone preamplifiers for voice analysis, camera interfaces for facial recognition, and timing circuits for reaction time assessment into a unified electronics platform.
[Week 4 progress and contributions to final project placeholder]
PCB fabrication, debugging, and assembly for the biomarker device.
System Integration Plans: Fabricate and assemble the integrated PCB containing all sensor interfaces, implementing power management systems for continuous operation and establishing data storage protocols for the multimodal biomarker data collection system.
[Week 5 progress and contributions to final project placeholder]
CAM and milling for precision components and device housings.
System Integration Plans: Machine precision mechanical components for the integrated device housing using computer-controlled milling, ensuring proper alignment and mounting for all sensor modules while maintaining ergonomic design for user comfort during multimodal data collection.
[Week 6 progress and contributions to final project placeholder]
Sensors and embedded architectures for data collection in the biomarker device.
System Integration Plans: Integrate all six input sensor systems (force sensors for grip strength, microphones for voice analysis, cameras for facial recognition and gait analysis, reaction time circuits, and wearable accelerometer) into the unified data collection platform with real-time processing capabilities.
[Week 7 progress and contributions to final project placeholder]
Actuators and system integration for the biomarker device outputs.
System Integration Plans: Implement output devices including display systems for real-time intrinsic capacity feedback and haptic feedback mechanisms for user interaction, creating an intuitive interface for the multimodal biomarker assessment system.
[Week 8 progress and contributions to final project placeholder]
Forming and resin techniques for bioprinting molds and device components.
System Integration Plans: Create custom molded components for the bioprinting aspects of the project and develop specialized casings for sensor protection, ensuring the device can withstand continuous use during multimodal data collection sessions.
[Week 9 progress and contributions to final project placeholder]
Kits and mechanical design for the bioprinting and biomarker device systems.
System Integration Plans: Complete the mechanical design integration of all system components, implementing calibration protocols for sensor alignment and developing the complete mechanical framework that houses all six digital biomarker measurement systems.
[Week 10 progress and contributions to final project placeholder]
BLE, Wi-Fi, and communication protocols for the biomarker device connectivity.
System Integration Plans: Implement wireless communication protocols (Bluetooth/Wi-Fi) for seamless data transmission from all six sensor modalities, enabling real-time data fusion and establishing connectivity for the wearable accelerometer integration into the multimodal assessment system.
[Week 11 progress and contributions to final project placeholder]
UI and application development for the biomarker device interface.
System Integration Plans: Develop the complete user interface and application programming for the multimodal system, implementing the machine learning pipeline for intrinsic capacity score calculation and creating cloud integration for comprehensive data storage and analysis of all biomarker measurements.
[Week 12 progress and contributions to final project placeholder]
Final orders and wildcard week activities for project completion.
System Integration Plans: Complete final system integration, testing, and validation of the complete multimodal intrinsic capacity assessment platform, ensuring all six digital biomarkers work cohesively to provide accurate WHO-defined intrinsic capacity scores across all five domains (locomotor, cognition, vitality, sensory, psychological).
[Week 13 progress and contributions to final project placeholder]
[Links to CAD/SVG/assets placeholder.]
[Reflection notes placeholder.]
Acknowledgements and contributions that made this project possible.
Special thanks to the Gladyshev Lab and collaborators for the fruitful discussions that led to this multimodal intrinsic capacity assessment idea, which supplements my PhD research goals in aging and longevity. The conceptual framework for integrating multiple digital biomarkers to assess intrinsic capacity domains emerged from collaborative research discussions on aging biomarkers and healthspan assessment.
Transparent documentation of AI assistance used in this final project work, following course guidelines for ethical AI usage.
Cursor AI assisted with creating the final project page structure, setting up navigation links, and organizing weekly progress sections. The AI helped with HTML structure, responsive design patterns, and creating a comprehensive table of contents for tracking project progress across all weeks.
ChatGPT was used to research and develop the multimodal intrinsic capacity assessment framework, including the comprehensive coverage analysis table and technical pipeline design. The AI assisted with structuring the WHO-defined intrinsic capacity domains, identifying appropriate digital biomarkers, and designing the fusion architecture for multimodal data processing.
Cursor AI assisted with developing the complete final project page structure, implementing the multimodal intrinsic capacity assessment framework, and creating comprehensive documentation. The AI helped with HTML structure, responsive design, weekly system integration plans, and organizing the technical documentation for the complete biomarker assessment system.
Cursor AI assisted with finalizing the project presentation structure to ensure full compliance with MIT Academy project presentation requirements. The AI helped implement all required sections including answering questions, design documentation, bill of materials, individual mastery requirements, course presentation structure, and spiral model development approach visualization.
ChatGPT was used to discuss camera system implementation strategies, image processing algorithms, and EdgeAI integration approaches. The AI assisted with understanding Floyd-Steinberg dithering implementation, touch sensor integration, and exploring TinyML deployment options for the XIAO ESP32S3 platform.
This work is licensed under a
Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License
Bioprinting & Aging Biomarker Device · HTMAA 2025
XIAO ESP32S3 camera successfully captures and displays images on OLED screen using Floyd-Steinberg dithering
Conceptual visualization of the multimodal intrinsic capacity assessment system integrating grip strength, voice analysis, facial recognition, video motion capture, reaction time measurement, and wearable accelerometer data.
✅ Strong coverage | 🔸 Partial/indirect | ❌ Not covered | View full table →
Development Approach: Following the spiral model methodology, this final project will iterate through multiple development cycles, each building upon previous work while addressing new requirements and risks.
[Project introduction placeholder - describing the bioprinting and aging biomarker device project.]
Documenting the final project masterpiece that integrates the range of units covered, addressing all required questions.
[Placeholder: Describe the multimodal intrinsic capacity assessment system functionality]
[Placeholder: Literature review and prior work in aging biomarkers and intrinsic capacity assessment]
[Placeholder: Research papers, open-source projects, and reference materials]
[Placeholder: Custom hardware, software, and integration systems designed]
[Placeholder: Complete list of materials, sensors, microcontrollers, and components]
[Placeholder: Suppliers, vendors, and sourcing information]
[Placeholder: Detailed cost breakdown and budget analysis]
[Placeholder: Custom fabricated parts, 3D printed components, and assembled systems]
[Placeholder: Laser cutting, 3D printing, PCB fabrication, programming, and testing processes]
[Placeholder: Research questions addressed and hypotheses tested]
[Placeholder: Successes, challenges, and lessons learned]
[Placeholder: Testing protocols, validation methods, and evaluation criteria]
[Placeholder: Impact on aging research, clinical applications, and future development]
Your project should incorporate 2D and 3D design, multiple additive and subtractive fabrication processes, electronics design and production, embedded microcontroller design, interfacing, and programming, system integration and packaging.
2D design work for the multimodal intrinsic capacity assessment system:
Tools Used: Inkscape, KiCad, Adobe Illustrator, Figma
3D design work for device components and integration:
Tools Used: Fusion 360, FreeCAD, OpenSCAD, PrusaSlicer
Where possible, you should make rather than buy the parts of your project. Complete breakdown of materials, components, and sourcing information.
Complete list of materials and components:
[Placeholder: Detailed cost analysis and budget tracking]
Strategic decisions on fabrication vs. purchasing:
[Fabrication process documentation placeholder.]
[Testing and validation documentation placeholder.]
Projects can be separate or joint, but need to show individual mastery of the skills, and be independently operable.
Demonstration of individual skills across all course units:
Project operates independently without external dependencies:
Present your final project, weekly and group assignments, and documentation.
Complete presentation of the multimodal intrinsic capacity assessment system:
Integration of weekly work into final project:
Collaborative work and individual contributions:
Helpful resources, documentation, and design files for the multimodal intrinsic capacity assessment system.
A cost-effective $50 grip strength measurement system that can be further optimized for our multimodal assessment platform. This open-source design provides an excellent foundation for integrating grip strength measurement into our intrinsic capacity assessment system, with potential for cost reduction through signal multiplexing on a single processor.
Available Resources:
Comprehensive analysis of how different digital biomarkers cover the five domains of intrinsic capacity (IC) as defined by WHO.
Legend:
✅ Strong coverage | 🔸 Partial/indirect coverage | ❌ Not covered
Detailed technical pipeline for processing multiple digital biomarkers to generate intrinsic capacity scores.
Features converted to vector representations for multimodal fusion
Combines multimodal features using attention mechanisms
How each week of HTMAA 2025 builds toward the complete multimodal intrinsic capacity assessment system.
Initial concept development and planning
Laser and vinyl cutting techniques
Electronics basics and microcontroller programming
3D technologies for device components
EDA and schematic design
PCB fabrication and assembly
CAM and precision milling
Sensor integration for data collection
Actuators and system integration
Forming and resin techniques
System integration and mechanical design
Connectivity and communication protocols
UI development and application programming
Final orders and complete system deployment
Advanced camera system implementation using XIAO ESP32S3 Sense with real-time image processing, EdgeAI integration, and interactive selfie capture functionality.
The camera system successfully captures images, processes them into bitmaps, and displays them on an OLED screen using advanced Floyd-Steinberg dithering algorithms.
Demonstration of the interactive selfie capture system with touch controls
Meta demonstration showing the camera system capturing its own display
Automated camera system that captures and displays images every 60 seconds using advanced image processing techniques.
1. Initialize camera with PSRAM frame buffers
2. Configure OLED display (128x64 pixels)
3. Set up 60-second capture interval timer
4. In main loop:
a. Check if 60 seconds have elapsed
b. Capture image from camera
c. Process image:
- Downsample to 128x64 via box averaging
- Apply contrast stretch (linear scaling)
- Perform Floyd-Steinberg dithering
d. Display processed bitmap on OLED
e. Release frame buffer
5. Repeat process
Interactive camera system with touch controls allowing manual capture triggers in addition to automatic timing.
1. Initialize camera and OLED display
2. Set up touch pins (GPIO1 & GPIO2) with threshold detection
3. Configure 60-second auto-capture timer
4. In main loop:
a. Update touch sensor readings
b. Detect touch press events (justPressed)
c. Check for capture trigger:
- Touch press OR 60-second timer elapsed
d. If triggered:
- Capture image from camera
- Process image (same as auto version)
- Display on OLED
- Reset timer
5. Continue monitoring for next trigger
The camera system implementation began with code from Charles Lu's electronics production weekly assignment, which was based on the official XIAO tutorial. Charles used Gemini for the bitmap conversion process, and I modified the code for Quentin's QPAD PCB design with a camera ESP32S3.
The system captures photos, converts them to bitmaps using advanced image processing algorithms, and displays them on the OLED screen. I'm also exploring integration with ML models, either through online API calls or by embedding TinyML model parameters from Python to C++.
Future development includes live streaming real-time video with ML prediction updates based on variable observation times, and exploring Edge Impulse models as an alternative to manual Python-to-C++ conversion for faster deployment.
Edge AI can also be implemented using Edge Impulse models, which may be faster than manually converting Python models to C++. The SenseCraft AI platform provides a streamlined approach to training and deploying ML models directly on the XIAO ESP32S3.
Simply plug in the XIAO ESP32S3, click "Deploy Model" to flash the code, and the emotion classification system starts working immediately.
Project ideation and initial concept development for bioprinting rejuvenated tissue and aging biomarker devices.
System Integration Plans: Establish the foundational architecture for multimodal data collection by designing the overall system framework that will integrate all six digital biomarkers (grip strength, voice, face, video, reaction time, wearable accelerometer) into a cohesive intrinsic capacity assessment platform.
[Week 0 progress and contributions to final project placeholder]
Version control, laser cutting, and vinyl cutting techniques applied to final project components.
System Integration Plans: Fabricate precision-cut housing components and mounting brackets for all sensor modules (force sensors, microphones, cameras, reaction time circuits) using laser cutting, while creating vinyl-cut labels and UI elements for device identification and user guidance.
[Week 1 progress and contributions to final project placeholder]
Electronics basics and embedded programming for the aging biomarker device components.
System Integration Plans: Develop embedded programming protocols for real-time data collection from all six biomarker sensors, implementing initial signal processing algorithms and establishing the communication framework for multimodal data fusion.
[Week 2 progress and contributions to final project placeholder]
3D scanning and printing techniques for bioprinting components and device housings.
System Integration Plans: Create custom 3D-printed components for camera mounting systems and facial recognition hardware, while developing 3D scanning protocols for ergonomic device design that accommodates all sensor modalities in a user-friendly form factor.
[Week 3 progress and contributions to final project placeholder]
EDA and schematic design for the aging biomarker device electronics.
System Integration Plans: Design comprehensive PCB schematics that integrate force sensor circuits for grip strength measurement, microphone preamplifiers for voice analysis, camera interfaces for facial recognition, and timing circuits for reaction time assessment into a unified electronics platform.
[Week 4 progress and contributions to final project placeholder]
PCB fabrication, debugging, and assembly for the biomarker device.
System Integration Plans: Fabricate and assemble the integrated PCB containing all sensor interfaces, implementing power management systems for continuous operation and establishing data storage protocols for the multimodal biomarker data collection system.
[Week 5 progress and contributions to final project placeholder]
CAM and milling for precision components and device housings.
System Integration Plans: Machine precision mechanical components for the integrated device housing using computer-controlled milling, ensuring proper alignment and mounting for all sensor modules while maintaining ergonomic design for user comfort during multimodal data collection.
[Week 6 progress and contributions to final project placeholder]
Sensors and embedded architectures for data collection in the biomarker device.
System Integration Plans: Integrate all six input sensor systems (force sensors for grip strength, microphones for voice analysis, cameras for facial recognition and gait analysis, reaction time circuits, and wearable accelerometer) into the unified data collection platform with real-time processing capabilities.
[Week 7 progress and contributions to final project placeholder]
Actuators and system integration for the biomarker device outputs.
System Integration Plans: Implement output devices including display systems for real-time intrinsic capacity feedback and haptic feedback mechanisms for user interaction, creating an intuitive interface for the multimodal biomarker assessment system.
[Week 8 progress and contributions to final project placeholder]
Forming and resin techniques for bioprinting molds and device components.
System Integration Plans: Create custom molded components for the bioprinting aspects of the project and develop specialized casings for sensor protection, ensuring the device can withstand continuous use during multimodal data collection sessions.
[Week 9 progress and contributions to final project placeholder]
Kits and mechanical design for the bioprinting and biomarker device systems.
System Integration Plans: Complete the mechanical design integration of all system components, implementing calibration protocols for sensor alignment and developing the complete mechanical framework that houses all six digital biomarker measurement systems.
[Week 10 progress and contributions to final project placeholder]
BLE, Wi-Fi, and communication protocols for the biomarker device connectivity.
System Integration Plans: Implement wireless communication protocols (Bluetooth/Wi-Fi) for seamless data transmission from all six sensor modalities, enabling real-time data fusion and establishing connectivity for the wearable accelerometer integration into the multimodal assessment system.
[Week 11 progress and contributions to final project placeholder]
UI and application development for the biomarker device interface.
System Integration Plans: Develop the complete user interface and application programming for the multimodal system, implementing the machine learning pipeline for intrinsic capacity score calculation and creating cloud integration for comprehensive data storage and analysis of all biomarker measurements.
[Week 12 progress and contributions to final project placeholder]
Final orders and wildcard week activities for project completion.
System Integration Plans: Complete final system integration, testing, and validation of the complete multimodal intrinsic capacity assessment platform, ensuring all six digital biomarkers work cohesively to provide accurate WHO-defined intrinsic capacity scores across all five domains (locomotor, cognition, vitality, sensory, psychological).
[Week 13 progress and contributions to final project placeholder]
[Links to CAD/SVG/assets placeholder.]
[Reflection notes placeholder.]
Acknowledgements and contributions that made this project possible.
Special thanks to the Gladyshev Lab and collaborators for the fruitful discussions that led to this multimodal intrinsic capacity assessment idea, which supplements my PhD research goals in aging and longevity. The conceptual framework for integrating multiple digital biomarkers to assess intrinsic capacity domains emerged from collaborative research discussions on aging biomarkers and healthspan assessment.
Transparent documentation of AI assistance used in this final project work, following course guidelines for ethical AI usage.
Cursor AI assisted with creating the final project page structure, setting up navigation links, and organizing weekly progress sections. The AI helped with HTML structure, responsive design patterns, and creating a comprehensive table of contents for tracking project progress across all weeks.
ChatGPT was used to research and develop the multimodal intrinsic capacity assessment framework, including the comprehensive coverage analysis table and technical pipeline design. The AI assisted with structuring the WHO-defined intrinsic capacity domains, identifying appropriate digital biomarkers, and designing the fusion architecture for multimodal data processing.
Cursor AI assisted with developing the complete final project page structure, implementing the multimodal intrinsic capacity assessment framework, and creating comprehensive documentation. The AI helped with HTML structure, responsive design, weekly system integration plans, and organizing the technical documentation for the complete biomarker assessment system.
Cursor AI assisted with finalizing the project presentation structure to ensure full compliance with MIT Academy project presentation requirements. The AI helped implement all required sections including answering questions, design documentation, bill of materials, individual mastery requirements, course presentation structure, and spiral model development approach visualization.
ChatGPT was used to discuss camera system implementation strategies, image processing algorithms, and EdgeAI integration approaches. The AI assisted with understanding Floyd-Steinberg dithering implementation, touch sensor integration, and exploring TinyML deployment options for the XIAO ESP32S3 platform.
This work is licensed under a
Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License