NASA is working diligently to bring astronauts back to the moon in the upcoming Artemis missions. There are two main goals of the Artemis missions:
In line with the goals of the Artemis mission, NASA's JARVIS (aka Joint-AR) project investigates the potential of augmented reality (AR) displays in astronaut suits. There are numerous benefits of an astronaut AR informatics system:
Stemming from JARVIS, the NASA SUITS Challenge asks university teams to develop augmented reality (AR) interfaces for astronauts. The interface should aid the astronaut in navigation, geological sampling, tracking mission tasks, and more. The challenge outlines some requirements to define how our interface will be tested, but teams are also given a lot of flexibility in their designs and implementation.
Each year, nine teams are selected as finalists from universities nationwide, including Stanford, Duke, USC, Berkeley, Umich, Carnegie Mellon, etc. Although the SUITS Challenge is not a competition, the University of Michigan has (in my opinion) been the leader each year, demonstrating both high technical proficiency and a comprehensive user-centered design process.
I was the president of a 60-member development team at the University of Michigan called CLAWS. As both the project mnaager and AR development lead, I simultaneously oversaw the development and integration of features across 5 cross-functional teams whilst onboarding and assisting all AR developers.
Here are the final deliverables from my 2 years in CLAWS.
Each year, we spearheaded major changes, from development, to design, to team culture. We constantly had an outlook of improvement, which led to major success each year.
As the AR development lead, I did three main tasks:
The AR team consisted of around 12 new AR developers. I spent the first 2 months of the year to onboard the new team on augmented reality basics, the ins-and-outs of the HoloLens 2, Microsoft's Mixed Reality Toolkit, Unity Engine, C# scripting, software architecture patterns, and GitHub workflow. Each week, I gave a 1 hour interactive (and fun) lecture about a topic and also gave a project to put the learning into practice.
It was very fun to onboard new developers and create lessons plans, but next year, I have even more ideas for onboarding, involving having multiple instructors, a faster onboarding timeline, more projects, and new onboarding topics.
Since the vast majority of the AR team was new to AR development, I decided to develop most of the foundational software to lessen the learning curve for other developers. Additionally, since most of this foundational software would be used by all developers across all cross-functional feature teams, I developed most of the foundational software during the onboarding phase.
This foundational software included:
Throughout the year, I also developed more foundational software with certain members of the team, including:
During the onboarding phase, I clearly outlined our software architecture and created a data-flow model to help developers understand how each subsystem fits within the overall product.
Throughout the year, I oversaw the progress of AR development across all features and assisted with GitHub integration across all the features. I also helped members communicate well with other teams, especially the UX team.
This is one of the biggest areas where our team can improve. I am planning to create a CI/CD pipeline for next year in order to allow for easier integration and faster user testing with the UX team.
Patrick Halim - 8/4/23
I led a 60-member development team selected by NASA to develop 2 full augmented reality interfaces for astronauts. As both the project manager and AR development lead, I simultaneously oversaw the development and integration of features across 5 cross-functional teams whilst onboarding and assisting all AR developers. I also created foundational software for the AR team, such as a singleton interface for the backend, a publisher-subscriber event system, and a system state machine.
Technologies used: Unity Engine, C#, MRTK, HoloLens 2