Discussions will be organized around the following questions:
1) What areas of the LGT program in general, and the topic you are representing in particular, require a comprehensive study to be conducted as part of the Snowmass process in order to quantify the impact of the LGT results on improving phenomenological constraints and the overall experimental programs. i.e., are there areas for which we need to go beyond the USQCD whitepapers and do a more thorough study?
2) What are the computational, algorithmic, and human resource requirements of the program to achieve the impact identified and quantified in the previous question? What is the best HPC model that facilitates scientific progress in our community? If we were to have an input in the development of the upcoming machines and technologies, what would we propose? What is the significance of new classical algorithms, and how can they be combined with developing paradigms based on Machine Learning and Quantum Computing to expedite our scientific output already in the next decade?