Workflow Deep Analysis Workflow Simple: Getting a simple workflow running, repeatably. Log Collection and Analysis: Gathering the logs from at least 2 sources and seeing if each gives us the data we need. Apache Airflow, Prefect, and/or live data from AmSC. a. The more sources the better. We can find what is missing with each and come up with an average structure across all the workflow systems. Workflow Reconstruction: Reconstructing a workflow from what it did along with timings and other metadata Workflow Serialization: Define what a workflow would be serialized into using existing or new methods. Some other AmSC interfaces are missing this structure. Workflow Visualization: Displaying the workflow, timeline based and/or graph based. Event Aggregation: Developing a tool, possibly using existing tools(ex: process_scenario.py) to generically create aggregate logs(points in time) into events(start->end something) by some identifier, is a workflow. I have a good base for event code in in here: https://git.cels.anl.gov/aig-public/octeres/-/tree/develop/src. Containerization: Creating a container that does all this for gathering every component of this in one easily digestible place. Workflow Inefficient: Create inefficient workflows and run them. Workflow AI4Ops: Using AI algorithms and other methods identify inefficiencies in workflows based on "happens-before" and other methods.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Intern
Education Level
No Education Listed