S: One of our team's main responsibility is to provide utilization forecasts for the Supercharging team. They use our forecasted plan to understand when and where to expand in order to meet the growing EV demand from our Tesla fleet, as well as other EVs as we opened our network.
T: One component of the model is to take forecasted daily energy time series and transform it into daily utilization and potential congestion events. This is challenging because it is a rare observable event, with only less than 1% of site-days experiencing this on the network. When simplified into a classification problem, the training data is zero-inflated, showing a huge imbalance between days without waiters (0) and days with waiters (1), which is what we are trying to predict. Therefore, the challenge is two-fold: develop a model robust against the imbalanced data and ensure the validation set is representative of the future when the usage is close to full.
A: As the technical lead, I iterated on this model to improve its performance. The first iteration involved using a Tobit model formulation for "censored" data and class balancing techniques. Through extensive internal collaborations, I was able to repurpose a site-level queue simulation model built for a different purpose. I then created a site-level model with a Bayesian-like correction add-on, drastically improving the performance.
R: This approach maintained recall (don't miss from false negatives) while increasing precision (less false positives). Although the recommended network size globally did not change, the increased precision meant a much better allocation of our half a billion dollar budget that is invested into the expansion of the network.
S: After the 2021 winter storm in Texas, which triggered one of the worst energy infrastructure crises, Elon asked us to determine how to make our network "resilient" to ensure people can freely travel out of the state if needed.
T: This was an urgent request with a vague direction, so our team was tasked with both defining the metrics and creating a framework to calculate this metric. We decided to create a framework to simulate a set covering problem with existing and forecasted Supercharger sites. This framework could optimize against different layers, including historical EV trips, third-party trips, population, and roads, weighted against risks such as earthquakes, storms, and heat.
A: I was responsible for coding the component that ingests any user-defined period and prepares the various inputs and risk layers, ensuring they conform to a common dataset using SQL and PySpark. Additionally, I prepared the network scenarios for simulation. I helped design a modular framework with a clear interface between each module to enable effective collaborative coding.
R: The output included Pareto diagrams for each scenario, showing the number of sites (and their ranked order) versus the coverage of the target metric. We were able to answer Elon's resilience question by identifying which stations to back up in a disaster scenario, thus enabling the freedom to travel. Additionally, we repurposed the framework for new coverage and ranked the importance of planned sites.
S: In various discussions, it became clear that we lacked a simple way to distinguish the real-world differences between Supercharger technologies.
T: One of our team's core responsibilities is to measure the charging network. It fell to me to design an analysis that would become the de facto assessment of Supercharger technology differences.
A: On paper, the power difference between the technologies is 1.6x (250kW vs. 150kW). However, I had to consider several factors such as ambient temperature associated with geography and time of the year, battery level, vehicle platforms, software-enabled assistance, and site occupancy to provide a clear and confident measure of the difference between the technology types.
R: I was able to develop an evaluation framework that can be performed on a timely basis to provide the latest differences between Supercharger technologies. This framework allows us to make informed decisions and clearly communicate the performance of each technology....
S: We run a model that produces an energy forecast, which is a daily series of the energy expected to be demanded at each Supercharger site. This model is built using regression with several input variables. While this method is flexible to input changes and explainable, we faced a problem due to the lack of data cleaning in the training data and a bias towards capturing peaks. This led to unexpected forecasts that we could visually flag and use fallback models for, which is tedious given there are over 4,000 sites and we run this model every month.
T: To automate the anomaly flagging without relying on "clear" rules, we needed to implement several different anomaly detection checks. Some checks were based on heuristic rules such as negative energy growth, variability being x times higher, and large zero values. Others were implemented using anomaly detection packages, including change point detection.
A: I fine-tuned the algorithm using labeled data from previous rounds, ensuring that recall was maintained while increasing precision.
R: The automated label flags roughly 20%-30% of sites (800 sites), and we manually label less than 10% of those. We successfully fine-tuned the algorithm to reduce false positives by almost 50%.