When looking at individual college baseball data for a season, one needs to be extremely careful. During this past season, the highest amount of Plate Appearances for a single player in the SEC, Big Ten, Big 12, ACC, or PAC-12 was 346. For the MLB, a single season worth of plate appearances ~500 is typically not enough to make certain conclusions about a player unless you’re Mike Trout.
For most college baseball players, there is significant variation in their rate statistics like strikeout rate, walk rate, and etc. This is a big reason why college baseball models tend to be pretty bad because they’re typically based on the mean values. This doesn’t mean that these models are useless, but need to be viewed as distributions in order to see the uncertainty of these predictions in small samples.
The big mistake that I have made and see many people make are viewing these statistics as definitive. For example, if a player with 250 PA has a higher strikeout rate than the league average. The league average strikeout rate is normalized because it is operating under a very large sample of plate appearances. However, an individual player’s strikeout rate will have a ton of variation because of the small sample of plate appearances. Because of the different distributions between player and league statistics, it is typically hard to make this conclusion definitive based on seasonal statistics. This is why modeling the rate statistics at the per-pitch level will likely help a ton with driving down this variation because pitches normalize significantly faster than plate appearances based on sample size. As I will stress a lot, normalization helps us make better predictions!
Below is a table with the 95% confidence intervals for individual player statistics with at least 50 PA in the SEC, Big Ten, Big 12, ACC, and PAC-12