nchammas commented on code in PR #44546: URL: https://github.com/apache/spark/pull/44546#discussion_r1438914015
########## dev/py-tests/requirements-sql.txt: ########## @@ -0,0 +1,11 @@ +# PySpark dependencies for SQL tests + +numpy==1.26.2 Review Comment: Related prior discussion on pinning development dependencies: https://github.com/apache/spark/pull/27928#pullrequestreview-375712684 ########## dev/py-tests/requirements-sql.txt: ########## @@ -0,0 +1,11 @@ +# PySpark dependencies for SQL tests + +numpy==1.26.2 Review Comment: > actually, I think maybe we should always specify the versions I agree with this, and this is something I tried to do in the PR I linked to just above, but several committers were against it. When I look at the [number of PRs related to pinning dev dependencies][1] over the past three years, I wonder if committers still feel the same way today. Not pinning development dependencies creates constant breakages that can pop up whenever an upstream library releases a new version. When we pin dependencies, by contrast, we choose when to upgrade and deal with the potential breakage. [1]: https://github.com/apache/spark/pulls?q=is%3Apr+infra+pin -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
