[ https://issues.apache.org/jira/browse/SPARK-8706?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14611611#comment-14611611 ]
Manoj Kumar commented on SPARK-8706: ------------------------------------ Sorry for sounding dumb, but the present code downloads pep8 as a script. However it seems that pylint is a repo, which again has two dependencies. What is the preferred way to do this in Spark? > Implement Pylint / Prospector checks for PySpark > ------------------------------------------------ > > Key: SPARK-8706 > URL: https://issues.apache.org/jira/browse/SPARK-8706 > Project: Spark > Issue Type: New Feature > Components: Project Infra, PySpark > Reporter: Josh Rosen > > It would be nice to implement Pylint / Prospector > (https://github.com/landscapeio/prospector) checks for PySpark. As with the > style checker rules, I'll imagine that we'll want to roll out new rules > gradually in order to avoid a mass refactoring commit. > For starters, we should create a pull request that introduces the harness for > running the linters, add a configuration file which enables only the lint > checks that currently pass, and install the required dependencies on Jenkins. > Once we've done this, we can open a series of smaller followup PRs to > gradually enable more linting checks and to fix existing violations. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org