sfood-checker(1) debian man page | unix.com

Man Page: sfood-checker

Operating Environment: debian

Section: 1

SFOOD-CHECKER(1)					      General Commands Manual						  SFOOD-CHECKER(1)

NAME
sfood-checker - check for superfluous import statements in Python source code
SYNOPSIS
sfood-checker [options] files...
DESCRIPTION
This script is used to detect forgotten imports that are not used anymore. When writing Python code (which happens so fast), it is often the case that we forget to remove useless imports. This is implemented using a search in the AST, and as such we do not require to import the module in order to run the checks. This is a major advantage over all the other lint/checker programs, and the main reason for taking the time to write it. As inputs, it can receive either files or directories; in case no argument is passed, it parses the current directory recursively.
OPTIONS
-h, --help show the help message and exit --debug Debugging output. -I IGNORES, --ignore=IGNORES Add the given directory name to the list to be ignored. -d, --disable-pragmas Disable processing of pragma directives as strings after imports. -D, --duplicates, --enable-duplicates Enable experimental heuristic for finding duplicate imports. -M, --missing, --enable-missing Enable experimental heuristic for finding missing imports.
SEE ALSO
sfood(1), sfood-cluster(1), sfood-copy(1), sfood-flatten(1), sfood-graph(1), sfood-imports(1).
AUTHOR
sfood-checker was written by Martin Blais <blais@furius.ca> and it's part of snakefood suite. This manual page was written by Sandro Tosi <morph@debian.org>, for the Debian project (and may be used by others). January 2, 2009 SFOOD-CHECKER(1)
Related Man Pages
cruft(8) - debian
module::load(3pm) - suse
swf::constants(3pm) - debian
namespace::autoclean(3) - osx
namespace::autoclean(3pm) - debian
Similar Topics in the Unix Linux Community
finding duplicates with perl
is /. superfluous? why not just say / ?
Finding duplicates from positioned substring across lines
For loop to check the files availability
script to remove duplicates per line