changeset 327:d2cede794997

Fix core pep8
author Amine Sehili <amine.sehili@gmail.com>
date Mon, 21 Oct 2019 21:16:31 +0100
parents 1c1bec063103
children 85b4ba237e0f
files .pre-commit-config.yaml auditok/core.py tests/test_core.py
diffstat 3 files changed, 14 insertions(+), 13 deletions(-) [+]
line wrap: on
line diff
--- a/.pre-commit-config.yaml	Sun Oct 20 15:41:28 2019 +0100
+++ b/.pre-commit-config.yaml	Mon Oct 21 21:16:31 2019 +0100
@@ -7,4 +7,6 @@
 -   repo: https://github.com/pre-commit/pre-commit-hooks
     rev: v2.4.0
     hooks:
-    - id: flake8
\ No newline at end of file
+    - id: flake8
+    -   id: end-of-file-fixer
+    -   id: trailing-whitespace
--- a/auditok/core.py	Sun Oct 20 15:41:28 2019 +0100
+++ b/auditok/core.py	Mon Oct 21 21:16:31 2019 +0100
@@ -280,7 +280,6 @@
     try:
         while True:
             frame = reader.read()
-            print("read:", frame)
             if frame is None:
                 break
             data.append(frame)
@@ -831,7 +830,7 @@
 
         `validator` :
             Callable or an instance of DataValidator that implements
-            `is_valid` method. 
+            `is_valid` method.
 
         `min_length` : *(int)*
             Minimum number of frames of a valid token. This includes all \
@@ -851,8 +850,8 @@
             Minimum number of consecutive valid frames that must be **initially** \
             gathered before any sequence of non valid frames can be tolerated. This
             option is not always needed, it can be used to drop non-valid tokens as
-            early as possible. **Default = 0** means that the option is by default 
-            ineffective. 
+            early as possible. **Default = 0** means that the option is by default
+            ineffective.
 
         `init_max_silence` : *(int, default=0)*
             Maximum number of tolerated consecutive non-valid frames if the \
@@ -864,10 +863,10 @@
             `mode` can be:
 
         1. `StreamTokenizer.NORMAL`:
-        Do not drop trailing silence, and accept a token shorter than 
+        Do not drop trailing silence, and accept a token shorter than
         `min_length` if it is the continuation of the latest delivered token.
-        
-        2. `StreamTokenizer.STRICT_MIN_LENGTH`: 
+
+        2. `StreamTokenizer.STRICT_MIN_LENGTH`:
         if token *i* is delivered because `max_length`
         is reached, and token *i+1* is immediately adjacent to
         token *i* (i.e. token *i* ends at frame *k* and token *i+1* starts
@@ -911,7 +910,7 @@
         .. code:: python
 
             dsource = StringDataSource("aaaAAAABBbbb")
-            tokenizer = StreamTokenizer(validator=UpperCaseChecker(), 
+            tokenizer = StreamTokenizer(validator=UpperCaseChecker(),
                                         min_length=3, max_length=4,
                                         max_continuous_silence=0,
                                         mode=StreamTokenizer.STRICT_MIN_LENGTH)
@@ -1048,7 +1047,7 @@
     def _set_mode(self, mode):
         strict_min_and_drop_trailing = StreamTokenizer.STRICT_MIN_LENGTH
         strict_min_and_drop_trailing |= StreamTokenizer.DROP_TRAILING_SILENCE
-        if not mode in [
+        if mode not in [
             StreamTokenizer.NORMAL,
             StreamTokenizer.STRICT_MIN_LENGTH,
             StreamTokenizer.DROP_TRAILING_SILENCE,
@@ -1091,7 +1090,7 @@
                 (data, start, end)
 
            where `data` is a list of read frames, `start`: index of the first frame in the
-           original data and `end` : index of the last frame. 
+           original data and `end` : index of the last frame.
         """
         token_gen = self._iter_tokens(data_source)
         if callback:
@@ -1116,7 +1115,7 @@
             if token is not None:
                 yield token
 
-    def _process(self, frame):
+    def _process(self, frame):  # noqa: C901
 
         frame_is_valid = self._is_valid(frame)
 
--- a/tests/test_core.py	Sun Oct 20 15:41:28 2019 +0100
+++ b/tests/test_core.py	Mon Oct 21 21:16:31 2019 +0100
@@ -1656,6 +1656,6 @@
         try:
             # for numpy
             equal = equal.all()
-        except:
+        except AttributeError:
             pass
         self.assertTrue(equal)