Skip to content

Commit

Permalink
Fix imports
Browse files Browse the repository at this point in the history
  • Loading branch information
Jonatas Leon committed Jun 21, 2022
1 parent 015a8d0 commit c056d67
Show file tree
Hide file tree
Showing 13 changed files with 36 additions and 48 deletions.
9 changes: 5 additions & 4 deletions das/das_generate_file.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,15 +3,16 @@
import os
import re
import shutil
from typing import Iterator, Optional
from typing import Iterator

from couchbase import exceptions as cb_exceptions
from couchbase.auth import PasswordAuthenticator
from couchbase.cluster import Cluster
from couchbase.management.collections import CollectionSpec
from helpers import get_logger, get_mongodb
from pymongo.collection import Collection
from util import AccumulatorClock, Clock, Statistics

from das.helpers import get_logger, get_mongodb
from das.util import AccumulatorClock, Clock, Statistics

logger = get_logger()

Expand All @@ -36,7 +37,7 @@
acc_clock_block5 = AccumulatorClock()
acc_clock_full = AccumulatorClock()

from hashing import Hasher
from das.hashing import Hasher


def populate_sets(hasher: Hasher, fh, collection: Collection, bucket, composite_keys_masks: dict[str, list[set[int]]]):
Expand Down
7 changes: 4 additions & 3 deletions das/das_upload_to_couch_from_file.py
Original file line number Diff line number Diff line change
@@ -1,13 +1,14 @@
import argparse
import datetime
import os
from typing import Iterator
from typing import Iterator, Tuple, List

from couchbase import exceptions as cb_exceptions
from couchbase.auth import PasswordAuthenticator
from couchbase.cluster import Cluster
from couchbase.management.collections import CollectionSpec
from helpers import get_logger

from das.helpers import get_logger

logger = get_logger()

Expand All @@ -23,7 +24,7 @@
def key_value_generator(
input_filename: str,
*,
block_size: int = MAX_BLOCK_SIZE) -> Iterator[tuple[str, list[str], int]]:
block_size: int = MAX_BLOCK_SIZE) -> Iterator[Tuple[str, List[str], int]]:
last_key = ''
last_list = []
counter = 0
Expand Down
3 changes: 1 addition & 2 deletions das/hashing.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,6 @@
from collections import defaultdict
from hashlib import md5
from operator import itemgetter
from typing import Any, Union
from typing import Union

from das.atomese2metta.translator import AtomType, Expression, MSet

Expand Down
2 changes: 1 addition & 1 deletion das/helpers.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@

from pymongo import MongoClient

from .atomese2metta.translator import AtomType, Type
from das.atomese2metta.translator import AtomType, Type


def get_mongodb(mongodb_specs):
Expand Down
3 changes: 2 additions & 1 deletion das/lex.py
Original file line number Diff line number Diff line change
@@ -1,8 +1,9 @@
import argparse

from helpers import get_logger
from ply.lex import lex

from das.helpers import get_logger

logger = get_logger()


Expand Down
9 changes: 5 additions & 4 deletions das/metta_lex.py
Original file line number Diff line number Diff line change
@@ -1,12 +1,13 @@
import argparse
from typing import Any, List, Union

from atomese2metta.translator import (AtomType, Expression, MSet,
UnorderedExpression)
from hashing import Hasher
from helpers import get_logger
from ply.lex import lex

from das.atomese2metta.translator import (AtomType, Expression, MSet,
UnorderedExpression)
from das.hashing import Hasher
from das.helpers import get_logger

logger = get_logger()


Expand Down
6 changes: 3 additions & 3 deletions das/parser.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,9 +3,9 @@
from datetime import datetime
from os import path

from atomese2metta.parser import LexParser
from atomese2metta.translator import Translator
from helpers import get_filesize_mb, get_logger, human_time
from das.atomese2metta.parser import LexParser
from das.atomese2metta.translator import Translator
from das.helpers import get_filesize_mb, get_logger, human_time

logger = get_logger()

Expand Down
15 changes: 4 additions & 11 deletions das/pattern_matcher/regression.py
Original file line number Diff line number Diff line change
@@ -1,17 +1,10 @@
from das.pattern_matcher.db_interface import DBInterface
from das.pattern_matcher.pattern_matcher import (And, Link, LogicalExpression,
Node, Not, Or,
PatternMatchingAnswer,
Variable)
from das.pattern_matcher.stub_db import StubDB

from das.pattern_matcher.pattern_matcher import (
And,
Link,
LogicalExpression,
Node,
Not,
Or,
PatternMatchingAnswer,
Variable,
)


def match(db_api: DBInterface, expression: LogicalExpression):
print(f"Matching {expression}")
Expand Down
14 changes: 2 additions & 12 deletions das/query.py
Original file line number Diff line number Diff line change
@@ -1,17 +1,7 @@
import argparse
import glob
import os
from datetime import datetime

from atomese2metta.translator import AtomType, Expression
from hashing import Hasher
from helpers import get_mongodb
from metta_lex import MettaParser
from pymongo.collection import Collection
from pymongo.database import Database
from pymongo.errors import DuplicateKeyError
from pymongo.operations import DeleteMany
from pymongo.results import InsertOneResult

from das.helpers import get_mongodb


class QueryDas:
Expand Down
2 changes: 1 addition & 1 deletion das/research/das_couch_cached.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,11 +7,11 @@
from couchbase.management.collections import CollectionSpec
from pymongo.collection import Collection
from pymongo.mongo_client import MongoClient
from util import AccumulatorClock, Clock, Statistics

from das.helpers import get_logger
from das.research.cache import (CachedCouchbaseClient, CouchbaseClient,
DocumentNotFoundException)
from das.util import AccumulatorClock, Clock, Statistics

logger = get_logger()

Expand Down
9 changes: 5 additions & 4 deletions das/unifier.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,12 +4,13 @@
from pprint import pprint
from typing import Any, List, Union

from atomese2metta.translator import (AtomType, Expression, MSet,
UnorderedExpression)
from hashing import Hasher
from helpers import get_mongodb
from ply.lex import lex

from das.atomese2metta.translator import (AtomType, Expression, MSet,
UnorderedExpression)
from das.hashing import Hasher
from das.helpers import get_mongodb


class DAS:
NODE_COLLS = [
Expand Down
3 changes: 2 additions & 1 deletion das/util.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,8 @@

import numpy as np
import pytest
from helpers import get_logger

from das.helpers import get_logger

logger = get_logger()

Expand Down
2 changes: 1 addition & 1 deletion tests/test_parser.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
from das.atomese2metta.parser import Parser, MultiprocessingParser, LexParser
from das.atomese2metta.parser import LexParser, MultiprocessingParser, Parser


def test_when_split_string_to_two_chunks():
Expand Down

0 comments on commit c056d67

Please sign in to comment.