diff options
author | Kevin Tindall <kevinkjt2000@users.noreply.github.com> | 2018-02-06 14:10:45 -0600 |
---|---|---|
committer | Dana Powers <dana.powers@gmail.com> | 2018-02-06 12:10:45 -0800 |
commit | 9ae0e590b5ea93a9c07d71b55b593e0fcc973bc6 (patch) | |
tree | 7baee93ef095c3694a9ddbc50809b2a681ef56f8 | |
parent | c0df771cf51bc27f2dd5d5af333666f4074e68c8 (diff) | |
download | kafka-python-9ae0e590b5ea93a9c07d71b55b593e0fcc973bc6.tar.gz |
use absolute imports everywhere (#1362)
43 files changed, 144 insertions, 144 deletions
diff --git a/kafka/__init__.py b/kafka/__init__.py index 6a80418..f108eff 100644 --- a/kafka/__init__.py +++ b/kafka/__init__.py @@ -1,7 +1,7 @@ from __future__ import absolute_import __title__ = 'kafka' -from .version import __version__ +from kafka.version import __version__ __author__ = 'Dana Powers' __license__ = 'Apache License 2.0' __copyright__ = 'Copyright 2016 Dana Powers, David Arthur, and Contributors' diff --git a/kafka/client_async.py b/kafka/client_async.py index 2e78e3b..4962d9f 100644 --- a/kafka/client_async.py +++ b/kafka/client_async.py @@ -13,26 +13,26 @@ try: import selectors # pylint: disable=import-error except ImportError: # vendored backport module - from .vendor import selectors34 as selectors + from kafka.vendor import selectors34 as selectors import socket import time from kafka.vendor import six -from .cluster import ClusterMetadata -from .conn import BrokerConnection, ConnectionStates, collect_hosts, get_ip_port_afi -from . import errors as Errors -from .future import Future -from .metrics import AnonMeasurable -from .metrics.stats import Avg, Count, Rate -from .metrics.stats.rate import TimeUnit -from .protocol.metadata import MetadataRequest -from .util import Dict, WeakMethod +from kafka.cluster import ClusterMetadata +from kafka.conn import BrokerConnection, ConnectionStates, collect_hosts, get_ip_port_afi +from kafka import errors as Errors +from kafka.future import Future +from kafka.metrics import AnonMeasurable +from kafka.metrics.stats import Avg, Count, Rate +from kafka.metrics.stats.rate import TimeUnit +from kafka.protocol.metadata import MetadataRequest +from kafka.util import Dict, WeakMethod # Although this looks unused, it actually monkey-patches socket.socketpair() # and should be left in as long as we're using socket.socketpair() in this file -from .vendor import socketpair -from .version import __version__ +from kafka.vendor import socketpair +from kafka.version import __version__ if six.PY2: ConnectionError = None diff --git a/kafka/cluster.py b/kafka/cluster.py index 1ab4218..5be3c2f 100644 --- a/kafka/cluster.py +++ b/kafka/cluster.py @@ -8,9 +8,9 @@ import time from kafka.vendor import six -from . import errors as Errors -from .future import Future -from .structs import BrokerMetadata, PartitionMetadata, TopicPartition +from kafka import errors as Errors +from kafka.future import Future +from kafka.structs import BrokerMetadata, PartitionMetadata, TopicPartition log = logging.getLogger(__name__) diff --git a/kafka/conn.py b/kafka/conn.py index 4fe5e21..0c8ae9a 100644 --- a/kafka/conn.py +++ b/kafka/conn.py @@ -12,7 +12,7 @@ try: import selectors # pylint: disable=import-error except ImportError: # vendored backport module - from .vendor import selectors34 as selectors + from kafka.vendor import selectors34 as selectors import socket import struct @@ -858,8 +858,8 @@ class BrokerConnection(object): # vanilla MetadataRequest. If the server did not recognize the first # request, both will be failed with a ConnectionError that wraps # socket.error (32, 54, or 104) - from .protocol.admin import ApiVersionRequest, ListGroupsRequest - from .protocol.commit import OffsetFetchRequest, GroupCoordinatorRequest + from kafka.protocol.admin import ApiVersionRequest, ListGroupsRequest + from kafka.protocol.commit import OffsetFetchRequest, GroupCoordinatorRequest # Socket errors are logged as exceptions and can alarm users. Mute them from logging import Filter diff --git a/kafka/consumer/__init__.py b/kafka/consumer/__init__.py index 36c8ff0..4b900ac 100644 --- a/kafka/consumer/__init__.py +++ b/kafka/consumer/__init__.py @@ -1,8 +1,8 @@ from __future__ import absolute_import -from .simple import SimpleConsumer -from .multiprocess import MultiProcessConsumer -from .group import KafkaConsumer +from kafka.consumer.simple import SimpleConsumer +from kafka.consumer.multiprocess import MultiProcessConsumer +from kafka.consumer.group import KafkaConsumer __all__ = [ 'SimpleConsumer', 'MultiProcessConsumer', 'KafkaConsumer' diff --git a/kafka/consumer/multiprocess.py b/kafka/consumer/multiprocess.py index 58e3e07..1da4a33 100644 --- a/kafka/consumer/multiprocess.py +++ b/kafka/consumer/multiprocess.py @@ -8,15 +8,15 @@ import warnings from kafka.vendor.six.moves import queue # pylint: disable=import-error -from ..common import KafkaError -from .base import ( +from kafka.common import KafkaError +from kafka.consumer.base import ( Consumer, AUTO_COMMIT_MSG_COUNT, AUTO_COMMIT_INTERVAL, NO_MESSAGES_WAIT_TIME_SECONDS, FULL_QUEUE_WAIT_TIME_SECONDS, MAX_BACKOFF_SECONDS, ) -from .simple import SimpleConsumer +from kafka.consumer.simple import SimpleConsumer log = logging.getLogger(__name__) diff --git a/kafka/consumer/simple.py b/kafka/consumer/simple.py index f5b6a99..c0c1b1e 100644 --- a/kafka/consumer/simple.py +++ b/kafka/consumer/simple.py @@ -12,7 +12,7 @@ import warnings from kafka.vendor import six from kafka.vendor.six.moves import queue # pylint: disable=import-error -from .base import ( +from kafka.consumer.base import ( Consumer, FETCH_DEFAULT_BLOCK_TIMEOUT, AUTO_COMMIT_MSG_COUNT, @@ -24,7 +24,7 @@ from .base import ( ITER_TIMEOUT_SECONDS, NO_MESSAGES_WAIT_TIME_SECONDS ) -from ..common import ( +from kafka.common import ( FetchRequestPayload, KafkaError, OffsetRequestPayload, ConsumerFetchSizeTooSmall, UnknownTopicOrPartitionError, NotLeaderForPartitionError, diff --git a/kafka/coordinator/assignors/range.py b/kafka/coordinator/assignors/range.py index cbf411e..c232d9e 100644 --- a/kafka/coordinator/assignors/range.py +++ b/kafka/coordinator/assignors/range.py @@ -5,8 +5,8 @@ import logging from kafka.vendor import six -from .abstract import AbstractPartitionAssignor -from ..protocol import ConsumerProtocolMemberMetadata, ConsumerProtocolMemberAssignment +from kafka.coordinator.assignors.abstract import AbstractPartitionAssignor +from kafka.coordinator.protocol import ConsumerProtocolMemberMetadata, ConsumerProtocolMemberAssignment log = logging.getLogger(__name__) diff --git a/kafka/coordinator/assignors/roundrobin.py b/kafka/coordinator/assignors/roundrobin.py index c241685..a831033 100644 --- a/kafka/coordinator/assignors/roundrobin.py +++ b/kafka/coordinator/assignors/roundrobin.py @@ -6,9 +6,9 @@ import logging from kafka.vendor import six -from .abstract import AbstractPartitionAssignor -from ...common import TopicPartition -from ..protocol import ConsumerProtocolMemberMetadata, ConsumerProtocolMemberAssignment +from kafka.coordinator.assignors.abstract import AbstractPartitionAssignor +from kafka.common import TopicPartition +from kafka.coordinator.protocol import ConsumerProtocolMemberMetadata, ConsumerProtocolMemberAssignment log = logging.getLogger(__name__) diff --git a/kafka/coordinator/base.py b/kafka/coordinator/base.py index 820fc1f..57da971 100644 --- a/kafka/coordinator/base.py +++ b/kafka/coordinator/base.py @@ -10,13 +10,13 @@ import weakref from kafka.vendor import six -from .heartbeat import Heartbeat -from .. import errors as Errors -from ..future import Future -from ..metrics import AnonMeasurable -from ..metrics.stats import Avg, Count, Max, Rate -from ..protocol.commit import GroupCoordinatorRequest, OffsetCommitRequest -from ..protocol.group import (HeartbeatRequest, JoinGroupRequest, +from kafka.coordinator.heartbeat import Heartbeat +from kafka import errors as Errors +from kafka.future import Future +from kafka.metrics import AnonMeasurable +from kafka.metrics.stats import Avg, Count, Max, Rate +from kafka.protocol.commit import GroupCoordinatorRequest, OffsetCommitRequest +from kafka.protocol.group import (HeartbeatRequest, JoinGroupRequest, LeaveGroupRequest, SyncGroupRequest) log = logging.getLogger('kafka.coordinator') diff --git a/kafka/coordinator/consumer.py b/kafka/coordinator/consumer.py index 9438a7e..7a22c61 100644 --- a/kafka/coordinator/consumer.py +++ b/kafka/coordinator/consumer.py @@ -7,17 +7,17 @@ import time from kafka.vendor import six -from .base import BaseCoordinator, Generation -from .assignors.range import RangePartitionAssignor -from .assignors.roundrobin import RoundRobinPartitionAssignor -from .protocol import ConsumerProtocol -from .. import errors as Errors -from ..future import Future -from ..metrics import AnonMeasurable -from ..metrics.stats import Avg, Count, Max, Rate -from ..protocol.commit import OffsetCommitRequest, OffsetFetchRequest -from ..structs import OffsetAndMetadata, TopicPartition -from ..util import WeakMethod +from kafka.coordinator.base import BaseCoordinator, Generation +from kafka.coordinator.assignors.range import RangePartitionAssignor +from kafka.coordinator.assignors.roundrobin import RoundRobinPartitionAssignor +from kafka.coordinator.protocol import ConsumerProtocol +from kafka import errors as Errors +from kafka.future import Future +from kafka.metrics import AnonMeasurable +from kafka.metrics.stats import Avg, Count, Max, Rate +from kafka.protocol.commit import OffsetCommitRequest, OffsetFetchRequest +from kafka.structs import OffsetAndMetadata, TopicPartition +from kafka.util import WeakMethod log = logging.getLogger(__name__) diff --git a/kafka/metrics/__init__.py b/kafka/metrics/__init__.py index 6055142..2a62d63 100644 --- a/kafka/metrics/__init__.py +++ b/kafka/metrics/__init__.py @@ -1,13 +1,13 @@ from __future__ import absolute_import -from .compound_stat import NamedMeasurable -from .dict_reporter import DictReporter -from .kafka_metric import KafkaMetric -from .measurable import AnonMeasurable -from .metric_config import MetricConfig -from .metric_name import MetricName -from .metrics import Metrics -from .quota import Quota +from kafka.metrics.compound_stat import NamedMeasurable +from kafka.metrics.dict_reporter import DictReporter +from kafka.metrics.kafka_metric import KafkaMetric +from kafka.metrics.measurable import AnonMeasurable +from kafka.metrics.metric_config import MetricConfig +from kafka.metrics.metric_name import MetricName +from kafka.metrics.metrics import Metrics +from kafka.metrics.quota import Quota __all__ = [ 'AnonMeasurable', 'DictReporter', 'KafkaMetric', 'MetricConfig', diff --git a/kafka/metrics/stats/__init__.py b/kafka/metrics/stats/__init__.py index ab1fb71..a3d535d 100644 --- a/kafka/metrics/stats/__init__.py +++ b/kafka/metrics/stats/__init__.py @@ -1,15 +1,15 @@ from __future__ import absolute_import -from .avg import Avg -from .count import Count -from .histogram import Histogram -from .max_stat import Max -from .min_stat import Min -from .percentile import Percentile -from .percentiles import Percentiles -from .rate import Rate -from .sensor import Sensor -from .total import Total +from kafka.metrics.stats.avg import Avg +from kafka.metrics.stats.count import Count +from kafka.metrics.stats.histogram import Histogram +from kafka.metrics.stats.max_stat import Max +from kafka.metrics.stats.min_stat import Min +from kafka.metrics.stats.percentile import Percentile +from kafka.metrics.stats.percentiles import Percentiles +from kafka.metrics.stats.rate import Rate +from kafka.metrics.stats.sensor import Sensor +from kafka.metrics.stats.total import Total __all__ = [ 'Avg', 'Count', 'Histogram', 'Max', 'Min', 'Percentile', 'Percentiles', diff --git a/kafka/partitioner/__init__.py b/kafka/partitioner/__init__.py index 299b485..a9dbbdc 100644 --- a/kafka/partitioner/__init__.py +++ b/kafka/partitioner/__init__.py @@ -1,8 +1,8 @@ from __future__ import absolute_import -from .default import DefaultPartitioner -from .hashed import HashedPartitioner, Murmur2Partitioner, LegacyPartitioner -from .roundrobin import RoundRobinPartitioner +from kafka.partitioner.default import DefaultPartitioner +from kafka.partitioner.hashed import HashedPartitioner, Murmur2Partitioner, LegacyPartitioner +from kafka.partitioner.roundrobin import RoundRobinPartitioner __all__ = [ 'DefaultPartitioner', 'RoundRobinPartitioner', 'HashedPartitioner', diff --git a/kafka/partitioner/default.py b/kafka/partitioner/default.py index 087166c..e4d9df5 100644 --- a/kafka/partitioner/default.py +++ b/kafka/partitioner/default.py @@ -2,7 +2,7 @@ from __future__ import absolute_import import random -from .hashed import murmur2 +from kafka.partitioner.hashed import murmur2 class DefaultPartitioner(object): diff --git a/kafka/partitioner/hashed.py b/kafka/partitioner/hashed.py index 06307f0..be92daf 100644 --- a/kafka/partitioner/hashed.py +++ b/kafka/partitioner/hashed.py @@ -2,7 +2,7 @@ from __future__ import absolute_import from kafka.vendor import six -from .base import Partitioner +from kafka.partitioner.base import Partitioner class Murmur2Partitioner(Partitioner): diff --git a/kafka/partitioner/roundrobin.py b/kafka/partitioner/roundrobin.py index 9ac2ed0..e68c372 100644 --- a/kafka/partitioner/roundrobin.py +++ b/kafka/partitioner/roundrobin.py @@ -1,6 +1,6 @@ from __future__ import absolute_import -from .base import Partitioner +from kafka.partitioner.base import Partitioner class RoundRobinPartitioner(Partitioner): diff --git a/kafka/producer/__init__.py b/kafka/producer/__init__.py index 5213fe8..54fd8d2 100644 --- a/kafka/producer/__init__.py +++ b/kafka/producer/__init__.py @@ -1,8 +1,8 @@ from __future__ import absolute_import -from .kafka import KafkaProducer -from .simple import SimpleProducer -from .keyed import KeyedProducer +from kafka.producer.kafka import KafkaProducer +from kafka.producer.simple import SimpleProducer +from kafka.producer.keyed import KeyedProducer __all__ = [ 'KafkaProducer', diff --git a/kafka/producer/buffer.py b/kafka/producer/buffer.py index 19ea732..8a8d717 100644 --- a/kafka/producer/buffer.py +++ b/kafka/producer/buffer.py @@ -5,7 +5,7 @@ import io import threading import time -from ..metrics.stats import Rate +from kafka.metrics.stats import Rate import kafka.errors as Errors diff --git a/kafka/producer/future.py b/kafka/producer/future.py index e39a0a9..aa216c4 100644 --- a/kafka/producer/future.py +++ b/kafka/producer/future.py @@ -3,8 +3,8 @@ from __future__ import absolute_import import collections import threading -from .. import errors as Errors -from ..future import Future +from kafka import errors as Errors +from kafka.future import Future class FutureProduceResult(Future): diff --git a/kafka/producer/kafka.py b/kafka/producer/kafka.py index 5e2fb69..ae28779 100644 --- a/kafka/producer/kafka.py +++ b/kafka/producer/kafka.py @@ -8,20 +8,20 @@ import threading import time import weakref -from ..vendor import six - -from .. import errors as Errors -from ..client_async import KafkaClient, selectors -from ..codec import has_gzip, has_snappy, has_lz4 -from ..metrics import MetricConfig, Metrics -from ..partitioner.default import DefaultPartitioner -from ..record.default_records import DefaultRecordBatchBuilder -from ..record.legacy_records import LegacyRecordBatchBuilder -from ..serializer import Serializer -from ..structs import TopicPartition -from .future import FutureRecordMetadata, FutureProduceResult -from .record_accumulator import AtomicInteger, RecordAccumulator -from .sender import Sender +from kafka.vendor import six + +from kafka import errors as Errors +from kafka.client_async import KafkaClient, selectors +from kafka.codec import has_gzip, has_snappy, has_lz4 +from kafka.metrics import MetricConfig, Metrics +from kafka.partitioner.default import DefaultPartitioner +from kafka.record.default_records import DefaultRecordBatchBuilder +from kafka.record.legacy_records import LegacyRecordBatchBuilder +from kafka.serializer import Serializer +from kafka.structs import TopicPartition +from kafka.producer.future import FutureRecordMetadata, FutureProduceResult +from kafka.producer.record_accumulator import AtomicInteger, RecordAccumulator +from kafka.producer.sender import Sender log = logging.getLogger(__name__) diff --git a/kafka/producer/keyed.py b/kafka/producer/keyed.py index 9fba33b..8de3ad8 100644 --- a/kafka/producer/keyed.py +++ b/kafka/producer/keyed.py @@ -3,8 +3,8 @@ from __future__ import absolute_import import logging import warnings -from .base import Producer -from ..partitioner import HashedPartitioner +from kafka.producer.base import Producer +from kafka.partitioner import HashedPartitioner log = logging.getLogger(__name__) diff --git a/kafka/producer/record_accumulator.py b/kafka/producer/record_accumulator.py index 5158474..61f1e0e 100644 --- a/kafka/producer/record_accumulator.py +++ b/kafka/producer/record_accumulator.py @@ -6,10 +6,10 @@ import logging import threading import time -from .. import errors as Errors -from .buffer import SimpleBufferPool -from .future import FutureRecordMetadata, FutureProduceResult -from ..structs import TopicPartition +from kafka import errors as Errors +from kafka.producer.buffer import SimpleBufferPool +from kafka.producer.future import FutureRecordMetadata, FutureProduceResult +from kafka.structs import TopicPartition from kafka.record.memory_records import MemoryRecordsBuilder from kafka.record.legacy_records import LegacyRecordBatchBuilder diff --git a/kafka/producer/sender.py b/kafka/producer/sender.py index 48ad06e..895045d 100644 --- a/kafka/producer/sender.py +++ b/kafka/producer/sender.py @@ -8,12 +8,12 @@ import time from kafka.vendor import six -from .. import errors as Errors -from ..metrics.measurable import AnonMeasurable -from ..metrics.stats import Avg, Max, Rate -from ..protocol.produce import ProduceRequest -from ..structs import TopicPartition -from ..version import __version__ +from kafka import errors as Errors +from kafka.metrics.measurable import AnonMeasurable +from kafka.metrics.stats import Avg, Max, Rate +from kafka.protocol.produce import ProduceRequest +from kafka.structs import TopicPartition +from kafka.version import __version__ log = logging.getLogger(__name__) diff --git a/kafka/producer/simple.py b/kafka/producer/simple.py index 90b3d4a..589363c 100644 --- a/kafka/producer/simple.py +++ b/kafka/producer/simple.py @@ -6,7 +6,7 @@ import random from kafka.vendor.six.moves import xrange # pylint: disable=import-error -from .base import Producer +from kafka.producer.base import Producer log = logging.getLogger(__name__) diff --git a/kafka/protocol/__init__.py b/kafka/protocol/__init__.py index 4dcf4a4..55ecb86 100644 --- a/kafka/protocol/__init__.py +++ b/kafka/protocol/__init__.py @@ -1,6 +1,6 @@ from __future__ import absolute_import -from .legacy import ( +from kafka.protocol.legacy import ( create_message, create_gzip_message, create_snappy_message, create_message_set, CODEC_NONE, CODEC_GZIP, CODEC_SNAPPY, ALL_CODECS, diff --git a/kafka/protocol/admin.py b/kafka/protocol/admin.py index b787c5f..9d4f17f 100644 --- a/kafka/protocol/admin.py +++ b/kafka/protocol/admin.py @@ -1,7 +1,7 @@ from __future__ import absolute_import -from .api import Request, Response -from .types import Array, Boolean, Bytes, Int16, Int32, Schema, String +from kafka.protocol.api import Request, Response +from kafka.protocol.types import Array, Boolean, Bytes, Int16, Int32, Schema, String class ApiVersionResponse_v0(Response): diff --git a/kafka/protocol/api.py b/kafka/protocol/api.py index ec24a39..efaf63e 100644 --- a/kafka/protocol/api.py +++ b/kafka/protocol/api.py @@ -2,8 +2,8 @@ from __future__ import absolute_import import abc -from .struct import Struct -from .types import Int16, Int32, String, Schema +from kafka.protocol.struct import Struct +from kafka.protocol.types import Int16, Int32, String, Schema class RequestHeader(Struct): diff --git a/kafka/protocol/commit.py b/kafka/protocol/commit.py index 9d744c7..31fc237 100644 --- a/kafka/protocol/commit.py +++ b/kafka/protocol/commit.py @@ -1,7 +1,7 @@ from __future__ import absolute_import -from .api import Request, Response -from .types import Array, Int8, Int16, Int32, Int64, Schema, String +from kafka.protocol.api import Request, Response +from kafka.protocol.types import Array, Int8, Int16, Int32, Int64, Schema, String class OffsetCommitResponse_v0(Response): diff --git a/kafka/protocol/fetch.py b/kafka/protocol/fetch.py index 5fc17e0..dd3f648 100644 --- a/kafka/protocol/fetch.py +++ b/kafka/protocol/fetch.py @@ -1,7 +1,7 @@ from __future__ import absolute_import -from .api import Request, Response -from .types import Array, Int8, Int16, Int32, Int64, Schema, String, Bytes +from kafka.protocol.api import Request, Response +from kafka.protocol.types import Array, Int8, Int16, Int32, Int64, Schema, String, Bytes class FetchResponse_v0(Response): diff --git a/kafka/protocol/group.py b/kafka/protocol/group.py index db84427..bcb9655 100644 --- a/kafka/protocol/group.py +++ b/kafka/protocol/group.py @@ -1,8 +1,8 @@ from __future__ import absolute_import -from .api import Request, Response -from .struct import Struct -from .types import Array, Bytes, Int16, Int32, Schema, String +from kafka.protocol.api import Request, Response +from kafka.protocol.struct import Struct +from kafka.protocol.types import Array, Bytes, Int16, Int32, Schema, String class JoinGroupResponse_v0(Response): diff --git a/kafka/protocol/message.py b/kafka/protocol/message.py index a330ed8..19dcbd9 100644 --- a/kafka/protocol/message.py +++ b/kafka/protocol/message.py @@ -3,15 +3,15 @@ from __future__ import absolute_import import io import time -from ..codec import (has_gzip, has_snappy, has_lz4, +from kafka.codec import (has_gzip, has_snappy, has_lz4, gzip_decode, snappy_decode, lz4_decode, lz4_decode_old_kafka) -from .frame import KafkaBytes -from .struct import Struct -from .types import ( +from kafka.protocol.frame import KafkaBytes +from kafka.protocol.struct import Struct +from kafka.protocol.types import ( Int8, Int32, Int64, Bytes, Schema, AbstractType ) -from ..util import crc32, WeakMethod +from kafka.util import crc32, WeakMethod class Message(Struct): diff --git a/kafka/protocol/metadata.py b/kafka/protocol/metadata.py index 2aafdd3..414e5b8 100644 --- a/kafka/protocol/metadata.py +++ b/kafka/protocol/metadata.py @@ -1,7 +1,7 @@ from __future__ import absolute_import -from .api import Request, Response -from .types import Array, Boolean, Int16, Int32, Schema, String +from kafka.protocol.api import Request, Response +from kafka.protocol.types import Array, Boolean, Int16, Int32, Schema, String class MetadataResponse_v0(Response): diff --git a/kafka/protocol/offset.py b/kafka/protocol/offset.py index 5179658..3c254de 100644 --- a/kafka/protocol/offset.py +++ b/kafka/protocol/offset.py @@ -1,7 +1,7 @@ from __future__ import absolute_import -from .api import Request, Response -from .types import Array, Int8, Int16, Int32, Int64, Schema, String +from kafka.protocol.api import Request, Response +from kafka.protocol.types import Array, Int8, Int16, Int32, Int64, Schema, String UNKNOWN_OFFSET = -1 diff --git a/kafka/protocol/produce.py b/kafka/protocol/produce.py index 5fbddec..f4032b3 100644 --- a/kafka/protocol/produce.py +++ b/kafka/protocol/produce.py @@ -1,7 +1,7 @@ from __future__ import absolute_import -from .api import Request, Response -from .types import Int16, Int32, Int64, String, Array, Schema, Bytes +from kafka.protocol.api import Request, Response +from kafka.protocol.types import Int16, Int32, Int64, String, Array, Schema, Bytes class ProduceResponse_v0(Response): diff --git a/kafka/protocol/struct.py b/kafka/protocol/struct.py index 3288172..676de1b 100644 --- a/kafka/protocol/struct.py +++ b/kafka/protocol/struct.py @@ -2,10 +2,10 @@ from __future__ import absolute_import from io import BytesIO -from .abstract import AbstractType -from .types import Schema +from kafka.protocol.abstract import AbstractType +from kafka.protocol.types import Schema -from ..util import WeakMethod +from kafka.util import WeakMethod class Struct(AbstractType): diff --git a/kafka/protocol/types.py b/kafka/protocol/types.py index 6a6e89e..d5e446a 100644 --- a/kafka/protocol/types.py +++ b/kafka/protocol/types.py @@ -2,7 +2,7 @@ from __future__ import absolute_import from struct import pack, unpack, error -from .abstract import AbstractType +from kafka.protocol.abstract import AbstractType def _pack(f, value): diff --git a/kafka/record/__init__.py b/kafka/record/__init__.py index 4c75acb..cbd70d9 100644 --- a/kafka/record/__init__.py +++ b/kafka/record/__init__.py @@ -1,3 +1,3 @@ -from .memory_records import MemoryRecords +from kafka.record.memory_records import MemoryRecords __all__ = ["MemoryRecords"] diff --git a/kafka/record/default_records.py b/kafka/record/default_records.py index 3d517af..2bbd47e 100644 --- a/kafka/record/default_records.py +++ b/kafka/record/default_records.py @@ -57,8 +57,8 @@ import io import struct import time -from .abc import ABCRecord, ABCRecordBatch, ABCRecordBatchBuilder -from .util import decode_varint, encode_varint, calc_crc32c, size_of_varint +from kafka.record.abc import ABCRecord, ABCRecordBatch, ABCRecordBatchBuilder +from kafka.record.util import decode_varint, encode_varint, calc_crc32c, size_of_varint from kafka.errors import CorruptRecordException from kafka.codec import ( diff --git a/kafka/record/legacy_records.py b/kafka/record/legacy_records.py index 8c0791e..036e6c4 100644 --- a/kafka/record/legacy_records.py +++ b/kafka/record/legacy_records.py @@ -44,8 +44,8 @@ import struct import time -from .abc import ABCRecord, ABCRecordBatch, ABCRecordBatchBuilder -from .util import calc_crc32 +from kafka.record.abc import ABCRecord, ABCRecordBatch, ABCRecordBatchBuilder +from kafka.record.util import calc_crc32 from kafka.codec import ( gzip_encode, snappy_encode, lz4_encode, lz4_encode_old_kafka, diff --git a/kafka/record/memory_records.py b/kafka/record/memory_records.py index 56aa51f..cb1cc01 100644 --- a/kafka/record/memory_records.py +++ b/kafka/record/memory_records.py @@ -22,9 +22,9 @@ import struct from kafka.errors import CorruptRecordException -from .abc import ABCRecords -from .legacy_records import LegacyRecordBatch, LegacyRecordBatchBuilder -from .default_records import DefaultRecordBatch, DefaultRecordBatchBuilder +from kafka.record.abc import ABCRecords +from kafka.record.legacy_records import LegacyRecordBatch, LegacyRecordBatchBuilder +from kafka.record.default_records import DefaultRecordBatch, DefaultRecordBatchBuilder class MemoryRecords(ABCRecords): diff --git a/kafka/record/util.py b/kafka/record/util.py index 88135f1..55d7adb 100644 --- a/kafka/record/util.py +++ b/kafka/record/util.py @@ -1,6 +1,6 @@ import binascii -from ._crc32c import crc as crc32c_py +from kafka.record._crc32c import crc as crc32c_py def encode_varint(value, write): diff --git a/kafka/serializer/__init__.py b/kafka/serializer/__init__.py index c08cffe..90cd93a 100644 --- a/kafka/serializer/__init__.py +++ b/kafka/serializer/__init__.py @@ -1,3 +1,3 @@ from __future__ import absolute_import -from .abstract import Serializer, Deserializer +from kafka.serializer.abstract import Serializer, Deserializer |