summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorKevin Tindall <kevinkjt2000@users.noreply.github.com>2018-02-06 14:10:45 -0600
committerDana Powers <dana.powers@gmail.com>2018-02-06 12:10:45 -0800
commit9ae0e590b5ea93a9c07d71b55b593e0fcc973bc6 (patch)
tree7baee93ef095c3694a9ddbc50809b2a681ef56f8
parentc0df771cf51bc27f2dd5d5af333666f4074e68c8 (diff)
downloadkafka-python-9ae0e590b5ea93a9c07d71b55b593e0fcc973bc6.tar.gz
use absolute imports everywhere (#1362)
-rw-r--r--kafka/__init__.py2
-rw-r--r--kafka/client_async.py24
-rw-r--r--kafka/cluster.py6
-rw-r--r--kafka/conn.py6
-rw-r--r--kafka/consumer/__init__.py6
-rw-r--r--kafka/consumer/multiprocess.py6
-rw-r--r--kafka/consumer/simple.py4
-rw-r--r--kafka/coordinator/assignors/range.py4
-rw-r--r--kafka/coordinator/assignors/roundrobin.py6
-rw-r--r--kafka/coordinator/base.py14
-rw-r--r--kafka/coordinator/consumer.py22
-rw-r--r--kafka/metrics/__init__.py16
-rw-r--r--kafka/metrics/stats/__init__.py20
-rw-r--r--kafka/partitioner/__init__.py6
-rw-r--r--kafka/partitioner/default.py2
-rw-r--r--kafka/partitioner/hashed.py2
-rw-r--r--kafka/partitioner/roundrobin.py2
-rw-r--r--kafka/producer/__init__.py6
-rw-r--r--kafka/producer/buffer.py2
-rw-r--r--kafka/producer/future.py4
-rw-r--r--kafka/producer/kafka.py28
-rw-r--r--kafka/producer/keyed.py4
-rw-r--r--kafka/producer/record_accumulator.py8
-rw-r--r--kafka/producer/sender.py12
-rw-r--r--kafka/producer/simple.py2
-rw-r--r--kafka/protocol/__init__.py2
-rw-r--r--kafka/protocol/admin.py4
-rw-r--r--kafka/protocol/api.py4
-rw-r--r--kafka/protocol/commit.py4
-rw-r--r--kafka/protocol/fetch.py4
-rw-r--r--kafka/protocol/group.py6
-rw-r--r--kafka/protocol/message.py10
-rw-r--r--kafka/protocol/metadata.py4
-rw-r--r--kafka/protocol/offset.py4
-rw-r--r--kafka/protocol/produce.py4
-rw-r--r--kafka/protocol/struct.py6
-rw-r--r--kafka/protocol/types.py2
-rw-r--r--kafka/record/__init__.py2
-rw-r--r--kafka/record/default_records.py4
-rw-r--r--kafka/record/legacy_records.py4
-rw-r--r--kafka/record/memory_records.py6
-rw-r--r--kafka/record/util.py2
-rw-r--r--kafka/serializer/__init__.py2
43 files changed, 144 insertions, 144 deletions
diff --git a/kafka/__init__.py b/kafka/__init__.py
index 6a80418..f108eff 100644
--- a/kafka/__init__.py
+++ b/kafka/__init__.py
@@ -1,7 +1,7 @@
from __future__ import absolute_import
__title__ = 'kafka'
-from .version import __version__
+from kafka.version import __version__
__author__ = 'Dana Powers'
__license__ = 'Apache License 2.0'
__copyright__ = 'Copyright 2016 Dana Powers, David Arthur, and Contributors'
diff --git a/kafka/client_async.py b/kafka/client_async.py
index 2e78e3b..4962d9f 100644
--- a/kafka/client_async.py
+++ b/kafka/client_async.py
@@ -13,26 +13,26 @@ try:
import selectors # pylint: disable=import-error
except ImportError:
# vendored backport module
- from .vendor import selectors34 as selectors
+ from kafka.vendor import selectors34 as selectors
import socket
import time
from kafka.vendor import six
-from .cluster import ClusterMetadata
-from .conn import BrokerConnection, ConnectionStates, collect_hosts, get_ip_port_afi
-from . import errors as Errors
-from .future import Future
-from .metrics import AnonMeasurable
-from .metrics.stats import Avg, Count, Rate
-from .metrics.stats.rate import TimeUnit
-from .protocol.metadata import MetadataRequest
-from .util import Dict, WeakMethod
+from kafka.cluster import ClusterMetadata
+from kafka.conn import BrokerConnection, ConnectionStates, collect_hosts, get_ip_port_afi
+from kafka import errors as Errors
+from kafka.future import Future
+from kafka.metrics import AnonMeasurable
+from kafka.metrics.stats import Avg, Count, Rate
+from kafka.metrics.stats.rate import TimeUnit
+from kafka.protocol.metadata import MetadataRequest
+from kafka.util import Dict, WeakMethod
# Although this looks unused, it actually monkey-patches socket.socketpair()
# and should be left in as long as we're using socket.socketpair() in this file
-from .vendor import socketpair
-from .version import __version__
+from kafka.vendor import socketpair
+from kafka.version import __version__
if six.PY2:
ConnectionError = None
diff --git a/kafka/cluster.py b/kafka/cluster.py
index 1ab4218..5be3c2f 100644
--- a/kafka/cluster.py
+++ b/kafka/cluster.py
@@ -8,9 +8,9 @@ import time
from kafka.vendor import six
-from . import errors as Errors
-from .future import Future
-from .structs import BrokerMetadata, PartitionMetadata, TopicPartition
+from kafka import errors as Errors
+from kafka.future import Future
+from kafka.structs import BrokerMetadata, PartitionMetadata, TopicPartition
log = logging.getLogger(__name__)
diff --git a/kafka/conn.py b/kafka/conn.py
index 4fe5e21..0c8ae9a 100644
--- a/kafka/conn.py
+++ b/kafka/conn.py
@@ -12,7 +12,7 @@ try:
import selectors # pylint: disable=import-error
except ImportError:
# vendored backport module
- from .vendor import selectors34 as selectors
+ from kafka.vendor import selectors34 as selectors
import socket
import struct
@@ -858,8 +858,8 @@ class BrokerConnection(object):
# vanilla MetadataRequest. If the server did not recognize the first
# request, both will be failed with a ConnectionError that wraps
# socket.error (32, 54, or 104)
- from .protocol.admin import ApiVersionRequest, ListGroupsRequest
- from .protocol.commit import OffsetFetchRequest, GroupCoordinatorRequest
+ from kafka.protocol.admin import ApiVersionRequest, ListGroupsRequest
+ from kafka.protocol.commit import OffsetFetchRequest, GroupCoordinatorRequest
# Socket errors are logged as exceptions and can alarm users. Mute them
from logging import Filter
diff --git a/kafka/consumer/__init__.py b/kafka/consumer/__init__.py
index 36c8ff0..4b900ac 100644
--- a/kafka/consumer/__init__.py
+++ b/kafka/consumer/__init__.py
@@ -1,8 +1,8 @@
from __future__ import absolute_import
-from .simple import SimpleConsumer
-from .multiprocess import MultiProcessConsumer
-from .group import KafkaConsumer
+from kafka.consumer.simple import SimpleConsumer
+from kafka.consumer.multiprocess import MultiProcessConsumer
+from kafka.consumer.group import KafkaConsumer
__all__ = [
'SimpleConsumer', 'MultiProcessConsumer', 'KafkaConsumer'
diff --git a/kafka/consumer/multiprocess.py b/kafka/consumer/multiprocess.py
index 58e3e07..1da4a33 100644
--- a/kafka/consumer/multiprocess.py
+++ b/kafka/consumer/multiprocess.py
@@ -8,15 +8,15 @@ import warnings
from kafka.vendor.six.moves import queue # pylint: disable=import-error
-from ..common import KafkaError
-from .base import (
+from kafka.common import KafkaError
+from kafka.consumer.base import (
Consumer,
AUTO_COMMIT_MSG_COUNT, AUTO_COMMIT_INTERVAL,
NO_MESSAGES_WAIT_TIME_SECONDS,
FULL_QUEUE_WAIT_TIME_SECONDS,
MAX_BACKOFF_SECONDS,
)
-from .simple import SimpleConsumer
+from kafka.consumer.simple import SimpleConsumer
log = logging.getLogger(__name__)
diff --git a/kafka/consumer/simple.py b/kafka/consumer/simple.py
index f5b6a99..c0c1b1e 100644
--- a/kafka/consumer/simple.py
+++ b/kafka/consumer/simple.py
@@ -12,7 +12,7 @@ import warnings
from kafka.vendor import six
from kafka.vendor.six.moves import queue # pylint: disable=import-error
-from .base import (
+from kafka.consumer.base import (
Consumer,
FETCH_DEFAULT_BLOCK_TIMEOUT,
AUTO_COMMIT_MSG_COUNT,
@@ -24,7 +24,7 @@ from .base import (
ITER_TIMEOUT_SECONDS,
NO_MESSAGES_WAIT_TIME_SECONDS
)
-from ..common import (
+from kafka.common import (
FetchRequestPayload, KafkaError, OffsetRequestPayload,
ConsumerFetchSizeTooSmall,
UnknownTopicOrPartitionError, NotLeaderForPartitionError,
diff --git a/kafka/coordinator/assignors/range.py b/kafka/coordinator/assignors/range.py
index cbf411e..c232d9e 100644
--- a/kafka/coordinator/assignors/range.py
+++ b/kafka/coordinator/assignors/range.py
@@ -5,8 +5,8 @@ import logging
from kafka.vendor import six
-from .abstract import AbstractPartitionAssignor
-from ..protocol import ConsumerProtocolMemberMetadata, ConsumerProtocolMemberAssignment
+from kafka.coordinator.assignors.abstract import AbstractPartitionAssignor
+from kafka.coordinator.protocol import ConsumerProtocolMemberMetadata, ConsumerProtocolMemberAssignment
log = logging.getLogger(__name__)
diff --git a/kafka/coordinator/assignors/roundrobin.py b/kafka/coordinator/assignors/roundrobin.py
index c241685..a831033 100644
--- a/kafka/coordinator/assignors/roundrobin.py
+++ b/kafka/coordinator/assignors/roundrobin.py
@@ -6,9 +6,9 @@ import logging
from kafka.vendor import six
-from .abstract import AbstractPartitionAssignor
-from ...common import TopicPartition
-from ..protocol import ConsumerProtocolMemberMetadata, ConsumerProtocolMemberAssignment
+from kafka.coordinator.assignors.abstract import AbstractPartitionAssignor
+from kafka.common import TopicPartition
+from kafka.coordinator.protocol import ConsumerProtocolMemberMetadata, ConsumerProtocolMemberAssignment
log = logging.getLogger(__name__)
diff --git a/kafka/coordinator/base.py b/kafka/coordinator/base.py
index 820fc1f..57da971 100644
--- a/kafka/coordinator/base.py
+++ b/kafka/coordinator/base.py
@@ -10,13 +10,13 @@ import weakref
from kafka.vendor import six
-from .heartbeat import Heartbeat
-from .. import errors as Errors
-from ..future import Future
-from ..metrics import AnonMeasurable
-from ..metrics.stats import Avg, Count, Max, Rate
-from ..protocol.commit import GroupCoordinatorRequest, OffsetCommitRequest
-from ..protocol.group import (HeartbeatRequest, JoinGroupRequest,
+from kafka.coordinator.heartbeat import Heartbeat
+from kafka import errors as Errors
+from kafka.future import Future
+from kafka.metrics import AnonMeasurable
+from kafka.metrics.stats import Avg, Count, Max, Rate
+from kafka.protocol.commit import GroupCoordinatorRequest, OffsetCommitRequest
+from kafka.protocol.group import (HeartbeatRequest, JoinGroupRequest,
LeaveGroupRequest, SyncGroupRequest)
log = logging.getLogger('kafka.coordinator')
diff --git a/kafka/coordinator/consumer.py b/kafka/coordinator/consumer.py
index 9438a7e..7a22c61 100644
--- a/kafka/coordinator/consumer.py
+++ b/kafka/coordinator/consumer.py
@@ -7,17 +7,17 @@ import time
from kafka.vendor import six
-from .base import BaseCoordinator, Generation
-from .assignors.range import RangePartitionAssignor
-from .assignors.roundrobin import RoundRobinPartitionAssignor
-from .protocol import ConsumerProtocol
-from .. import errors as Errors
-from ..future import Future
-from ..metrics import AnonMeasurable
-from ..metrics.stats import Avg, Count, Max, Rate
-from ..protocol.commit import OffsetCommitRequest, OffsetFetchRequest
-from ..structs import OffsetAndMetadata, TopicPartition
-from ..util import WeakMethod
+from kafka.coordinator.base import BaseCoordinator, Generation
+from kafka.coordinator.assignors.range import RangePartitionAssignor
+from kafka.coordinator.assignors.roundrobin import RoundRobinPartitionAssignor
+from kafka.coordinator.protocol import ConsumerProtocol
+from kafka import errors as Errors
+from kafka.future import Future
+from kafka.metrics import AnonMeasurable
+from kafka.metrics.stats import Avg, Count, Max, Rate
+from kafka.protocol.commit import OffsetCommitRequest, OffsetFetchRequest
+from kafka.structs import OffsetAndMetadata, TopicPartition
+from kafka.util import WeakMethod
log = logging.getLogger(__name__)
diff --git a/kafka/metrics/__init__.py b/kafka/metrics/__init__.py
index 6055142..2a62d63 100644
--- a/kafka/metrics/__init__.py
+++ b/kafka/metrics/__init__.py
@@ -1,13 +1,13 @@
from __future__ import absolute_import
-from .compound_stat import NamedMeasurable
-from .dict_reporter import DictReporter
-from .kafka_metric import KafkaMetric
-from .measurable import AnonMeasurable
-from .metric_config import MetricConfig
-from .metric_name import MetricName
-from .metrics import Metrics
-from .quota import Quota
+from kafka.metrics.compound_stat import NamedMeasurable
+from kafka.metrics.dict_reporter import DictReporter
+from kafka.metrics.kafka_metric import KafkaMetric
+from kafka.metrics.measurable import AnonMeasurable
+from kafka.metrics.metric_config import MetricConfig
+from kafka.metrics.metric_name import MetricName
+from kafka.metrics.metrics import Metrics
+from kafka.metrics.quota import Quota
__all__ = [
'AnonMeasurable', 'DictReporter', 'KafkaMetric', 'MetricConfig',
diff --git a/kafka/metrics/stats/__init__.py b/kafka/metrics/stats/__init__.py
index ab1fb71..a3d535d 100644
--- a/kafka/metrics/stats/__init__.py
+++ b/kafka/metrics/stats/__init__.py
@@ -1,15 +1,15 @@
from __future__ import absolute_import
-from .avg import Avg
-from .count import Count
-from .histogram import Histogram
-from .max_stat import Max
-from .min_stat import Min
-from .percentile import Percentile
-from .percentiles import Percentiles
-from .rate import Rate
-from .sensor import Sensor
-from .total import Total
+from kafka.metrics.stats.avg import Avg
+from kafka.metrics.stats.count import Count
+from kafka.metrics.stats.histogram import Histogram
+from kafka.metrics.stats.max_stat import Max
+from kafka.metrics.stats.min_stat import Min
+from kafka.metrics.stats.percentile import Percentile
+from kafka.metrics.stats.percentiles import Percentiles
+from kafka.metrics.stats.rate import Rate
+from kafka.metrics.stats.sensor import Sensor
+from kafka.metrics.stats.total import Total
__all__ = [
'Avg', 'Count', 'Histogram', 'Max', 'Min', 'Percentile', 'Percentiles',
diff --git a/kafka/partitioner/__init__.py b/kafka/partitioner/__init__.py
index 299b485..a9dbbdc 100644
--- a/kafka/partitioner/__init__.py
+++ b/kafka/partitioner/__init__.py
@@ -1,8 +1,8 @@
from __future__ import absolute_import
-from .default import DefaultPartitioner
-from .hashed import HashedPartitioner, Murmur2Partitioner, LegacyPartitioner
-from .roundrobin import RoundRobinPartitioner
+from kafka.partitioner.default import DefaultPartitioner
+from kafka.partitioner.hashed import HashedPartitioner, Murmur2Partitioner, LegacyPartitioner
+from kafka.partitioner.roundrobin import RoundRobinPartitioner
__all__ = [
'DefaultPartitioner', 'RoundRobinPartitioner', 'HashedPartitioner',
diff --git a/kafka/partitioner/default.py b/kafka/partitioner/default.py
index 087166c..e4d9df5 100644
--- a/kafka/partitioner/default.py
+++ b/kafka/partitioner/default.py
@@ -2,7 +2,7 @@ from __future__ import absolute_import
import random
-from .hashed import murmur2
+from kafka.partitioner.hashed import murmur2
class DefaultPartitioner(object):
diff --git a/kafka/partitioner/hashed.py b/kafka/partitioner/hashed.py
index 06307f0..be92daf 100644
--- a/kafka/partitioner/hashed.py
+++ b/kafka/partitioner/hashed.py
@@ -2,7 +2,7 @@ from __future__ import absolute_import
from kafka.vendor import six
-from .base import Partitioner
+from kafka.partitioner.base import Partitioner
class Murmur2Partitioner(Partitioner):
diff --git a/kafka/partitioner/roundrobin.py b/kafka/partitioner/roundrobin.py
index 9ac2ed0..e68c372 100644
--- a/kafka/partitioner/roundrobin.py
+++ b/kafka/partitioner/roundrobin.py
@@ -1,6 +1,6 @@
from __future__ import absolute_import
-from .base import Partitioner
+from kafka.partitioner.base import Partitioner
class RoundRobinPartitioner(Partitioner):
diff --git a/kafka/producer/__init__.py b/kafka/producer/__init__.py
index 5213fe8..54fd8d2 100644
--- a/kafka/producer/__init__.py
+++ b/kafka/producer/__init__.py
@@ -1,8 +1,8 @@
from __future__ import absolute_import
-from .kafka import KafkaProducer
-from .simple import SimpleProducer
-from .keyed import KeyedProducer
+from kafka.producer.kafka import KafkaProducer
+from kafka.producer.simple import SimpleProducer
+from kafka.producer.keyed import KeyedProducer
__all__ = [
'KafkaProducer',
diff --git a/kafka/producer/buffer.py b/kafka/producer/buffer.py
index 19ea732..8a8d717 100644
--- a/kafka/producer/buffer.py
+++ b/kafka/producer/buffer.py
@@ -5,7 +5,7 @@ import io
import threading
import time
-from ..metrics.stats import Rate
+from kafka.metrics.stats import Rate
import kafka.errors as Errors
diff --git a/kafka/producer/future.py b/kafka/producer/future.py
index e39a0a9..aa216c4 100644
--- a/kafka/producer/future.py
+++ b/kafka/producer/future.py
@@ -3,8 +3,8 @@ from __future__ import absolute_import
import collections
import threading
-from .. import errors as Errors
-from ..future import Future
+from kafka import errors as Errors
+from kafka.future import Future
class FutureProduceResult(Future):
diff --git a/kafka/producer/kafka.py b/kafka/producer/kafka.py
index 5e2fb69..ae28779 100644
--- a/kafka/producer/kafka.py
+++ b/kafka/producer/kafka.py
@@ -8,20 +8,20 @@ import threading
import time
import weakref
-from ..vendor import six
-
-from .. import errors as Errors
-from ..client_async import KafkaClient, selectors
-from ..codec import has_gzip, has_snappy, has_lz4
-from ..metrics import MetricConfig, Metrics
-from ..partitioner.default import DefaultPartitioner
-from ..record.default_records import DefaultRecordBatchBuilder
-from ..record.legacy_records import LegacyRecordBatchBuilder
-from ..serializer import Serializer
-from ..structs import TopicPartition
-from .future import FutureRecordMetadata, FutureProduceResult
-from .record_accumulator import AtomicInteger, RecordAccumulator
-from .sender import Sender
+from kafka.vendor import six
+
+from kafka import errors as Errors
+from kafka.client_async import KafkaClient, selectors
+from kafka.codec import has_gzip, has_snappy, has_lz4
+from kafka.metrics import MetricConfig, Metrics
+from kafka.partitioner.default import DefaultPartitioner
+from kafka.record.default_records import DefaultRecordBatchBuilder
+from kafka.record.legacy_records import LegacyRecordBatchBuilder
+from kafka.serializer import Serializer
+from kafka.structs import TopicPartition
+from kafka.producer.future import FutureRecordMetadata, FutureProduceResult
+from kafka.producer.record_accumulator import AtomicInteger, RecordAccumulator
+from kafka.producer.sender import Sender
log = logging.getLogger(__name__)
diff --git a/kafka/producer/keyed.py b/kafka/producer/keyed.py
index 9fba33b..8de3ad8 100644
--- a/kafka/producer/keyed.py
+++ b/kafka/producer/keyed.py
@@ -3,8 +3,8 @@ from __future__ import absolute_import
import logging
import warnings
-from .base import Producer
-from ..partitioner import HashedPartitioner
+from kafka.producer.base import Producer
+from kafka.partitioner import HashedPartitioner
log = logging.getLogger(__name__)
diff --git a/kafka/producer/record_accumulator.py b/kafka/producer/record_accumulator.py
index 5158474..61f1e0e 100644
--- a/kafka/producer/record_accumulator.py
+++ b/kafka/producer/record_accumulator.py
@@ -6,10 +6,10 @@ import logging
import threading
import time
-from .. import errors as Errors
-from .buffer import SimpleBufferPool
-from .future import FutureRecordMetadata, FutureProduceResult
-from ..structs import TopicPartition
+from kafka import errors as Errors
+from kafka.producer.buffer import SimpleBufferPool
+from kafka.producer.future import FutureRecordMetadata, FutureProduceResult
+from kafka.structs import TopicPartition
from kafka.record.memory_records import MemoryRecordsBuilder
from kafka.record.legacy_records import LegacyRecordBatchBuilder
diff --git a/kafka/producer/sender.py b/kafka/producer/sender.py
index 48ad06e..895045d 100644
--- a/kafka/producer/sender.py
+++ b/kafka/producer/sender.py
@@ -8,12 +8,12 @@ import time
from kafka.vendor import six
-from .. import errors as Errors
-from ..metrics.measurable import AnonMeasurable
-from ..metrics.stats import Avg, Max, Rate
-from ..protocol.produce import ProduceRequest
-from ..structs import TopicPartition
-from ..version import __version__
+from kafka import errors as Errors
+from kafka.metrics.measurable import AnonMeasurable
+from kafka.metrics.stats import Avg, Max, Rate
+from kafka.protocol.produce import ProduceRequest
+from kafka.structs import TopicPartition
+from kafka.version import __version__
log = logging.getLogger(__name__)
diff --git a/kafka/producer/simple.py b/kafka/producer/simple.py
index 90b3d4a..589363c 100644
--- a/kafka/producer/simple.py
+++ b/kafka/producer/simple.py
@@ -6,7 +6,7 @@ import random
from kafka.vendor.six.moves import xrange # pylint: disable=import-error
-from .base import Producer
+from kafka.producer.base import Producer
log = logging.getLogger(__name__)
diff --git a/kafka/protocol/__init__.py b/kafka/protocol/__init__.py
index 4dcf4a4..55ecb86 100644
--- a/kafka/protocol/__init__.py
+++ b/kafka/protocol/__init__.py
@@ -1,6 +1,6 @@
from __future__ import absolute_import
-from .legacy import (
+from kafka.protocol.legacy import (
create_message, create_gzip_message,
create_snappy_message, create_message_set,
CODEC_NONE, CODEC_GZIP, CODEC_SNAPPY, ALL_CODECS,
diff --git a/kafka/protocol/admin.py b/kafka/protocol/admin.py
index b787c5f..9d4f17f 100644
--- a/kafka/protocol/admin.py
+++ b/kafka/protocol/admin.py
@@ -1,7 +1,7 @@
from __future__ import absolute_import
-from .api import Request, Response
-from .types import Array, Boolean, Bytes, Int16, Int32, Schema, String
+from kafka.protocol.api import Request, Response
+from kafka.protocol.types import Array, Boolean, Bytes, Int16, Int32, Schema, String
class ApiVersionResponse_v0(Response):
diff --git a/kafka/protocol/api.py b/kafka/protocol/api.py
index ec24a39..efaf63e 100644
--- a/kafka/protocol/api.py
+++ b/kafka/protocol/api.py
@@ -2,8 +2,8 @@ from __future__ import absolute_import
import abc
-from .struct import Struct
-from .types import Int16, Int32, String, Schema
+from kafka.protocol.struct import Struct
+from kafka.protocol.types import Int16, Int32, String, Schema
class RequestHeader(Struct):
diff --git a/kafka/protocol/commit.py b/kafka/protocol/commit.py
index 9d744c7..31fc237 100644
--- a/kafka/protocol/commit.py
+++ b/kafka/protocol/commit.py
@@ -1,7 +1,7 @@
from __future__ import absolute_import
-from .api import Request, Response
-from .types import Array, Int8, Int16, Int32, Int64, Schema, String
+from kafka.protocol.api import Request, Response
+from kafka.protocol.types import Array, Int8, Int16, Int32, Int64, Schema, String
class OffsetCommitResponse_v0(Response):
diff --git a/kafka/protocol/fetch.py b/kafka/protocol/fetch.py
index 5fc17e0..dd3f648 100644
--- a/kafka/protocol/fetch.py
+++ b/kafka/protocol/fetch.py
@@ -1,7 +1,7 @@
from __future__ import absolute_import
-from .api import Request, Response
-from .types import Array, Int8, Int16, Int32, Int64, Schema, String, Bytes
+from kafka.protocol.api import Request, Response
+from kafka.protocol.types import Array, Int8, Int16, Int32, Int64, Schema, String, Bytes
class FetchResponse_v0(Response):
diff --git a/kafka/protocol/group.py b/kafka/protocol/group.py
index db84427..bcb9655 100644
--- a/kafka/protocol/group.py
+++ b/kafka/protocol/group.py
@@ -1,8 +1,8 @@
from __future__ import absolute_import
-from .api import Request, Response
-from .struct import Struct
-from .types import Array, Bytes, Int16, Int32, Schema, String
+from kafka.protocol.api import Request, Response
+from kafka.protocol.struct import Struct
+from kafka.protocol.types import Array, Bytes, Int16, Int32, Schema, String
class JoinGroupResponse_v0(Response):
diff --git a/kafka/protocol/message.py b/kafka/protocol/message.py
index a330ed8..19dcbd9 100644
--- a/kafka/protocol/message.py
+++ b/kafka/protocol/message.py
@@ -3,15 +3,15 @@ from __future__ import absolute_import
import io
import time
-from ..codec import (has_gzip, has_snappy, has_lz4,
+from kafka.codec import (has_gzip, has_snappy, has_lz4,
gzip_decode, snappy_decode,
lz4_decode, lz4_decode_old_kafka)
-from .frame import KafkaBytes
-from .struct import Struct
-from .types import (
+from kafka.protocol.frame import KafkaBytes
+from kafka.protocol.struct import Struct
+from kafka.protocol.types import (
Int8, Int32, Int64, Bytes, Schema, AbstractType
)
-from ..util import crc32, WeakMethod
+from kafka.util import crc32, WeakMethod
class Message(Struct):
diff --git a/kafka/protocol/metadata.py b/kafka/protocol/metadata.py
index 2aafdd3..414e5b8 100644
--- a/kafka/protocol/metadata.py
+++ b/kafka/protocol/metadata.py
@@ -1,7 +1,7 @@
from __future__ import absolute_import
-from .api import Request, Response
-from .types import Array, Boolean, Int16, Int32, Schema, String
+from kafka.protocol.api import Request, Response
+from kafka.protocol.types import Array, Boolean, Int16, Int32, Schema, String
class MetadataResponse_v0(Response):
diff --git a/kafka/protocol/offset.py b/kafka/protocol/offset.py
index 5179658..3c254de 100644
--- a/kafka/protocol/offset.py
+++ b/kafka/protocol/offset.py
@@ -1,7 +1,7 @@
from __future__ import absolute_import
-from .api import Request, Response
-from .types import Array, Int8, Int16, Int32, Int64, Schema, String
+from kafka.protocol.api import Request, Response
+from kafka.protocol.types import Array, Int8, Int16, Int32, Int64, Schema, String
UNKNOWN_OFFSET = -1
diff --git a/kafka/protocol/produce.py b/kafka/protocol/produce.py
index 5fbddec..f4032b3 100644
--- a/kafka/protocol/produce.py
+++ b/kafka/protocol/produce.py
@@ -1,7 +1,7 @@
from __future__ import absolute_import
-from .api import Request, Response
-from .types import Int16, Int32, Int64, String, Array, Schema, Bytes
+from kafka.protocol.api import Request, Response
+from kafka.protocol.types import Int16, Int32, Int64, String, Array, Schema, Bytes
class ProduceResponse_v0(Response):
diff --git a/kafka/protocol/struct.py b/kafka/protocol/struct.py
index 3288172..676de1b 100644
--- a/kafka/protocol/struct.py
+++ b/kafka/protocol/struct.py
@@ -2,10 +2,10 @@ from __future__ import absolute_import
from io import BytesIO
-from .abstract import AbstractType
-from .types import Schema
+from kafka.protocol.abstract import AbstractType
+from kafka.protocol.types import Schema
-from ..util import WeakMethod
+from kafka.util import WeakMethod
class Struct(AbstractType):
diff --git a/kafka/protocol/types.py b/kafka/protocol/types.py
index 6a6e89e..d5e446a 100644
--- a/kafka/protocol/types.py
+++ b/kafka/protocol/types.py
@@ -2,7 +2,7 @@ from __future__ import absolute_import
from struct import pack, unpack, error
-from .abstract import AbstractType
+from kafka.protocol.abstract import AbstractType
def _pack(f, value):
diff --git a/kafka/record/__init__.py b/kafka/record/__init__.py
index 4c75acb..cbd70d9 100644
--- a/kafka/record/__init__.py
+++ b/kafka/record/__init__.py
@@ -1,3 +1,3 @@
-from .memory_records import MemoryRecords
+from kafka.record.memory_records import MemoryRecords
__all__ = ["MemoryRecords"]
diff --git a/kafka/record/default_records.py b/kafka/record/default_records.py
index 3d517af..2bbd47e 100644
--- a/kafka/record/default_records.py
+++ b/kafka/record/default_records.py
@@ -57,8 +57,8 @@
import io
import struct
import time
-from .abc import ABCRecord, ABCRecordBatch, ABCRecordBatchBuilder
-from .util import decode_varint, encode_varint, calc_crc32c, size_of_varint
+from kafka.record.abc import ABCRecord, ABCRecordBatch, ABCRecordBatchBuilder
+from kafka.record.util import decode_varint, encode_varint, calc_crc32c, size_of_varint
from kafka.errors import CorruptRecordException
from kafka.codec import (
diff --git a/kafka/record/legacy_records.py b/kafka/record/legacy_records.py
index 8c0791e..036e6c4 100644
--- a/kafka/record/legacy_records.py
+++ b/kafka/record/legacy_records.py
@@ -44,8 +44,8 @@
import struct
import time
-from .abc import ABCRecord, ABCRecordBatch, ABCRecordBatchBuilder
-from .util import calc_crc32
+from kafka.record.abc import ABCRecord, ABCRecordBatch, ABCRecordBatchBuilder
+from kafka.record.util import calc_crc32
from kafka.codec import (
gzip_encode, snappy_encode, lz4_encode, lz4_encode_old_kafka,
diff --git a/kafka/record/memory_records.py b/kafka/record/memory_records.py
index 56aa51f..cb1cc01 100644
--- a/kafka/record/memory_records.py
+++ b/kafka/record/memory_records.py
@@ -22,9 +22,9 @@
import struct
from kafka.errors import CorruptRecordException
-from .abc import ABCRecords
-from .legacy_records import LegacyRecordBatch, LegacyRecordBatchBuilder
-from .default_records import DefaultRecordBatch, DefaultRecordBatchBuilder
+from kafka.record.abc import ABCRecords
+from kafka.record.legacy_records import LegacyRecordBatch, LegacyRecordBatchBuilder
+from kafka.record.default_records import DefaultRecordBatch, DefaultRecordBatchBuilder
class MemoryRecords(ABCRecords):
diff --git a/kafka/record/util.py b/kafka/record/util.py
index 88135f1..55d7adb 100644
--- a/kafka/record/util.py
+++ b/kafka/record/util.py
@@ -1,6 +1,6 @@
import binascii
-from ._crc32c import crc as crc32c_py
+from kafka.record._crc32c import crc as crc32c_py
def encode_varint(value, write):
diff --git a/kafka/serializer/__init__.py b/kafka/serializer/__init__.py
index c08cffe..90cd93a 100644
--- a/kafka/serializer/__init__.py
+++ b/kafka/serializer/__init__.py
@@ -1,3 +1,3 @@
from __future__ import absolute_import
-from .abstract import Serializer, Deserializer
+from kafka.serializer.abstract import Serializer, Deserializer