You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
 
 
 
 
 
 

110 lines
3.5 KiB

#!/home/alpcentaur/ProjektA/PrototypeWebApp/venv/bin/python3.5
import argparse
import os
import boto
from boto.compat import json
from boto.compat import six
from boto.dynamodb.schema import Schema
DESCRIPTION = """Load data into one or more DynamoDB tables.
For each table, data is read from two files:
- {table_name}.metadata for the table's name, schema and provisioned
throughput (only required if creating the table).
- {table_name}.data for the table's actual contents.
Both files are searched for in the current directory. To read them from
somewhere else, use the --in-dir parameter.
This program does not wipe the tables prior to loading data. However, any
items present in the data files will overwrite the table's contents.
"""
def _json_iterload(fd):
"""Lazily load newline-separated JSON objects from a file-like object."""
buffer = ""
eof = False
while not eof:
try:
# Add a line to the buffer
buffer += fd.next()
except StopIteration:
# We can't let that exception bubble up, otherwise the last
# object in the file will never be decoded.
eof = True
try:
# Try to decode a JSON object.
json_object = json.loads(buffer.strip())
# Success: clear the buffer (everything was decoded).
buffer = ""
except ValueError:
if eof and buffer.strip():
# No more lines to load and the buffer contains something other
# than whitespace: the file is, in fact, malformed.
raise
# We couldn't decode a complete JSON object: load more lines.
continue
yield json_object
def create_table(metadata_fd):
"""Create a table from a metadata file-like object."""
def load_table(table, in_fd):
"""Load items into a table from a file-like object."""
for i in _json_iterload(in_fd):
# Convert lists back to sets.
data = {}
for k, v in six.iteritems(i):
if isinstance(v, list):
data[k] = set(v)
else:
data[k] = v
table.new_item(attrs=data).put()
def dynamodb_load(tables, in_dir, create_tables):
conn = boto.connect_dynamodb()
for t in tables:
metadata_file = os.path.join(in_dir, "%s.metadata" % t)
data_file = os.path.join(in_dir, "%s.data" % t)
if create_tables:
with open(metadata_file) as meta_fd:
metadata = json.load(meta_fd)
table = conn.create_table(
name=t,
schema=Schema(metadata["schema"]),
read_units=metadata["read_units"],
write_units=metadata["write_units"],
)
table.refresh(wait_for_active=True)
else:
table = conn.get_table(t)
with open(data_file) as in_fd:
load_table(table, in_fd)
if __name__ == "__main__":
parser = argparse.ArgumentParser(
prog="dynamodb_load",
description=DESCRIPTION
)
parser.add_argument(
"--create-tables",
action="store_true",
help="Create the tables if they don't exist already (without this flag, attempts to load data into non-existing tables fail)."
)
parser.add_argument("--in-dir", default=".")
parser.add_argument("tables", metavar="TABLES", nargs="+")
namespace = parser.parse_args()
dynamodb_load(namespace.tables, namespace.in_dir, namespace.create_tables)