python json dump to file
import json
with open('data.json', 'w') as f:
json.dump(data, f)
python json dump to file
import json
with open('data.json', 'w') as f:
json.dump(data, f)
json load from file python 3
import json
with open('file_to_load.json', 'r') as file:
data = json.load(file)
json load
import json
with open('path_to_file/person.json') as f:
data = json.load(f)
json load
import json
with open('path_to_file/person.json') as f:
data = json.load(f)
# Output: {'name': 'Bob', 'languages': ['English', 'Fench']}
print(data)
how to load cifar10 in python
"""Load from /home/USER/data/cifar10 or elsewhere; download if missing."""
import tarfile
import os
from urllib.request import urlretrieve
import numpy as np
def cifar10(path=None):
r"""Return (train_images, train_labels, test_images, test_labels).
Args:
path (str): Directory containing CIFAR-10. Default is
/home/USER/data/cifar10 or C:\Users\USER\data\cifar10.
Create if nonexistant. Download CIFAR-10 if missing.
Returns:
Tuple of (train_images, train_labels, test_images, test_labels), each
a matrix. Rows are examples. Columns of images are pixel values,
with the order (red -> blue -> green). Columns of labels are a
onehot encoding of the correct class.
"""
url = 'https://www.cs.toronto.edu/~kriz/'
tar = 'cifar-10-binary.tar.gz'
files = ['cifar-10-batches-bin/data_batch_1.bin',
'cifar-10-batches-bin/data_batch_2.bin',
'cifar-10-batches-bin/data_batch_3.bin',
'cifar-10-batches-bin/data_batch_4.bin',
'cifar-10-batches-bin/data_batch_5.bin',
'cifar-10-batches-bin/test_batch.bin']
if path is None:
# Set path to /home/USER/data/mnist or C:\Users\USER\data\mnist
path = os.path.join(os.path.expanduser('~'), 'data', 'cifar10')
# Create path if it doesn't exist
os.makedirs(path, exist_ok=True)
# Download tarfile if missing
if tar not in os.listdir(path):
urlretrieve(''.join((url, tar)), os.path.join(path, tar))
print("Downloaded %s to %s" % (tar, path))
# Load data from tarfile
with tarfile.open(os.path.join(path, tar)) as tar_object:
# Each file contains 10,000 color images and 10,000 labels
fsize = 10000 * (32 * 32 * 3) + 10000
# There are 6 files (5 train and 1 test)
buffr = np.zeros(fsize * 6, dtype='uint8')
# Get members of tar corresponding to data files
# -- The tar contains README's and other extraneous stuff
members = [file for file in tar_object if file.name in files]
# Sort those members by name
# -- Ensures we load train data in the proper order
# -- Ensures that test data is the last file in the list
members.sort(key=lambda member: member.name)
# Extract data from members
for i, member in enumerate(members):
# Get member as a file object
f = tar_object.extractfile(member)
# Read bytes from that file object into buffr
buffr[i * fsize:(i + 1) * fsize] = np.frombuffer(f.read(), 'B')
# Parse data from buffer
# -- Examples are in chunks of 3,073 bytes
# -- First byte of each chunk is the label
# -- Next 32 * 32 * 3 = 3,072 bytes are its corresponding image
# Labels are the first byte of every chunk
labels = buffr[::3073]
# Pixels are everything remaining after we delete the labels
pixels = np.delete(buffr, np.arange(0, buffr.size, 3073))
images = pixels.reshape(-1, 3072).astype('float32') / 255
# Split into train and test
train_images, test_images = images[:50000], images[50000:]
train_labels, test_labels = labels[:50000], labels[50000:]
def _onehot(integer_labels):
"""Return matrix whose rows are onehot encodings of integers."""
n_rows = len(integer_labels)
n_cols = integer_labels.max() + 1
onehot = np.zeros((n_rows, n_cols), dtype='uint8')
onehot[np.arange(n_rows), integer_labels] = 1
return onehot
return train_images, _onehot(train_labels), \
test_images, _onehot(test_labels)
Copyright © 2021 Codeinu
Forgot your account's password or having trouble logging into your Account? Don't worry, we'll help you to get back your account. Enter your email address and we'll send you a recovery link to reset your password. If you are experiencing problems resetting your password contact us