Pytorch 資料載入與資料預處理方式
資料載入分為載入torchvision.datasets中的資料集以及載入自己使用的資料集兩種情況。
torchvision.datasets中的資料集
torchvision.datasets中自帶MNIST,Imagenet-12,CIFAR等資料集,所有的資料集都是torch.utils.data.Dataset的子類,都包含 _ _ len _ (獲取資料集長度)和 _ getItem _ _ (獲取資料集中每一項)兩個子方法。
Dataset原始碼如上,可以看到其中包含了兩個沒有實現的子方法,之後所有的Dataet類都繼承該類,並根據資料情況定製這兩個子方法的具體實現。
因此當我們需要載入自己的資料集的時候也可以借鑑這種方法,只需要繼承torch.utils.data.Dataset類並重寫 init,len,以及getitem這三個方法即可。這樣組著的類可以直接作為引數傳入到torch.util.data.DataLoader中去。
以CIFAR10為例 原始碼:
class torchvision.datasets.CIFAR10(root,train=True,transform=None,target_transform=None,download=False)
root (string) – Root directory of dataset where directory cifar-10-batches-py exists or will be saved to if download is set to True. train (bool,optional) – If True,creates dataset from training set,otherwise creates from test set. transform (callable,optional) – A function/transform that takes in an PIL image and returns a transformed version. E.g,transforms.RandomCrop target_transform (callable,optional) – A function/transform that takes in the target and transforms it. download (bool,optional) – If true,downloads the dataset from the internet and puts it in root directory. If dataset is already downloaded,it is not downloaded again.
載入自己的資料集
對於torchvision.datasets中有兩個不同的類,分別為DatasetFolder和ImageFolder,ImageFolder是繼承自DatasetFolder。
下面我們通過原始碼來看一看folder檔案中DatasetFolder和ImageFolder分別做了些什麼
import torch.utils.data as data from PIL import Image import os import os.path def has_file_allowed_extension(filename,extensions): //檢查輸入是否是規定的副檔名 """Checks if a file is an allowed extension. Args: filename (string): path to a file Returns: bool: True if the filename ends with a known image extension """ filename_lower = filename.lower() return any(filename_lower.endswith(ext) for ext in extensions) def find_classes(dir): classes = [d for d in os.listdir(dir) if os.path.isdir(os.path.join(dir,d))] //獲取root目錄下所有的資料夾名稱 classes.sort() class_to_idx = {classes[i]: i for i in range(len(classes))} //生成類別名稱與類別id的對應Dictionary return classes,class_to_idx def make_dataset(dir,class_to_idx,extensions): images = [] dir = os.path.expanduser(dir)// 將~和~user轉化為使用者目錄,對引數中出現~進行處理 for target in sorted(os.listdir(dir)): d = os.path.join(dir,target) if not os.path.isdir(d): continue for root,_,fnames in sorted(os.walk(d)): //os.work包含三個部分,root代表該目錄路徑 _代表該路徑下的資料夾名稱集合,fnames代表該路徑下的檔名稱集合 for fname in sorted(fnames): if has_file_allowed_extension(fname,extensions): path = os.path.join(root,fname) item = (path,class_to_idx[target]) images.append(item) //生成(訓練樣本影象目錄,訓練樣本所屬類別)的元組 return images //返回上述元組的列表 class DatasetFolder(data.Dataset): """A generic data loader where the samples are arranged in this way: :: root/class_x/xxx.ext root/class_x/xxy.ext root/class_x/xxz.ext root/class_y/123.ext root/class_y/nsdf3.ext root/class_y/asd932_.ext Args: root (string): Root directory path. loader (callable): A function to load a sample given its path. extensions (list[string]): A list of allowed extensions. transform (callable,optional): A function/transform that takes in a sample and returns a transformed version. E.g,``transforms.RandomCrop`` for images. target_transform (callable,optional): A function/transform that takes in the target and transforms it. Attributes: classes (list): List of the class names. class_to_idx (dict): Dict with items (class_name,class_index). samples (list): List of (sample path,class_index) tuples """ def __init__(self,root,loader,extensions,target_transform=None): classes,class_to_idx = find_classes(root) samples = make_dataset(root,extensions) if len(samples) == 0: raise(RuntimeError("Found 0 files in subfolders of: " + root + "\n" "Supported extensions are: " + ",".join(extensions))) self.root = root self.loader = loader self.extensions = extensions self.classes = classes self.class_to_idx = class_to_idx self.samples = samples self.transform = transform self.target_transform = target_transform def __getitem__(self,index): """ 根據index獲取sample 返回值為(sample,target)元組,同時如果該類輸入引數中有transform和target_transform,torchvision.transforms型別的引數時,將獲取的元組分別執行transform和target_transform中的資料轉換方法。 Args: index (int): Index Returns: tuple: (sample,target) where target is class_index of the target class. """ path,target = self.samples[index] sample = self.loader(path) if self.transform is not None: sample = self.transform(sample) if self.target_transform is not None: target = self.target_transform(target) return sample,target def __len__(self): return len(self.samples) def __repr__(self): //定義輸出物件格式 其中和__str__的區別是__repr__無論是print輸出還是直接輸出物件自身 都是以定義的格式進行輸出,而__str__ 只有在print輸出的時候會是以定義的格式進行輸出 fmt_str = 'Dataset ' + self.__class__.__name__ + '\n' fmt_str += ' Number of datapoints: {}\n'.format(self.__len__()) fmt_str += ' Root Location: {}\n'.format(self.root) tmp = ' Transforms (if any): ' fmt_str += '{0}{1}\n'.format(tmp,self.transform.__repr__().replace('\n','\n' + ' ' * len(tmp))) tmp = ' Target Transforms (if any): ' fmt_str += '{0}{1}'.format(tmp,self.target_transform.__repr__().replace('\n','\n' + ' ' * len(tmp))) return fmt_str IMG_EXTENSIONS = ['.jpg','.jpeg','.png','.ppm','.bmp','.pgm','.tif'] def pil_loader(path): # open path as file to avoid ResourceWarning (https://github.com/python-pillow/Pillow/issues/835) with open(path,'rb') as f: img = Image.open(f) return img.convert('RGB') def accimage_loader(path): import accimage try: return accimage.Image(path) except IOError: # Potentially a decoding problem,fall back to PIL.Image return pil_loader(path) def default_loader(path): from torchvision import get_image_backend if get_image_backend() == 'accimage': return accimage_loader(path) else: return pil_loader(path) class ImageFolder(DatasetFolder): """A generic data loader where the images are arranged in this way: :: root/dog/xxx.png root/dog/xxy.png root/dog/xxz.png root/cat/123.png root/cat/nsdf3.png root/cat/asd932_.png Args: root (string): Root directory path. transform (callable,optional): A function/transform that takes in an PIL image and returns a transformed version. E.g,``transforms.RandomCrop`` target_transform (callable,optional): A function/transform that takes in the target and transforms it. loader (callable,optional): A function to load an image given its path. Attributes: classes (list): List of the class names. class_to_idx (dict): Dict with items (class_name,class_index). imgs (list): List of (image path,class_index) tuples """ def __init__(self,loader=default_loader): super(ImageFolder,self).__init__(root,IMG_EXTENSIONS,transform=transform,target_transform=target_transform) self.imgs = self.samples
如果自己所要載入的資料組織形式如下
root/dog/xxx.png root/dog/xxy.png root/dog/xxz.png root/cat/123.png root/cat/nsdf3.png root/cat/asd932_.png
即不同類別的訓練資料分別儲存在不同的資料夾中,這些資料夾都在root(即形如 D:/animals 或者 /usr/animals )路徑下
class torchvision.datasets.ImageFolder(root,loader=<function default_loader>)
引數如下:
root (string) – Root directory path. transform (callable,optional) – A function/transform that takes in the target and transforms it. loader – A function to load an image given its path. 就是上述原始碼中 __getitem__(index) Parameters: index (int) – Index Returns: (sample,target) where target is class_index of the target class. Return type: tuple
可以通過torchvision.datasets.ImageFolder進行載入
img_data = torchvision.datasets.ImageFolder('D:/bnu/database/flower',transform=transforms.Compose([ transforms.Scale(256),transforms.CenterCrop(224),transforms.ToTensor()]) ) print(len(img_data)) data_loader = torch.utils.data.DataLoader(img_data,batch_size=20,shuffle=True) print(len(data_loader))
對於所有的訓練樣本都在一個資料夾中 同時有一個對應的txt檔案每一行分別是對應影象的路徑以及其所屬的類別,可以參照上述class寫出對應的載入類
def default_loader(path): return Image.open(path).convert('RGB') class MyDataset(Dataset): def __init__(self,txt,loader=default_loader): fh = open(txt,'r') imgs = [] for line in fh: line = line.strip('\n') line = line.rstrip() words = line.split() imgs.append((words[0],int(words[1]))) self.imgs = imgs self.transform = transform self.target_transform = target_transform self.loader = loader def __getitem__(self,index): fn,label = self.imgs[index] img = self.loader(fn) if self.transform is not None: img = self.transform(img) return img,label def __len__(self): return len(self.imgs) train_data=MyDataset(txt='mnist_test.txt',transform=transforms.ToTensor()) data_loader = DataLoader(train_data,batch_size=100,shuffle=True) print(len(data_loader))
DataLoader解析
位於torch.util.data.DataLoader中 原始碼
該介面的主要目的是將pytorch中已有的資料介面如torchvision.datasets.ImageFolder,或者自定義的資料讀取介面轉化按照
batch_size的大小封裝為Tensor,即相當於在內建資料介面或者自定義資料介面的基礎上增加一維,大小為batch_size的大小,
得到的資料在之後可以通過封裝為Variable,作為模型的輸出
_ _ init _ _中所需的引數如下
1. dataset torch.utils.data.Dataset類的子類,可以是torchvision.datasets.ImageFolder等內建類,也可是繼承了torch.utils.data.Dataset的自定義類 2. batch_size 每一個batch中包含的樣本個數,預設是1 3. shuffle 一般在訓練集中採用,預設是false,設定為true則每一個epoch都會將訓練樣本打亂 4. sampler 訓練樣本選取策略,和shuffle是互斥的 如果 shuffle為true,該引數一定要為None 5. batch_sampler BatchSampler 一次產生一個 batch 的 indices,和sampler以及shuffle互斥,一般使用預設的即可 上述Sampler的原始碼地址如下[原始碼](https://github.com/pytorch/pytorch/blob/master/torch/utils/data/sampler.py) 6. num_workers 用於資料載入的執行緒數量 預設為0 即只有主執行緒用來載入資料 7. collate_fn 用來聚合資料生成mini_batch
使用的時候一般為如下使用方法:
train_data=torch.utils.data.DataLoader(...) for i,(input,target) in enumerate(train_data): ...
迴圈取DataLoader中的資料會觸發類中_ _ iter __方法,檢視原始碼可知 其中呼叫的方法為 return _DataLoaderIter(self),因此需要檢視 DataLoaderIter 這一內部類
class DataLoaderIter(object): "Iterates once over the DataLoader's dataset,as specified by the sampler" def __init__(self,loader): self.dataset = loader.dataset self.collate_fn = loader.collate_fn self.batch_sampler = loader.batch_sampler self.num_workers = loader.num_workers self.pin_memory = loader.pin_memory and torch.cuda.is_available() self.timeout = loader.timeout self.done_event = threading.Event() self.sample_iter = iter(self.batch_sampler) if self.num_workers > 0: self.worker_init_fn = loader.worker_init_fn self.index_queue = multiprocessing.SimpleQueue() self.worker_result_queue = multiprocessing.SimpleQueue() self.batches_outstanding = 0 self.worker_pids_set = False self.shutdown = False self.send_idx = 0 self.rcvd_idx = 0 self.reorder_dict = {} base_seed = torch.LongTensor(1).random_()[0] self.workers = [ multiprocessing.Process( target=_worker_loop,args=(self.dataset,self.index_queue,self.worker_result_queue,self.collate_fn,base_seed + i,self.worker_init_fn,i)) for i in range(self.num_workers)] if self.pin_memory or self.timeout > 0: self.data_queue = queue.Queue() self.worker_manager_thread = threading.Thread( target=_worker_manager_loop,args=(self.worker_result_queue,self.data_queue,self.done_event,self.pin_memory,torch.cuda.current_device())) self.worker_manager_thread.daemon = True self.worker_manager_thread.start() else: self.data_queue = self.worker_result_queue for w in self.workers: w.daemon = True # ensure that the worker exits on process exit w.start() _update_worker_pids(id(self),tuple(w.pid for w in self.workers)) _set_SIGCHLD_handler() self.worker_pids_set = True # prime the prefetch loop for _ in range(2 * self.num_workers): self._put_indices()
以上這篇Pytorch 資料載入與資料預處理方式就是小編分享給大家的全部內容了,希望能給大家一個參考,也希望大家多多支援我們。