Instructions to use nvidia/E-RADIO with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use nvidia/E-RADIO with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("feature-extraction", model="nvidia/E-RADIO", trust_remote_code=True)# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("nvidia/E-RADIO", trust_remote_code=True, dtype="auto") - Notebooks
- Google Colab
- Kaggle
| # Copyright (c) 2024, NVIDIA CORPORATION. All rights reserved. | |
| # | |
| # NVIDIA CORPORATION and its licensors retain all intellectual property | |
| # and proprietary rights in and to this software, related documentation | |
| # and any modifications thereto. Any use, reproduction, disclosure or | |
| # distribution of this software and related documentation without an express | |
| # license agreement from NVIDIA CORPORATION is strictly prohibited. | |
| import math | |
| from typing import Dict | |
| import torch | |
| from torch import nn | |
| from einops import rearrange | |
| from timm.models.vision_transformer import Block | |
| class MLP(nn.Module): | |
| def __init__(self, input_size: int, hidden_size: int, output_size: int, | |
| num_inner: int = 0, device: torch.device = None, **kwargs): | |
| super(MLP, self).__init__() | |
| self.fc1 = nn.Linear(input_size, hidden_size, device=device) | |
| self.norm = nn.LayerNorm(hidden_size, device=device) | |
| self.relu = nn.ReLU() | |
| inner = [] | |
| for _ in range(num_inner): | |
| inner.extend([ | |
| nn.Linear(hidden_size, hidden_size, device=device), | |
| nn.LayerNorm(hidden_size, device=device), | |
| nn.ReLU(), | |
| ]) | |
| if inner: | |
| self.inner = nn.Sequential(*inner) | |
| else: | |
| self.inner = nn.Identity() | |
| self.fc2 = nn.Linear(hidden_size, output_size, device=device) | |
| def forward(self, x: torch.Tensor) -> torch.Tensor: | |
| x = self.fc1(x) | |
| x = self.norm(x) | |
| x = self.relu(x) | |
| x = self.inner(x) | |
| x = self.fc2(x) | |
| return x | |
| class MLP2(nn.Module): | |
| def __init__(self, input_size: int, hidden_size: int, output_size: int, | |
| num_inner: int = 0, | |
| pre_norm: bool = False, device: torch.device = None, | |
| upsample_factor: int = 1, | |
| **kwargs): | |
| super().__init__() | |
| self.pre_norm = nn.Sequential( | |
| nn.LayerNorm(input_size), | |
| nn.GELU(), | |
| ) if pre_norm else nn.Identity() | |
| self.upsample_factor = upsample_factor | |
| self._real_output_dim = output_size | |
| hidden_size *= upsample_factor | |
| output_size *= (upsample_factor ** 2) | |
| self.fc1 = nn.Linear(input_size, hidden_size, device=device) | |
| blocks = [] | |
| for _ in range(num_inner): | |
| blocks.append(nn.Sequential( | |
| nn.LayerNorm(hidden_size, device=device), | |
| nn.GELU(), | |
| nn.Linear(hidden_size, hidden_size, device=device), | |
| )) | |
| self.blocks = nn.ModuleList(blocks) | |
| self.final = nn.Sequential( | |
| nn.LayerNorm(hidden_size, device=device), | |
| nn.GELU(), | |
| nn.Linear(hidden_size, output_size, device=device), | |
| ) | |
| def forward(self, x: torch.Tensor) -> torch.Tensor: | |
| x = self.pre_norm(x) | |
| x = self.fc1(x) | |
| for block in self.blocks: | |
| x = x + block(x) | |
| x = self.final(x) | |
| if self.upsample_factor > 1: | |
| h = w = int(math.sqrt(x.shape[1])) | |
| x = rearrange(x, 'b (h w) (u1 u2 c) -> b (u1 h u2 w) c', | |
| h=h, w=w, u1=self.upsample_factor, u2=self.upsample_factor, | |
| c=self._real_output_dim) | |
| return x | |
| MLP_FACTORY = { | |
| 'v1': MLP, | |
| 'v2': MLP2, | |
| } | |
| def strip_prefix(state: Dict[str, torch.Tensor], prefix: str): | |
| state = { | |
| k[len(prefix):]: v | |
| for k, v in state.items() | |
| if k.startswith(prefix) | |
| } | |
| return state | |
| def get_mlp_info_from_state(version: str, state: Dict[str, torch.Tensor], prefix: str = ''): | |
| state = strip_prefix(state, prefix) | |
| if version == 'v1': | |
| hidden_dim, input_dim = state['fc1.weight'].shape | |
| output_dim = state['fc2.weight'].shape[0] | |
| for num_inner in range(1000): | |
| k = f'inner.{num_inner}.0.weight' | |
| if k not in state: | |
| break | |
| elif version == 'v2': | |
| hidden_dim, input_dim = state['fc1.weight'].shape | |
| output_dim = state['final.2.weight'].shape[0] | |
| for num_inner in range(1000): | |
| k = f'blocks.{num_inner}.0.weight' | |
| if k not in state: | |
| break | |
| else: | |
| raise ValueError(f'Unsupported MLP version: {version}') | |
| return input_dim, hidden_dim, output_dim, num_inner | |
| def create_mlp_from_state(version: str, state: Dict[str, torch.Tensor], prefix: str = ''): | |
| state = strip_prefix(state, prefix) | |
| input_dim, hidden_dim, output_dim, num_inner = get_mlp_info_from_state(version, state) | |
| ret: nn.Module = MLP_FACTORY[version](input_dim, hidden_dim, output_dim, num_inner) | |
| ret.load_state_dict(state) | |
| return ret | |