Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -237,11 +237,18 @@ tokenizer = CharTokenizer()
|
|
| 237 |
# =========================================================
|
| 238 |
# 2. Model Definition (CUSTOM – REQUIRED)
|
| 239 |
# =========================================================
|
| 240 |
-
class
|
| 241 |
-
|
| 242 |
-
def __init__(self,
|
| 243 |
-
super().__init__(
|
| 244 |
-
self.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 245 |
|
| 246 |
class PositionalEncoding(nn.Module):
|
| 247 |
def __init__(self, d_model, max_len=512):
|
|
@@ -264,15 +271,25 @@ class TransformerCopyModel(nn.Module):
|
|
| 264 |
self.embedding = nn.Embedding(vocab_size, d_model)
|
| 265 |
self.pos = PositionalEncoding(d_model)
|
| 266 |
|
| 267 |
-
|
| 268 |
-
|
| 269 |
-
|
| 270 |
-
|
| 271 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 272 |
|
|
|
|
|
|
|
| 273 |
self.fc = nn.Linear(d_model, vocab_size)
|
| 274 |
|
| 275 |
|
|
|
|
| 276 |
def forward(self, src, tgt):
|
| 277 |
src_emb = self.pos(self.embedding(src))
|
| 278 |
tgt_emb = self.pos(self.embedding(tgt))
|
|
|
|
| 237 |
# =========================================================
|
| 238 |
# 2. Model Definition (CUSTOM – REQUIRED)
|
| 239 |
# =========================================================
|
| 240 |
+
class TransformerCopyHF(PreTrainedModel):
|
| 241 |
+
config_class = TransformerCopyConfig
|
| 242 |
+
def __init__(self, config):
|
| 243 |
+
super().__init__(config)
|
| 244 |
+
self.model = TransformerCopyModel(
|
| 245 |
+
vocab_size=config.vocab_size,
|
| 246 |
+
d_model=256,
|
| 247 |
+
nhead=8,
|
| 248 |
+
num_layers=4,
|
| 249 |
+
dim_ff=512
|
| 250 |
+
)
|
| 251 |
+
|
| 252 |
|
| 253 |
class PositionalEncoding(nn.Module):
|
| 254 |
def __init__(self, d_model, max_len=512):
|
|
|
|
| 271 |
self.embedding = nn.Embedding(vocab_size, d_model)
|
| 272 |
self.pos = PositionalEncoding(d_model)
|
| 273 |
|
| 274 |
+
encoder_layer = nn.TransformerEncoderLayer(
|
| 275 |
+
d_model=d_model,
|
| 276 |
+
nhead=nhead,
|
| 277 |
+
dim_feedforward=dim_ff,
|
| 278 |
+
dropout=dropout
|
| 279 |
+
)
|
| 280 |
+
decoder_layer = nn.TransformerDecoderLayer(
|
| 281 |
+
d_model=d_model,
|
| 282 |
+
nhead=nhead,
|
| 283 |
+
dim_feedforward=dim_ff,
|
| 284 |
+
dropout=dropout
|
| 285 |
+
)
|
| 286 |
|
| 287 |
+
self.encoder = nn.TransformerEncoder(encoder_layer, num_layers)
|
| 288 |
+
self.decoder = nn.TransformerDecoder(decoder_layer, num_layers)
|
| 289 |
self.fc = nn.Linear(d_model, vocab_size)
|
| 290 |
|
| 291 |
|
| 292 |
+
|
| 293 |
def forward(self, src, tgt):
|
| 294 |
src_emb = self.pos(self.embedding(src))
|
| 295 |
tgt_emb = self.pos(self.embedding(tgt))
|