bashyaldhiraj2067 commited on
Commit
eb87ee5
·
verified ·
1 Parent(s): 0ec5c0a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +27 -10
app.py CHANGED
@@ -237,11 +237,18 @@ tokenizer = CharTokenizer()
237
  # =========================================================
238
  # 2. Model Definition (CUSTOM – REQUIRED)
239
  # =========================================================
240
- class TransformerCopyConfig(PretrainedConfig):
241
- model_type = "transformer_copy"
242
- def __init__(self, vocab_size=tokenizer.vocab_size, **kwargs):
243
- super().__init__(**kwargs)
244
- self.vocab_size = vocab_size
 
 
 
 
 
 
 
245
 
246
  class PositionalEncoding(nn.Module):
247
  def __init__(self, d_model, max_len=512):
@@ -264,15 +271,25 @@ class TransformerCopyModel(nn.Module):
264
  self.embedding = nn.Embedding(vocab_size, d_model)
265
  self.pos = PositionalEncoding(d_model)
266
 
267
- enc_layer = nn.TransformerEncoderLayer(d_model, nhead, dim_ff, dropout=dropout)
268
- dec_layer = nn.TransformerDecoderLayer(d_model, nhead, dim_ff, dropout=dropout)
269
-
270
- self.encoder = nn.TransformerEncoder(enc_layer, num_layers)
271
- self.decoder = nn.TransformerDecoder(dec_layer, num_layers)
 
 
 
 
 
 
 
272
 
 
 
273
  self.fc = nn.Linear(d_model, vocab_size)
274
 
275
 
 
276
  def forward(self, src, tgt):
277
  src_emb = self.pos(self.embedding(src))
278
  tgt_emb = self.pos(self.embedding(tgt))
 
237
  # =========================================================
238
  # 2. Model Definition (CUSTOM – REQUIRED)
239
  # =========================================================
240
+ class TransformerCopyHF(PreTrainedModel):
241
+ config_class = TransformerCopyConfig
242
+ def __init__(self, config):
243
+ super().__init__(config)
244
+ self.model = TransformerCopyModel(
245
+ vocab_size=config.vocab_size,
246
+ d_model=256,
247
+ nhead=8,
248
+ num_layers=4,
249
+ dim_ff=512
250
+ )
251
+
252
 
253
  class PositionalEncoding(nn.Module):
254
  def __init__(self, d_model, max_len=512):
 
271
  self.embedding = nn.Embedding(vocab_size, d_model)
272
  self.pos = PositionalEncoding(d_model)
273
 
274
+ encoder_layer = nn.TransformerEncoderLayer(
275
+ d_model=d_model,
276
+ nhead=nhead,
277
+ dim_feedforward=dim_ff,
278
+ dropout=dropout
279
+ )
280
+ decoder_layer = nn.TransformerDecoderLayer(
281
+ d_model=d_model,
282
+ nhead=nhead,
283
+ dim_feedforward=dim_ff,
284
+ dropout=dropout
285
+ )
286
 
287
+ self.encoder = nn.TransformerEncoder(encoder_layer, num_layers)
288
+ self.decoder = nn.TransformerDecoder(decoder_layer, num_layers)
289
  self.fc = nn.Linear(d_model, vocab_size)
290
 
291
 
292
+
293
  def forward(self, src, tgt):
294
  src_emb = self.pos(self.embedding(src))
295
  tgt_emb = self.pos(self.embedding(tgt))