About 50 results
Open links in new tab
  1. How Seq2Seq Context Vector is generated? - Stack Overflow

    Feb 23, 2023 · In a sequence-to-sequence (seq2seq) model, the context vector is a representation of the input sequence generated by the encoder and used by the decoder to generate the output …

  2. python - How to fine tune a Huggingface Seq2Seq model with a …

    Mar 27, 2023 · How to fine tune a Huggingface Seq2Seq model with a dataset from the hub? Asked 3 years ago Modified 3 years ago Viewed 9k times

  3. What's the difference between LSTM and Seq2Seq (M to 1)

    Mar 23, 2021 · Seq2seq is a family of machine learning approaches used for language processing. Applications include language translation, image captioning, conversational models and text …

  4. tensorflow - Understanding Seq2Seq model - Stack Overflow

    Sep 22, 2017 · The value of initial_state should be a tensor or list of tensors representing the initial state of the RNN layer. EDIT: There's now an example script in Keras (lstm_seq2seq.py) showing how to …

  5. Adding attention to seq2seq LSTM Model - Stack Overflow

    Jan 14, 2024 · Adding attention to seq2seq LSTM Model Ask Question Asked 2 years, 2 months ago Modified 2 years, 2 months ago

  6. How to resolve ValueError while training Seq2Seq using ...

    Feb 4, 2025 · I want to fine tune a VisionEncoderDecoderModel.from_pretrained(model_name) I use a CustomOCRDataset from Learn Open CV. But the default_data_collator fails to stack the inputs …

  7. seq2seq - What are differences between T5 and Bart? - Stack Overflow

    Dec 29, 2023 · Both models are suitable for most seq2seq tasks such as summarization, translation QA tasks, comprehension tasks, etc. Both of them issued in 2019) T5 by Google, BART by Facebook AI …

  8. how to convert HuggingFace's Seq2seq models to onnx format

    Feb 9, 2021 · I am trying to convert the Pegasus newsroom in HuggingFace's transformers model to the ONNX format. I followed this guide published by Huggingface. After installing the prereqs, I ran this …

  9. tensorflow - Understanding states of a bidirectional LSTM in a seq2seq ...

    Aug 22, 2022 · I am creating a language model: A seq2seq model with 2 Bidirectional LSTM layers. I have got the model to train and the accuracy seems good, but whilst stuck on figuring out the …

  10. Simple import codes in transformers cause errors - Stack Overflow

    Sep 12, 2024 · RuntimeError: Failed to import transformers.trainer_seq2seq because of the following error (look up to see its traceback): Failed to import transformers.integrations.integration_utils …